diff options
Diffstat (limited to '')
-rw-r--r-- | src/mux_fcgi.c | 4268 |
1 files changed, 4268 insertions, 0 deletions
diff --git a/src/mux_fcgi.c b/src/mux_fcgi.c new file mode 100644 index 0000000..0230e6b --- /dev/null +++ b/src/mux_fcgi.c @@ -0,0 +1,4268 @@ +/* + * FastCGI mux-demux for connections + * + * Copyright (C) 2019 HAProxy Technologies, Christopher Faulet <cfaulet@haproxy.com> + * + * This program is free software; you can redistribute it and/or + * modify it under the terms of the GNU General Public License + * as published by the Free Software Foundation; either version + * 2 of the License, or (at your option) any later version. + * + */ + +#include <import/ist.h> +#include <import/eb32tree.h> +#include <import/ebmbtree.h> + +#include <haproxy/api.h> +#include <haproxy/cfgparse.h> +#include <haproxy/connection.h> +#include <haproxy/dynbuf.h> +#include <haproxy/errors.h> +#include <haproxy/fcgi-app.h> +#include <haproxy/fcgi.h> +#include <haproxy/h1.h> +#include <haproxy/h1_htx.h> +#include <haproxy/http_htx.h> +#include <haproxy/htx.h> +#include <haproxy/list.h> +#include <haproxy/log.h> +#include <haproxy/mux_fcgi-t.h> +#include <haproxy/net_helper.h> +#include <haproxy/proxy.h> +#include <haproxy/regex.h> +#include <haproxy/sc_strm.h> +#include <haproxy/server.h> +#include <haproxy/session-t.h> +#include <haproxy/stconn.h> +#include <haproxy/stream.h> +#include <haproxy/trace.h> +#include <haproxy/version.h> + +/* 32 buffers: one for the ring's root, rest for the mbuf itself */ +#define FCGI_C_MBUF_CNT 32 + +/* Size for a record header (also size of empty record) */ +#define FCGI_RECORD_HEADER_SZ 8 + +/* FCGI connection descriptor */ +struct fcgi_conn { + struct connection *conn; + + enum fcgi_conn_st state; /* FCGI connection state */ + int16_t max_id; /* highest ID known on this connection, <0 before mgmt records */ + uint32_t streams_limit; /* maximum number of concurrent streams the peer supports */ + uint32_t flags; /* Connection flags: FCGI_CF_* */ + + int16_t dsi; /* dmux stream ID (<0 = idle ) */ + uint16_t drl; /* demux record length (if dsi >= 0) */ + uint8_t drt; /* demux record type (if dsi >= 0) */ + uint8_t drp; /* demux record padding (if dsi >= 0) */ + + struct buffer dbuf; /* demux buffer */ + struct buffer mbuf[FCGI_C_MBUF_CNT]; /* mux buffers (ring) */ + + int timeout; /* idle timeout duration in ticks */ + int shut_timeout; /* idle timeout duration in ticks after shutdown */ + unsigned int nb_streams; /* number of streams in the tree */ + unsigned int nb_sc; /* number of attached stream connectors */ + unsigned int nb_reserved; /* number of reserved streams */ + unsigned int stream_cnt; /* total number of streams seen */ + + struct proxy *proxy; /* the proxy this connection was created for */ + struct fcgi_app *app; /* FCGI application used by this mux */ + struct task *task; /* timeout management task */ + struct eb_root streams_by_id; /* all active streams by their ID */ + + struct list send_list; /* list of blocked streams requesting to send */ + + struct buffer_wait buf_wait; /* Wait list for buffer allocation */ + struct wait_event wait_event; /* To be used if we're waiting for I/Os */ +}; + + +/* FCGI stream descriptor */ +struct fcgi_strm { + struct sedesc *sd; + struct session *sess; + struct fcgi_conn *fconn; + + int32_t id; /* stream ID */ + + uint32_t flags; /* Connection flags: FCGI_SF_* */ + enum fcgi_strm_st state; /* FCGI stream state */ + int proto_status; /* FCGI_PS_* */ + + struct h1m h1m; /* response parser state for H1 */ + + struct buffer rxbuf; /* receive buffer, always valid (buf_empty or real buffer) */ + + struct eb32_node by_id; /* place in fcgi_conn's streams_by_id */ + struct wait_event *subs; /* Address of the wait_event the stream connector associated is waiting on */ + struct list send_list; /* To be used when adding in fcgi_conn->send_list */ + struct tasklet *shut_tl; /* deferred shutdown tasklet, to retry to close after we failed to by lack of space */ +}; + +/* Flags representing all default FCGI parameters */ +#define FCGI_SP_CGI_GATEWAY 0x00000001 +#define FCGI_SP_DOC_ROOT 0x00000002 +#define FCGI_SP_SCRIPT_NAME 0x00000004 +#define FCGI_SP_PATH_INFO 0x00000008 +#define FCGI_SP_REQ_URI 0x00000010 +#define FCGI_SP_REQ_METH 0x00000020 +#define FCGI_SP_REQ_QS 0x00000040 +#define FCGI_SP_SRV_PORT 0x00000080 +#define FCGI_SP_SRV_PROTO 0x00000100 +#define FCGI_SP_SRV_NAME 0x00000200 +#define FCGI_SP_REM_ADDR 0x00000400 +#define FCGI_SP_REM_PORT 0x00000800 +#define FCGI_SP_SCRIPT_FILE 0x00001000 +#define FCGI_SP_PATH_TRANS 0x00002000 +#define FCGI_SP_CONT_LEN 0x00004000 +#define FCGI_SP_HTTPS 0x00008000 +#define FCGI_SP_SRV_SOFT 0x00010000 +#define FCGI_SP_MASK 0x0001FFFF +#define FCGI_SP_URI_MASK (FCGI_SP_SCRIPT_NAME|FCGI_SP_PATH_INFO|FCGI_SP_REQ_QS) + +/* FCGI parameters used when PARAMS record is sent */ +struct fcgi_strm_params { + uint32_t mask; + struct ist docroot; + struct ist scriptname; + struct ist pathinfo; + struct ist meth; + struct ist uri; + struct ist vsn; + struct ist qs; + struct ist srv_name; + struct ist srv_port; + struct ist rem_addr; + struct ist rem_port; + struct ist cont_len; + struct ist srv_soft; + int https; + struct buffer *p; +}; + +/* Maximum amount of data we're OK with re-aligning for buffer optimizations */ +#define MAX_DATA_REALIGN 1024 + +/* trace source and events */ +static void fcgi_trace(enum trace_level level, uint64_t mask, + const struct trace_source *src, + const struct ist where, const struct ist func, + const void *a1, const void *a2, const void *a3, const void *a4); + +/* The event representation is split like this : + * fconn - internal FCGI connection + * fstrm - internal FCGI stream + * strm - application layer + * rx - data receipt + * tx - data transmission + * rsp - response parsing + */ +static const struct trace_event fcgi_trace_events[] = { +#define FCGI_EV_FCONN_NEW (1ULL << 0) + { .mask = FCGI_EV_FCONN_NEW, .name = "fconn_new", .desc = "new FCGI connection" }, +#define FCGI_EV_FCONN_RECV (1ULL << 1) + { .mask = FCGI_EV_FCONN_RECV, .name = "fconn_recv", .desc = "Rx on FCGI connection" }, +#define FCGI_EV_FCONN_SEND (1ULL << 2) + { .mask = FCGI_EV_FCONN_SEND, .name = "fconn_send", .desc = "Tx on FCGI connection" }, +#define FCGI_EV_FCONN_BLK (1ULL << 3) + { .mask = FCGI_EV_FCONN_BLK, .name = "fconn_blk", .desc = "FCGI connection blocked" }, +#define FCGI_EV_FCONN_WAKE (1ULL << 4) + { .mask = FCGI_EV_FCONN_WAKE, .name = "fconn_wake", .desc = "FCGI connection woken up" }, +#define FCGI_EV_FCONN_END (1ULL << 5) + { .mask = FCGI_EV_FCONN_END, .name = "fconn_end", .desc = "FCGI connection terminated" }, +#define FCGI_EV_FCONN_ERR (1ULL << 6) + { .mask = FCGI_EV_FCONN_ERR, .name = "fconn_err", .desc = "error on FCGI connection" }, + +#define FCGI_EV_RX_FHDR (1ULL << 7) + { .mask = FCGI_EV_RX_FHDR, .name = "rx_fhdr", .desc = "FCGI record header received" }, +#define FCGI_EV_RX_RECORD (1ULL << 8) + { .mask = FCGI_EV_RX_RECORD, .name = "rx_record", .desc = "receipt of any FCGI record" }, +#define FCGI_EV_RX_EOI (1ULL << 9) + { .mask = FCGI_EV_RX_EOI, .name = "rx_eoi", .desc = "receipt of end of FCGI input" }, +#define FCGI_EV_RX_GETVAL (1ULL << 10) + { .mask = FCGI_EV_RX_GETVAL, .name = "rx_get_values", .desc = "receipt of FCGI GET_VALUES_RESULT record" }, +#define FCGI_EV_RX_STDOUT (1ULL << 11) + { .mask = FCGI_EV_RX_STDOUT, .name = "rx_stdout", .desc = "receipt of FCGI STDOUT record" }, +#define FCGI_EV_RX_STDERR (1ULL << 12) + { .mask = FCGI_EV_RX_STDERR, .name = "rx_stderr", .desc = "receipt of FCGI STDERR record" }, +#define FCGI_EV_RX_ENDREQ (1ULL << 13) + { .mask = FCGI_EV_RX_ENDREQ, .name = "rx_end_req", .desc = "receipt of FCGI END_REQUEST record" }, + +#define FCGI_EV_TX_RECORD (1ULL << 14) + { .mask = FCGI_EV_TX_RECORD, .name = "tx_record", .desc = "transmission of any FCGI record" }, +#define FCGI_EV_TX_EOI (1ULL << 15) + { .mask = FCGI_EV_TX_EOI, .name = "tx_eoi", .desc = "transmission of FCGI end of input" }, +#define FCGI_EV_TX_BEGREQ (1ULL << 16) + { .mask = FCGI_EV_TX_BEGREQ, .name = "tx_begin_request", .desc = "transmission of FCGI BEGIN_REQUEST record" }, +#define FCGI_EV_TX_GETVAL (1ULL << 17) + { .mask = FCGI_EV_TX_GETVAL, .name = "tx_get_values", .desc = "transmission of FCGI GET_VALUES record" }, +#define FCGI_EV_TX_PARAMS (1ULL << 18) + { .mask = FCGI_EV_TX_PARAMS, .name = "tx_params", .desc = "transmission of FCGI PARAMS record" }, +#define FCGI_EV_TX_STDIN (1ULL << 19) + { .mask = FCGI_EV_TX_STDIN, .name = "tx_stding", .desc = "transmission of FCGI STDIN record" }, +#define FCGI_EV_TX_ABORT (1ULL << 20) + { .mask = FCGI_EV_TX_ABORT, .name = "tx_abort", .desc = "transmission of FCGI ABORT record" }, + +#define FCGI_EV_RSP_DATA (1ULL << 21) + { .mask = FCGI_EV_RSP_DATA, .name = "rsp_data", .desc = "parse any data of H1 response" }, +#define FCGI_EV_RSP_EOM (1ULL << 22) + { .mask = FCGI_EV_RSP_EOM, .name = "rsp_eom", .desc = "reach the end of message of H1 response" }, +#define FCGI_EV_RSP_HDRS (1ULL << 23) + { .mask = FCGI_EV_RSP_HDRS, .name = "rsp_headers", .desc = "parse headers of H1 response" }, +#define FCGI_EV_RSP_BODY (1ULL << 24) + { .mask = FCGI_EV_RSP_BODY, .name = "rsp_body", .desc = "parse body part of H1 response" }, +#define FCGI_EV_RSP_TLRS (1ULL << 25) + { .mask = FCGI_EV_RSP_TLRS, .name = "rsp_trailerus", .desc = "parse trailers of H1 response" }, + +#define FCGI_EV_FSTRM_NEW (1ULL << 26) + { .mask = FCGI_EV_FSTRM_NEW, .name = "fstrm_new", .desc = "new FCGI stream" }, +#define FCGI_EV_FSTRM_BLK (1ULL << 27) + { .mask = FCGI_EV_FSTRM_BLK, .name = "fstrm_blk", .desc = "FCGI stream blocked" }, +#define FCGI_EV_FSTRM_END (1ULL << 28) + { .mask = FCGI_EV_FSTRM_END, .name = "fstrm_end", .desc = "FCGI stream terminated" }, +#define FCGI_EV_FSTRM_ERR (1ULL << 29) + { .mask = FCGI_EV_FSTRM_ERR, .name = "fstrm_err", .desc = "error on FCGI stream" }, + +#define FCGI_EV_STRM_NEW (1ULL << 30) + { .mask = FCGI_EV_STRM_NEW, .name = "strm_new", .desc = "app-layer stream creation" }, +#define FCGI_EV_STRM_RECV (1ULL << 31) + { .mask = FCGI_EV_STRM_RECV, .name = "strm_recv", .desc = "receiving data for stream" }, +#define FCGI_EV_STRM_SEND (1ULL << 32) + { .mask = FCGI_EV_STRM_SEND, .name = "strm_send", .desc = "sending data for stream" }, +#define FCGI_EV_STRM_FULL (1ULL << 33) + { .mask = FCGI_EV_STRM_FULL, .name = "strm_full", .desc = "stream buffer full" }, +#define FCGI_EV_STRM_WAKE (1ULL << 34) + { .mask = FCGI_EV_STRM_WAKE, .name = "strm_wake", .desc = "stream woken up" }, +#define FCGI_EV_STRM_SHUT (1ULL << 35) + { .mask = FCGI_EV_STRM_SHUT, .name = "strm_shut", .desc = "stream shutdown" }, +#define FCGI_EV_STRM_END (1ULL << 36) + { .mask = FCGI_EV_STRM_END, .name = "strm_end", .desc = "detaching app-layer stream" }, +#define FCGI_EV_STRM_ERR (1ULL << 37) + { .mask = FCGI_EV_STRM_ERR, .name = "strm_err", .desc = "stream error" }, + + { } +}; + +static const struct name_desc fcgi_trace_lockon_args[4] = { + /* arg1 */ { /* already used by the connection */ }, + /* arg2 */ { .name="fstrm", .desc="FCGI stream" }, + /* arg3 */ { }, + /* arg4 */ { } +}; + + +static const struct name_desc fcgi_trace_decoding[] = { +#define FCGI_VERB_CLEAN 1 + { .name="clean", .desc="only user-friendly stuff, generally suitable for level \"user\"" }, +#define FCGI_VERB_MINIMAL 2 + { .name="minimal", .desc="report only fconn/fstrm state and flags, no real decoding" }, +#define FCGI_VERB_SIMPLE 3 + { .name="simple", .desc="add request/response status line or htx info when available" }, +#define FCGI_VERB_ADVANCED 4 + { .name="advanced", .desc="add header fields or record decoding when available" }, +#define FCGI_VERB_COMPLETE 5 + { .name="complete", .desc="add full data dump when available" }, + { /* end */ } +}; + +static struct trace_source trace_fcgi __read_mostly = { + .name = IST("fcgi"), + .desc = "FastCGI multiplexer", + .arg_def = TRC_ARG1_CONN, // TRACE()'s first argument is always a connection + .default_cb = fcgi_trace, + .known_events = fcgi_trace_events, + .lockon_args = fcgi_trace_lockon_args, + .decoding = fcgi_trace_decoding, + .report_events = ~0, // report everything by default +}; + +#define TRACE_SOURCE &trace_fcgi +INITCALL1(STG_REGISTER, trace_register_source, TRACE_SOURCE); + +/* FCGI connection and stream pools */ +DECLARE_STATIC_POOL(pool_head_fcgi_conn, "fcgi_conn", sizeof(struct fcgi_conn)); +DECLARE_STATIC_POOL(pool_head_fcgi_strm, "fcgi_strm", sizeof(struct fcgi_strm)); + +struct task *fcgi_timeout_task(struct task *t, void *context, unsigned int state); +static int fcgi_process(struct fcgi_conn *fconn); +/* fcgi_io_cb is exported to see it resolved in "show fd" */ +struct task *fcgi_io_cb(struct task *t, void *ctx, unsigned int state); +static inline struct fcgi_strm *fcgi_conn_st_by_id(struct fcgi_conn *fconn, int id); +struct task *fcgi_deferred_shut(struct task *t, void *ctx, unsigned int state); +static struct fcgi_strm *fcgi_stconn_new(struct fcgi_conn *fconn, struct stconn *sc, struct session *sess); +static void fcgi_strm_notify_recv(struct fcgi_strm *fstrm); +static void fcgi_strm_notify_send(struct fcgi_strm *fstrm); +static void fcgi_strm_alert(struct fcgi_strm *fstrm); +static int fcgi_strm_send_abort(struct fcgi_conn *fconn, struct fcgi_strm *fstrm); + +/* a dummy closed endpoint */ +static const struct sedesc closed_ep = { + .sc = NULL, + .flags = SE_FL_DETACHED, +}; + +/* a dmumy management stream */ +static const struct fcgi_strm *fcgi_mgmt_stream = &(const struct fcgi_strm){ + .sd = (struct sedesc*)&closed_ep, + .fconn = NULL, + .state = FCGI_SS_CLOSED, + .flags = FCGI_SF_NONE, + .id = 0, +}; + +/* and a dummy idle stream for use with any unknown stream */ +static const struct fcgi_strm *fcgi_unknown_stream = &(const struct fcgi_strm){ + .sd = (struct sedesc*)&closed_ep, + .fconn = NULL, + .state = FCGI_SS_IDLE, + .flags = FCGI_SF_NONE, + .id = 0, +}; + +/* returns the stconn associated to the FCGI stream */ +static forceinline struct stconn *fcgi_strm_sc(const struct fcgi_strm *fstrm) +{ + return fstrm->sd->sc; +} + + +/* the FCGI traces always expect that arg1, if non-null, is of type connection + * (from which we can derive fconn), that arg2, if non-null, is of type fstrm, + * and that arg3, if non-null, is a htx for rx/tx headers. + */ +static void fcgi_trace(enum trace_level level, uint64_t mask, const struct trace_source *src, + const struct ist where, const struct ist func, + const void *a1, const void *a2, const void *a3, const void *a4) +{ + const struct connection *conn = a1; + struct fcgi_conn *fconn = conn ? conn->ctx : NULL; + const struct fcgi_strm *fstrm = a2; + const struct htx *htx = a3; + const size_t *val = a4; + + if (!fconn) + fconn = (fstrm ? fstrm->fconn : NULL); + + if (!fconn || src->verbosity < FCGI_VERB_CLEAN) + return; + + /* Display the response state if fstrm is defined */ + if (fstrm) + chunk_appendf(&trace_buf, " [rsp:%s]", h1m_state_str(fstrm->h1m.state)); + + if (src->verbosity == FCGI_VERB_CLEAN) + return; + + /* Display the value to the 4th argument (level > STATE) */ + if (src->level > TRACE_LEVEL_STATE && val) + chunk_appendf(&trace_buf, " - VAL=%lu", (long)*val); + + /* Display status-line if possible (verbosity > MINIMAL) */ + if (src->verbosity > FCGI_VERB_MINIMAL && htx && htx_nbblks(htx)) { + const struct htx_blk *blk = __htx_get_head_blk(htx); + const struct htx_sl *sl = htx_get_blk_ptr(htx, blk); + enum htx_blk_type type = htx_get_blk_type(blk); + + if (type == HTX_BLK_REQ_SL || type == HTX_BLK_RES_SL) + chunk_appendf(&trace_buf, " - \"%.*s %.*s %.*s\"", + HTX_SL_P1_LEN(sl), HTX_SL_P1_PTR(sl), + HTX_SL_P2_LEN(sl), HTX_SL_P2_PTR(sl), + HTX_SL_P3_LEN(sl), HTX_SL_P3_PTR(sl)); + } + + /* Display fconn info and, if defined, fstrm info */ + chunk_appendf(&trace_buf, " - fconn=%p(%s,0x%08x)", fconn, fconn_st_to_str(fconn->state), fconn->flags); + if (fstrm) + chunk_appendf(&trace_buf, " fstrm=%p(%d,%s,0x%08x)", fstrm, fstrm->id, fstrm_st_to_str(fstrm->state), fstrm->flags); + + if (!fstrm || fstrm->id <= 0) + chunk_appendf(&trace_buf, " dsi=%d", fconn->dsi); + if (fconn->dsi >= 0 && (mask & FCGI_EV_RX_FHDR)) + chunk_appendf(&trace_buf, " drt=%s", fcgi_rt_str(fconn->drt)); + + if (src->verbosity == FCGI_VERB_MINIMAL) + return; + + /* Display mbuf and dbuf info (level > USER & verbosity > SIMPLE) */ + if (src->level > TRACE_LEVEL_USER) { + if (src->verbosity == FCGI_VERB_COMPLETE || + (src->verbosity == FCGI_VERB_ADVANCED && (mask & (FCGI_EV_FCONN_RECV|FCGI_EV_RX_RECORD)))) + chunk_appendf(&trace_buf, " dbuf=%u@%p+%u/%u", + (unsigned int)b_data(&fconn->dbuf), b_orig(&fconn->dbuf), + (unsigned int)b_head_ofs(&fconn->dbuf), (unsigned int)b_size(&fconn->dbuf)); + if (src->verbosity == FCGI_VERB_COMPLETE || + (src->verbosity == FCGI_VERB_ADVANCED && (mask & (FCGI_EV_FCONN_SEND|FCGI_EV_TX_RECORD)))) { + struct buffer *hmbuf = br_head(fconn->mbuf); + struct buffer *tmbuf = br_tail(fconn->mbuf); + + chunk_appendf(&trace_buf, " .mbuf=[%u..%u|%u],h=[%u@%p+%u/%u],t=[%u@%p+%u/%u]", + br_head_idx(fconn->mbuf), br_tail_idx(fconn->mbuf), br_size(fconn->mbuf), + (unsigned int)b_data(hmbuf), b_orig(hmbuf), + (unsigned int)b_head_ofs(hmbuf), (unsigned int)b_size(hmbuf), + (unsigned int)b_data(tmbuf), b_orig(tmbuf), + (unsigned int)b_head_ofs(tmbuf), (unsigned int)b_size(tmbuf)); + } + + if (fstrm && (src->verbosity == FCGI_VERB_COMPLETE || + (src->verbosity == FCGI_VERB_ADVANCED && (mask & (FCGI_EV_STRM_RECV|FCGI_EV_RSP_DATA))))) + chunk_appendf(&trace_buf, " rxbuf=%u@%p+%u/%u", + (unsigned int)b_data(&fstrm->rxbuf), b_orig(&fstrm->rxbuf), + (unsigned int)b_head_ofs(&fstrm->rxbuf), (unsigned int)b_size(&fstrm->rxbuf)); + } + + /* Display htx info if defined (level > USER) */ + if (src->level > TRACE_LEVEL_USER && htx) { + int full = 0; + + /* Full htx info (level > STATE && verbosity > SIMPLE) */ + if (src->level > TRACE_LEVEL_STATE) { + if (src->verbosity == FCGI_VERB_COMPLETE) + full = 1; + else if (src->verbosity == FCGI_VERB_ADVANCED && (mask & (FCGI_EV_RSP_HDRS|FCGI_EV_TX_PARAMS))) + full = 1; + } + + chunk_memcat(&trace_buf, "\n\t", 2); + htx_dump(&trace_buf, htx, full); + } +} + +/*****************************************************/ +/* functions below are for dynamic buffer management */ +/*****************************************************/ + +/* Indicates whether or not the we may call the fcgi_recv() function to attempt + * to receive data into the buffer and/or demux pending data. The condition is + * a bit complex due to some API limits for now. The rules are the following : + * - if an error or a shutdown was detected on the connection and the buffer + * is empty, we must not attempt to receive + * - if the demux buf failed to be allocated, we must not try to receive and + * we know there is nothing pending + * - if no flag indicates a blocking condition, we may attempt to receive, + * regardless of whether the demux buffer is full or not, so that only + * de demux part decides whether or not to block. This is needed because + * the connection API indeed prevents us from re-enabling receipt that is + * already enabled in a polled state, so we must always immediately stop + * as soon as the demux can't proceed so as never to hit an end of read + * with data pending in the buffers. + * - otherwise must may not attempt + */ +static inline int fcgi_recv_allowed(const struct fcgi_conn *fconn) +{ + if (fconn->flags & (FCGI_CF_EOS|FCGI_CF_ERROR)) + return 0; + + if (b_data(&fconn->dbuf) == 0 && fconn->state == FCGI_CS_CLOSED) + return 0; + + if (!(fconn->flags & FCGI_CF_DEM_DALLOC) && + !(fconn->flags & FCGI_CF_DEM_BLOCK_ANY)) + return 1; + + return 0; +} + +/* Restarts reading on the connection if it was not enabled */ +static inline void fcgi_conn_restart_reading(const struct fcgi_conn *fconn, int consider_buffer) +{ + if (!fcgi_recv_allowed(fconn)) + return; + if ((!consider_buffer || !b_data(&fconn->dbuf)) && + (fconn->wait_event.events & SUB_RETRY_RECV)) + return; + tasklet_wakeup(fconn->wait_event.tasklet); +} + + +/* Tries to grab a buffer and to re-enable processing on mux <target>. The + * fcgi_conn flags are used to figure what buffer was requested. It returns 1 if + * the allocation succeeds, in which case the connection is woken up, or 0 if + * it's impossible to wake up and we prefer to be woken up later. + */ +static int fcgi_buf_available(void *target) +{ + struct fcgi_conn *fconn = target; + struct fcgi_strm *fstrm; + + if ((fconn->flags & FCGI_CF_DEM_DALLOC) && b_alloc(&fconn->dbuf)) { + TRACE_STATE("unblocking fconn, dbuf allocated", FCGI_EV_FCONN_RECV|FCGI_EV_FCONN_BLK|FCGI_EV_FCONN_WAKE, fconn->conn); + fconn->flags &= ~FCGI_CF_DEM_DALLOC; + fcgi_conn_restart_reading(fconn, 1); + return 1; + } + + if ((fconn->flags & FCGI_CF_MUX_MALLOC) && b_alloc(br_tail(fconn->mbuf))) { + TRACE_STATE("unblocking fconn, mbuf allocated", FCGI_EV_FCONN_SEND|FCGI_EV_FCONN_BLK|FCGI_EV_FCONN_WAKE, fconn->conn); + fconn->flags &= ~FCGI_CF_MUX_MALLOC; + if (fconn->flags & FCGI_CF_DEM_MROOM) { + fconn->flags &= ~FCGI_CF_DEM_MROOM; + fcgi_conn_restart_reading(fconn, 1); + } + return 1; + } + + if ((fconn->flags & FCGI_CF_DEM_SALLOC) && + (fstrm = fcgi_conn_st_by_id(fconn, fconn->dsi)) && fcgi_strm_sc(fstrm) && + b_alloc(&fstrm->rxbuf)) { + TRACE_STATE("unblocking fstrm, rxbuf allocated", FCGI_EV_STRM_RECV|FCGI_EV_FSTRM_BLK|FCGI_EV_STRM_WAKE, fconn->conn, fstrm); + fconn->flags &= ~FCGI_CF_DEM_SALLOC; + fcgi_conn_restart_reading(fconn, 1); + fcgi_strm_notify_recv(fstrm); + return 1; + } + + return 0; +} + +static inline struct buffer *fcgi_get_buf(struct fcgi_conn *fconn, struct buffer *bptr) +{ + struct buffer *buf = NULL; + + if (likely(!LIST_INLIST(&fconn->buf_wait.list)) && + unlikely((buf = b_alloc(bptr)) == NULL)) { + fconn->buf_wait.target = fconn; + fconn->buf_wait.wakeup_cb = fcgi_buf_available; + LIST_APPEND(&th_ctx->buffer_wq, &fconn->buf_wait.list); + } + return buf; +} + +static inline void fcgi_release_buf(struct fcgi_conn *fconn, struct buffer *bptr) +{ + if (bptr->size) { + b_free(bptr); + offer_buffers(NULL, 1); + } +} + +static inline void fcgi_release_mbuf(struct fcgi_conn *fconn) +{ + struct buffer *buf; + unsigned int count = 0; + + while (b_size(buf = br_head_pick(fconn->mbuf))) { + b_free(buf); + count++; + } + if (count) + offer_buffers(NULL, count); +} + +/* Returns the number of allocatable outgoing streams for the connection taking + * the number reserved streams into account. + */ +static inline int fcgi_streams_left(const struct fcgi_conn *fconn) +{ + int ret; + + ret = (unsigned int)(0x7FFF - fconn->max_id) - fconn->nb_reserved - 1; + if (ret < 0) + ret = 0; + return ret; +} + +/* Returns the number of streams in use on a connection to figure if it's + * idle or not. We check nb_sc and not nb_streams as the caller will want + * to know if it was the last one after a detach(). + */ +static int fcgi_used_streams(struct connection *conn) +{ + struct fcgi_conn *fconn = conn->ctx; + + return fconn->nb_sc; +} + +/* Returns the number of concurrent streams available on the connection */ +static int fcgi_avail_streams(struct connection *conn) +{ + struct server *srv = objt_server(conn->target); + struct fcgi_conn *fconn = conn->ctx; + int ret1, ret2; + + /* Don't open new stream if the connection is closed */ + if (fconn->state == FCGI_CS_CLOSED) + return 0; + + /* May be negative if this setting has changed */ + ret1 = (fconn->streams_limit - fconn->nb_streams); + + /* we must also consider the limit imposed by stream IDs */ + ret2 = fcgi_streams_left(fconn); + ret1 = MIN(ret1, ret2); + if (ret1 > 0 && srv && srv->max_reuse >= 0) { + ret2 = ((fconn->stream_cnt <= srv->max_reuse) ? srv->max_reuse - fconn->stream_cnt + 1: 0); + ret1 = MIN(ret1, ret2); + } + return ret1; +} + +/*****************************************************************/ +/* functions below are dedicated to the mux setup and management */ +/*****************************************************************/ + +/* Initializes the mux once it's attached. Only outgoing connections are + * supported. So the context is already initialized before installing the + * mux. <input> is always used as Input buffer and may contain data. It is the + * caller responsibility to not reuse it anymore. Returns < 0 on error. + */ +static int fcgi_init(struct connection *conn, struct proxy *px, struct session *sess, + struct buffer *input) +{ + struct fcgi_conn *fconn; + struct fcgi_strm *fstrm; + struct fcgi_app *app = get_px_fcgi_app(px); + struct task *t = NULL; + void *conn_ctx = conn->ctx; + + TRACE_ENTER(FCGI_EV_FSTRM_NEW); + + if (!app) { + TRACE_ERROR("No FCGI app found, don't create fconn", FCGI_EV_FCONN_NEW|FCGI_EV_FCONN_END|FCGI_EV_FCONN_ERR); + goto fail_conn; + } + + fconn = pool_alloc(pool_head_fcgi_conn); + if (!fconn) { + TRACE_ERROR("fconn allocation failure", FCGI_EV_FCONN_NEW|FCGI_EV_FCONN_END|FCGI_EV_FCONN_ERR); + goto fail_conn; + } + + fconn->shut_timeout = fconn->timeout = px->timeout.server; + if (tick_isset(px->timeout.serverfin)) + fconn->shut_timeout = px->timeout.serverfin; + + fconn->flags = FCGI_CF_NONE; + + /* Retrieve useful info from the FCGI app */ + if (app->flags & FCGI_APP_FL_KEEP_CONN) + fconn->flags |= FCGI_CF_KEEP_CONN; + if (app->flags & FCGI_APP_FL_GET_VALUES) + fconn->flags |= FCGI_CF_GET_VALUES; + if (app->flags & FCGI_APP_FL_MPXS_CONNS) + fconn->flags |= FCGI_CF_MPXS_CONNS; + + fconn->proxy = px; + fconn->app = app; + fconn->task = NULL; + if (tick_isset(fconn->timeout)) { + t = task_new_here(); + if (!t) { + TRACE_ERROR("fconn task allocation failure", FCGI_EV_FCONN_NEW|FCGI_EV_FCONN_END|FCGI_EV_FCONN_ERR); + goto fail; + } + + fconn->task = t; + t->process = fcgi_timeout_task; + t->context = fconn; + t->expire = tick_add(now_ms, fconn->timeout); + } + + fconn->wait_event.tasklet = tasklet_new(); + if (!fconn->wait_event.tasklet) + goto fail; + fconn->wait_event.tasklet->process = fcgi_io_cb; + fconn->wait_event.tasklet->context = fconn; + fconn->wait_event.events = 0; + + /* Initialise the context. */ + fconn->state = FCGI_CS_INIT; + fconn->conn = conn; + fconn->streams_limit = app->maxreqs; + fconn->max_id = -1; + fconn->nb_streams = 0; + fconn->nb_sc = 0; + fconn->nb_reserved = 0; + fconn->stream_cnt = 0; + + fconn->dbuf = *input; + fconn->dsi = -1; + + br_init(fconn->mbuf, sizeof(fconn->mbuf) / sizeof(fconn->mbuf[0])); + fconn->streams_by_id = EB_ROOT; + LIST_INIT(&fconn->send_list); + LIST_INIT(&fconn->buf_wait.list); + + conn->ctx = fconn; + + if (t) + task_queue(t); + + /* FIXME: this is temporary, for outgoing connections we need to + * immediately allocate a stream until the code is modified so that the + * caller calls ->attach(). For now the outgoing sc is stored as + * conn->ctx by the caller and saved in conn_ctx. + */ + fstrm = fcgi_stconn_new(fconn, conn_ctx, sess); + if (!fstrm) + goto fail; + + + /* Repare to read something */ + fcgi_conn_restart_reading(fconn, 1); + TRACE_LEAVE(FCGI_EV_FCONN_NEW, conn); + return 0; + + fail: + task_destroy(t); + tasklet_free(fconn->wait_event.tasklet); + pool_free(pool_head_fcgi_conn, fconn); + fail_conn: + conn->ctx = conn_ctx; // restore saved ctx + TRACE_DEVEL("leaving in error", FCGI_EV_FCONN_NEW|FCGI_EV_FCONN_END|FCGI_EV_FCONN_ERR); + return -1; +} + +/* Returns the next allocatable outgoing stream ID for the FCGI connection, or + * -1 if no more is allocatable. + */ +static inline int32_t fcgi_conn_get_next_sid(const struct fcgi_conn *fconn) +{ + int32_t id = (fconn->max_id + 1) | 1; + + if ((id & 0x80000000U)) + id = -1; + return id; +} + +/* Returns the stream associated with id <id> or NULL if not found */ +static inline struct fcgi_strm *fcgi_conn_st_by_id(struct fcgi_conn *fconn, int id) +{ + struct eb32_node *node; + + if (id == 0) + return (struct fcgi_strm *)fcgi_mgmt_stream; + + if (id > fconn->max_id) + return (struct fcgi_strm *)fcgi_unknown_stream; + + node = eb32_lookup(&fconn->streams_by_id, id); + if (!node) + return (struct fcgi_strm *)fcgi_unknown_stream; + return container_of(node, struct fcgi_strm, by_id); +} + + +/* Release function. This one should be called to free all resources allocated + * to the mux. + */ +static void fcgi_release(struct fcgi_conn *fconn) +{ + struct connection *conn = fconn->conn; + + TRACE_POINT(FCGI_EV_FCONN_END); + + if (LIST_INLIST(&fconn->buf_wait.list)) + LIST_DEL_INIT(&fconn->buf_wait.list); + + fcgi_release_buf(fconn, &fconn->dbuf); + fcgi_release_mbuf(fconn); + + if (fconn->task) { + fconn->task->context = NULL; + task_wakeup(fconn->task, TASK_WOKEN_OTHER); + fconn->task = NULL; + } + tasklet_free(fconn->wait_event.tasklet); + if (conn && fconn->wait_event.events != 0) + conn->xprt->unsubscribe(conn, conn->xprt_ctx, fconn->wait_event.events, + &fconn->wait_event); + + pool_free(pool_head_fcgi_conn, fconn); + + if (conn) { + conn->mux = NULL; + conn->ctx = NULL; + TRACE_DEVEL("freeing conn", FCGI_EV_FCONN_END, conn); + + conn_stop_tracking(conn); + conn_full_close(conn); + if (conn->destroy_cb) + conn->destroy_cb(conn); + conn_free(conn); + } +} + +/* Detect a pending read0 for a FCGI connection. It happens if a read0 is + * pending on the connection AND if there is no more data in the demux + * buffer. The function returns 1 to report a read0 or 0 otherwise. + */ +static int fcgi_conn_read0_pending(struct fcgi_conn *fconn) +{ + if ((fconn->flags & FCGI_CF_EOS) && !b_data(&fconn->dbuf)) + return 1; + return 0; +} + + +/* Returns true if the FCGI connection must be release */ +static inline int fcgi_conn_is_dead(struct fcgi_conn *fconn) +{ + if (eb_is_empty(&fconn->streams_by_id) && /* don't close if streams exist */ + (!(fconn->flags & FCGI_CF_KEEP_CONN) || /* don't keep the connection alive */ + (fconn->flags & FCGI_CF_ERROR) || /* errors close immediately */ + (fconn->state == FCGI_CS_CLOSED && !fconn->task) ||/* a timeout stroke earlier */ + (!(fconn->conn->owner)) || /* Nobody's left to take care of the connection, drop it now */ + (!br_data(fconn->mbuf) && /* mux buffer empty, also process clean events below */ + (fconn->flags & FCGI_CF_EOS)))) + return 1; + return 0; +} + + +/********************************************************/ +/* functions below are for the FCGI protocol processing */ +/********************************************************/ + +/* Marks an error on the stream. */ +static inline void fcgi_strm_error(struct fcgi_strm *fstrm) +{ + if (fstrm->id && fstrm->state != FCGI_SS_ERROR) { + TRACE_POINT(FCGI_EV_FSTRM_ERR, fstrm->fconn->conn, fstrm); + if (fstrm->state < FCGI_SS_ERROR) { + fstrm->state = FCGI_SS_ERROR; + TRACE_STATE("switching to ERROR", FCGI_EV_FSTRM_ERR, fstrm->fconn->conn, fstrm); + } + se_fl_set_error(fstrm->sd); + } +} + +/* Attempts to notify the data layer of recv availability */ +static void fcgi_strm_notify_recv(struct fcgi_strm *fstrm) +{ + if (fstrm->subs && (fstrm->subs->events & SUB_RETRY_RECV)) { + TRACE_POINT(FCGI_EV_STRM_WAKE, fstrm->fconn->conn, fstrm); + tasklet_wakeup(fstrm->subs->tasklet); + fstrm->subs->events &= ~SUB_RETRY_RECV; + if (!fstrm->subs->events) + fstrm->subs = NULL; + } +} + +/* Attempts to notify the data layer of send availability */ +static void fcgi_strm_notify_send(struct fcgi_strm *fstrm) +{ + if (fstrm->subs && (fstrm->subs->events & SUB_RETRY_SEND)) { + TRACE_POINT(FCGI_EV_STRM_WAKE, fstrm->fconn->conn, fstrm); + fstrm->flags |= FCGI_SF_NOTIFIED; + tasklet_wakeup(fstrm->subs->tasklet); + fstrm->subs->events &= ~SUB_RETRY_SEND; + if (!fstrm->subs->events) + fstrm->subs = NULL; + } + else if (fstrm->flags & (FCGI_SF_WANT_SHUTR | FCGI_SF_WANT_SHUTW)) { + TRACE_POINT(FCGI_EV_STRM_WAKE, fstrm->fconn->conn, fstrm); + tasklet_wakeup(fstrm->shut_tl); + } +} + +/* Alerts the data layer, trying to wake it up by all means, following + * this sequence : + * - if the fcgi stream' data layer is subscribed to recv, then it's woken up + * for recv + * - if its subscribed to send, then it's woken up for send + * - if it was subscribed to neither, its ->wake() callback is called + * It is safe to call this function with a closed stream which doesn't have a + * stream connector anymore. + */ +static void fcgi_strm_alert(struct fcgi_strm *fstrm) +{ + TRACE_POINT(FCGI_EV_STRM_WAKE, fstrm->fconn->conn, fstrm); + if (fstrm->subs || + (fstrm->flags & (FCGI_SF_WANT_SHUTR|FCGI_SF_WANT_SHUTW))) { + fcgi_strm_notify_recv(fstrm); + fcgi_strm_notify_send(fstrm); + } + else if (fcgi_strm_sc(fstrm) && fcgi_strm_sc(fstrm)->app_ops->wake != NULL) { + TRACE_POINT(FCGI_EV_STRM_WAKE, fstrm->fconn->conn, fstrm); + fcgi_strm_sc(fstrm)->app_ops->wake(fcgi_strm_sc(fstrm)); + } +} + +/* Writes the 16-bit record size <len> at address <record> */ +static inline void fcgi_set_record_size(void *record, uint16_t len) +{ + uint8_t *out = (record + 4); + + *out = (len >> 8); + *(out + 1) = (len & 0xff); +} + +/* Writes the 16-bit stream id <id> at address <record> */ +static inline void fcgi_set_record_id(void *record, uint16_t id) +{ + uint8_t *out = (record + 2); + + *out = (id >> 8); + *(out + 1) = (id & 0xff); +} + +/* Marks a FCGI stream as CLOSED and decrement the number of active streams for + * its connection if the stream was not yet closed. Please use this exclusively + * before closing a stream to ensure stream count is well maintained. + */ +static inline void fcgi_strm_close(struct fcgi_strm *fstrm) +{ + if (fstrm->state != FCGI_SS_CLOSED) { + TRACE_ENTER(FCGI_EV_FSTRM_END, fstrm->fconn->conn, fstrm); + fstrm->fconn->nb_streams--; + if (!fstrm->id) + fstrm->fconn->nb_reserved--; + if (fcgi_strm_sc(fstrm)) { + if (!se_fl_test(fstrm->sd, SE_FL_EOS) && !b_data(&fstrm->rxbuf)) + fcgi_strm_notify_recv(fstrm); + } + fstrm->state = FCGI_SS_CLOSED; + TRACE_STATE("switching to CLOSED", FCGI_EV_FSTRM_END, fstrm->fconn->conn, fstrm); + TRACE_LEAVE(FCGI_EV_FSTRM_END, fstrm->fconn->conn, fstrm); + } +} + +/* Detaches a FCGI stream from its FCGI connection and releases it to the + * fcgi_strm pool. + */ +static void fcgi_strm_destroy(struct fcgi_strm *fstrm) +{ + struct connection *conn = fstrm->fconn->conn; + + TRACE_ENTER(FCGI_EV_FSTRM_END, conn, fstrm); + + fcgi_strm_close(fstrm); + eb32_delete(&fstrm->by_id); + if (b_size(&fstrm->rxbuf)) { + b_free(&fstrm->rxbuf); + offer_buffers(NULL, 1); + } + if (fstrm->subs) + fstrm->subs->events = 0; + /* There's no need to explicitly call unsubscribe here, the only + * reference left would be in the fconn send_list/fctl_list, and if + * we're in it, we're getting out anyway + */ + LIST_DEL_INIT(&fstrm->send_list); + tasklet_free(fstrm->shut_tl); + BUG_ON(fstrm->sd && !se_fl_test(fstrm->sd, SE_FL_ORPHAN)); + sedesc_free(fstrm->sd); + pool_free(pool_head_fcgi_strm, fstrm); + + TRACE_LEAVE(FCGI_EV_FSTRM_END, conn); +} + +/* Allocates a new stream <id> for connection <fconn> and adds it into fconn's + * stream tree. In case of error, nothing is added and NULL is returned. The + * causes of errors can be any failed memory allocation. The caller is + * responsible for checking if the connection may support an extra stream prior + * to calling this function. + */ +static struct fcgi_strm *fcgi_strm_new(struct fcgi_conn *fconn, int id) +{ + struct fcgi_strm *fstrm; + + TRACE_ENTER(FCGI_EV_FSTRM_NEW, fconn->conn); + + fstrm = pool_alloc(pool_head_fcgi_strm); + if (!fstrm) { + TRACE_ERROR("fstrm allocation failure", FCGI_EV_FSTRM_NEW|FCGI_EV_FSTRM_ERR|FCGI_EV_FSTRM_END, fconn->conn); + goto out; + } + + fstrm->shut_tl = tasklet_new(); + if (!fstrm->shut_tl) { + TRACE_ERROR("fstrm shut tasklet allocation failure", FCGI_EV_FSTRM_NEW|FCGI_EV_FSTRM_ERR|FCGI_EV_FSTRM_END, fconn->conn); + pool_free(pool_head_fcgi_strm, fstrm); + goto out; + } + fstrm->subs = NULL; + fstrm->shut_tl->process = fcgi_deferred_shut; + fstrm->shut_tl->context = fstrm; + LIST_INIT(&fstrm->send_list); + fstrm->fconn = fconn; + fstrm->sd = NULL; + fstrm->flags = FCGI_SF_NONE; + fstrm->proto_status = 0; + fstrm->state = FCGI_SS_IDLE; + fstrm->rxbuf = BUF_NULL; + + h1m_init_res(&fstrm->h1m); + fstrm->h1m.err_pos = -1; // don't care about errors on the request path + fstrm->h1m.flags |= (H1_MF_NO_PHDR|H1_MF_CLEAN_CONN_HDR); + + fstrm->by_id.key = fstrm->id = id; + if (id > 0) + fconn->max_id = id; + else + fconn->nb_reserved++; + + eb32_insert(&fconn->streams_by_id, &fstrm->by_id); + fconn->nb_streams++; + fconn->stream_cnt++; + + TRACE_LEAVE(FCGI_EV_FSTRM_NEW, fconn->conn, fstrm); + return fstrm; + + out: + TRACE_DEVEL("leaving in error", FCGI_EV_FSTRM_NEW|FCGI_EV_FSTRM_ERR|FCGI_EV_FSTRM_END, fconn->conn); + return NULL; +} + +/* Allocates a new stream associated to stream connector <sc> on the FCGI connection + * <fconn> and returns it, or NULL in case of memory allocation error or if the + * highest possible stream ID was reached. + */ +static struct fcgi_strm *fcgi_stconn_new(struct fcgi_conn *fconn, struct stconn *sc, + struct session *sess) +{ + struct fcgi_strm *fstrm = NULL; + + TRACE_ENTER(FCGI_EV_FSTRM_NEW, fconn->conn); + if (fconn->nb_streams >= fconn->streams_limit) { + TRACE_ERROR("streams_limit reached", FCGI_EV_FSTRM_NEW|FCGI_EV_FSTRM_END|FCGI_EV_FSTRM_ERR, fconn->conn); + goto out; + } + + if (fcgi_streams_left(fconn) < 1) { + TRACE_ERROR("!streams_left", FCGI_EV_FSTRM_NEW|FCGI_EV_FSTRM_END|FCGI_EV_FSTRM_ERR, fconn->conn); + goto out; + } + + /* Defer choosing the ID until we send the first message to create the stream */ + fstrm = fcgi_strm_new(fconn, 0); + if (!fstrm) { + TRACE_ERROR("fstream allocation failure", FCGI_EV_FSTRM_NEW|FCGI_EV_FSTRM_END|FCGI_EV_FSTRM_ERR, fconn->conn); + goto out; + } + if (sc_attach_mux(sc, fstrm, fconn->conn) < 0) + goto out; + fstrm->sd = sc->sedesc; + fstrm->sess = sess; + fconn->nb_sc++; + + TRACE_LEAVE(FCGI_EV_FSTRM_NEW, fconn->conn, fstrm); + return fstrm; + + out: + TRACE_DEVEL("leaving on error", FCGI_EV_FSTRM_NEW|FCGI_EV_FSTRM_END|FCGI_EV_FSTRM_ERR, fconn->conn); + fcgi_strm_destroy(fstrm); + return NULL; +} + +/* Wakes a specific stream and assign its stream connector some SE_FL_* flags among + * SE_FL_ERR_PENDING and SE_FL_ERROR if needed. The stream's state is + * automatically updated accordingly. If the stream is orphaned, it is + * destroyed. + */ +static void fcgi_strm_wake_one_stream(struct fcgi_strm *fstrm) +{ + struct fcgi_conn *fconn = fstrm->fconn; + + TRACE_ENTER(FCGI_EV_STRM_WAKE, fconn->conn, fstrm); + + if (!fcgi_strm_sc(fstrm)) { + /* this stream was already orphaned */ + fcgi_strm_destroy(fstrm); + TRACE_DEVEL("leaving with no fstrm", FCGI_EV_STRM_WAKE, fconn->conn); + return; + } + + if (fcgi_conn_read0_pending(fconn)) { + if (fstrm->state == FCGI_SS_OPEN) { + fstrm->state = FCGI_SS_HREM; + TRACE_STATE("switching to HREM", FCGI_EV_STRM_WAKE|FCGI_EV_FSTRM_END, fconn->conn, fstrm); + } + else if (fstrm->state == FCGI_SS_HLOC) + fcgi_strm_close(fstrm); + } + + if (fconn->state == FCGI_CS_CLOSED || (fconn->flags & (FCGI_CF_ERR_PENDING|FCGI_CF_ERROR))) { + se_fl_set_error(fstrm->sd); + + if (fstrm->state < FCGI_SS_ERROR) { + fstrm->state = FCGI_SS_ERROR; + TRACE_STATE("switching to ERROR", FCGI_EV_STRM_WAKE|FCGI_EV_FSTRM_END, fconn->conn, fstrm); + } + } + + fcgi_strm_alert(fstrm); + + TRACE_LEAVE(FCGI_EV_STRM_WAKE, fconn->conn, fstrm); +} + +/* Wakes unassigned streams (ID == 0) attached to the connection. */ +static void fcgi_wake_unassigned_streams(struct fcgi_conn *fconn) +{ + struct eb32_node *node; + struct fcgi_strm *fstrm; + + node = eb32_lookup(&fconn->streams_by_id, 0); + while (node) { + fstrm = container_of(node, struct fcgi_strm, by_id); + if (fstrm->id > 0) + break; + node = eb32_next(node); + fcgi_strm_wake_one_stream(fstrm); + } +} + +/* Wakes the streams attached to the connection, whose id is greater than <last> + * or unassigned. + */ +static void fcgi_wake_some_streams(struct fcgi_conn *fconn, int last) +{ + struct eb32_node *node; + struct fcgi_strm *fstrm; + + TRACE_ENTER(FCGI_EV_STRM_WAKE, fconn->conn); + + /* Wake all streams with ID > last */ + node = eb32_lookup_ge(&fconn->streams_by_id, last + 1); + while (node) { + fstrm = container_of(node, struct fcgi_strm, by_id); + node = eb32_next(node); + fcgi_strm_wake_one_stream(fstrm); + } + fcgi_wake_unassigned_streams(fconn); + + TRACE_LEAVE(FCGI_EV_STRM_WAKE, fconn->conn); +} + +static int fcgi_set_default_param(struct fcgi_conn *fconn, struct fcgi_strm *fstrm, + struct htx *htx, struct htx_sl *sl, + struct fcgi_strm_params *params) +{ + struct connection *cli_conn = objt_conn(fstrm->sess->origin); + const struct sockaddr_storage *src = (sc_check(fcgi_strm_sc(fstrm)) ? conn_src(fconn->conn) : sc_src(sc_opposite(fcgi_strm_sc(fstrm)))); + const struct sockaddr_storage *dst = (sc_check(fcgi_strm_sc(fstrm)) ? conn_dst(fconn->conn) : sc_dst(sc_opposite(fcgi_strm_sc(fstrm)))); + struct ist p; + + if (!sl) + goto error; + + if (!(params->mask & FCGI_SP_DOC_ROOT)) + params->docroot = fconn->app->docroot; + + if (!(params->mask & FCGI_SP_REQ_METH)) { + p = htx_sl_req_meth(sl); + params->meth = ist2(b_tail(params->p), p.len); + chunk_istcat(params->p, p); + } + if (!(params->mask & FCGI_SP_REQ_URI)) { + p = h1_get_uri(sl); + params->uri = ist2(b_tail(params->p), p.len); + chunk_istcat(params->p, p); + } + if (!(params->mask & FCGI_SP_SRV_PROTO)) { + p = htx_sl_req_vsn(sl); + params->vsn = ist2(b_tail(params->p), p.len); + chunk_istcat(params->p, p); + } + if (!(params->mask & FCGI_SP_SRV_PORT)) { + char *end; + int port = 0; + if (dst) + port = get_host_port(dst); + end = ultoa_o(port, b_tail(params->p), b_room(params->p)); + if (!end) + goto error; + params->srv_port = ist2(b_tail(params->p), end - b_tail(params->p)); + params->p->data += params->srv_port.len; + } + if (!(params->mask & FCGI_SP_SRV_NAME)) { + /* If no Host header found, use the server address to fill + * srv_name */ + if (!istlen(params->srv_name)) { + char *ptr = NULL; + + if (dst) + if (addr_to_str(dst, b_tail(params->p), b_room(params->p)) != -1) + ptr = b_tail(params->p); + if (ptr) { + params->srv_name = ist(ptr); + params->p->data += params->srv_name.len; + } + } + } + if (!(params->mask & FCGI_SP_REM_ADDR)) { + char *ptr = NULL; + + if (src) + if (addr_to_str(src, b_tail(params->p), b_room(params->p)) != -1) + ptr = b_tail(params->p); + if (ptr) { + params->rem_addr = ist(ptr); + params->p->data += params->rem_addr.len; + } + } + if (!(params->mask & FCGI_SP_REM_PORT)) { + char *end; + int port = 0; + if (src) + port = get_host_port(src); + end = ultoa_o(port, b_tail(params->p), b_room(params->p)); + if (!end) + goto error; + params->rem_port = ist2(b_tail(params->p), end - b_tail(params->p)); + params->p->data += params->rem_port.len; + } + if (!(params->mask & FCGI_SP_CONT_LEN)) { + struct htx_blk *blk; + enum htx_blk_type type; + char *end; + size_t len = 0; + + for (blk = htx_get_head_blk(htx); blk; blk = htx_get_next_blk(htx, blk)) { + type = htx_get_blk_type(blk); + + if (type == HTX_BLK_TLR || type == HTX_BLK_EOT) + break; + if (type == HTX_BLK_DATA) + len += htx_get_blksz(blk); + } + end = ultoa_o(len, b_tail(params->p), b_room(params->p)); + if (!end) + goto error; + params->cont_len = ist2(b_tail(params->p), end - b_tail(params->p)); + params->p->data += params->cont_len.len; + } + + if (!(params->mask & FCGI_SP_HTTPS)) { + if (cli_conn) + params->https = conn_is_ssl(cli_conn); + } + + if ((params->mask & FCGI_SP_URI_MASK) != FCGI_SP_URI_MASK) { + /* one of scriptname, pathinfo or query_string is no set */ + struct http_uri_parser parser = http_uri_parser_init(params->uri); + struct ist path = http_parse_path(&parser); + int len; + + /* No scrit_name set but no valid path ==> error */ + if (!(params->mask & FCGI_SP_SCRIPT_NAME) && !istlen(path)) + goto error; + + /* If there is a query-string, Set it if not already set */ + if (!(params->mask & FCGI_SP_REQ_QS)) { + struct ist qs = istfind(path, '?'); + + /* Update the path length */ + path.len -= qs.len; + + /* Set the query-string skipping the '?', if any */ + if (istlen(qs)) + params->qs = istnext(qs); + } + + /* If the script_name is set, don't try to deduce the path_info + * too. The opposite is not true. + */ + if (params->mask & FCGI_SP_SCRIPT_NAME) { + params->mask |= FCGI_SP_PATH_INFO; + goto end; + } + + /* Decode the path. it must first be copied to keep the URI + * untouched. + */ + chunk_istcat(params->p, path); + path.ptr = b_tail(params->p) - path.len; + len = url_decode(ist0(path), 0); + if (len < 0) + goto error; + path.len = len; + + /* script_name not set, preset it with the path for now */ + params->scriptname = path; + + /* If there is no regex to match the pathinfo, just to the last + * part and see if the index must be used. + */ + if (!fconn->app->pathinfo_re) + goto check_index; + + /* If some special characters are found in the decoded path (\n + * or \0), the PATH_INFO regex cannot match. This is theoretically + * valid, but probably unexpected, to have such characters. So, + * to avoid any surprises, an error is triggered in this + * case. + */ + if (istchr(path, '\n') || istchr(path, '\0')) + goto error; + + /* The regex does not match, just to the last part and see if + * the index must be used. + */ + if (!regex_exec_match2(fconn->app->pathinfo_re, path.ptr, len, MAX_MATCH, pmatch, 0)) + goto check_index; + + /* We must have at least 1 capture for the script name, + * otherwise we do nothing and jump to the last part. + */ + if (pmatch[1].rm_so == -1 || pmatch[1].rm_eo == -1) + goto check_index; + + /* Finally we can set the script_name and the path_info. The + * path_info is set if not already defined, and if it was + * captured + */ + params->scriptname = ist2(path.ptr + pmatch[1].rm_so, pmatch[1].rm_eo - pmatch[1].rm_so); + if (!(params->mask & FCGI_SP_PATH_INFO) && !(pmatch[2].rm_so == -1 || pmatch[2].rm_eo == -1)) + params->pathinfo = ist2(path.ptr + pmatch[2].rm_so, pmatch[2].rm_eo - pmatch[2].rm_so); + + check_index: + len = params->scriptname.len; + /* the script_name if finished by a '/' so we can add the index + * part, if any. + */ + if (istlen(fconn->app->index) && params->scriptname.ptr[len-1] == '/') { + struct ist sn = params->scriptname; + + params->scriptname = ist2(b_tail(params->p), len+fconn->app->index.len); + chunk_istcat(params->p, sn); + chunk_istcat(params->p, fconn->app->index); + } + } + + if (!(params->mask & FCGI_SP_SRV_SOFT)) { + params->srv_soft = ist2(b_tail(params->p), 0); + chunk_appendf(params->p, "HAProxy %s", haproxy_version); + params->srv_soft.len = b_tail(params->p) - params->srv_soft.ptr; + } + + end: + return 1; + error: + return 0; +} + +static int fcgi_encode_default_param(struct fcgi_conn *fconn, struct fcgi_strm *fstrm, + struct fcgi_strm_params *params, struct buffer *outbuf, int flag) +{ + struct fcgi_param p; + + if (params->mask & flag) + return 1; + + chunk_reset(&trash); + + switch (flag) { + case FCGI_SP_CGI_GATEWAY: + p.n = ist("GATEWAY_INTERFACE"); + p.v = ist("CGI/1.1"); + goto encode; + case FCGI_SP_DOC_ROOT: + p.n = ist("DOCUMENT_ROOT"); + p.v = params->docroot; + goto encode; + case FCGI_SP_SCRIPT_NAME: + p.n = ist("SCRIPT_NAME"); + p.v = params->scriptname; + goto encode; + case FCGI_SP_PATH_INFO: + p.n = ist("PATH_INFO"); + p.v = params->pathinfo; + goto encode; + case FCGI_SP_REQ_URI: + p.n = ist("REQUEST_URI"); + p.v = params->uri; + goto encode; + case FCGI_SP_REQ_METH: + p.n = ist("REQUEST_METHOD"); + p.v = params->meth; + goto encode; + case FCGI_SP_REQ_QS: + p.n = ist("QUERY_STRING"); + p.v = params->qs; + goto encode; + case FCGI_SP_SRV_NAME: + p.n = ist("SERVER_NAME"); + p.v = params->srv_name; + goto encode; + case FCGI_SP_SRV_PORT: + p.n = ist("SERVER_PORT"); + p.v = params->srv_port; + goto encode; + case FCGI_SP_SRV_PROTO: + p.n = ist("SERVER_PROTOCOL"); + p.v = params->vsn; + goto encode; + case FCGI_SP_REM_ADDR: + p.n = ist("REMOTE_ADDR"); + p.v = params->rem_addr; + goto encode; + case FCGI_SP_REM_PORT: + p.n = ist("REMOTE_PORT"); + p.v = params->rem_port; + goto encode; + case FCGI_SP_SCRIPT_FILE: + p.n = ist("SCRIPT_FILENAME"); + chunk_istcat(&trash, params->docroot); + chunk_istcat(&trash, params->scriptname); + p.v = ist2(b_head(&trash), b_data(&trash)); + goto encode; + case FCGI_SP_PATH_TRANS: + if (!istlen(params->pathinfo)) + goto skip; + p.n = ist("PATH_TRANSLATED"); + chunk_istcat(&trash, params->docroot); + chunk_istcat(&trash, params->pathinfo); + p.v = ist2(b_head(&trash), b_data(&trash)); + goto encode; + case FCGI_SP_CONT_LEN: + p.n = ist("CONTENT_LENGTH"); + p.v = params->cont_len; + goto encode; + case FCGI_SP_HTTPS: + if (!params->https) + goto skip; + p.n = ist("HTTPS"); + p.v = ist("on"); + goto encode; + case FCGI_SP_SRV_SOFT: + p.n = ist("SERVER_SOFTWARE"); + p.v = params->srv_soft; + goto encode; + default: + goto skip; + } + + encode: + if (!istlen(p.v)) + goto skip; + if (!fcgi_encode_param(outbuf, &p)) + return 0; + skip: + params->mask |= flag; + return 1; +} + +/* Sends a GET_VALUES record. Returns > 0 on success, 0 if it couldn't do + * anything. It is highly unexpected, but if the record is larger than a buffer + * and cannot be encoded in one time, an error is triggered and the connection is + * closed. GET_VALUES record cannot be split. + */ +static int fcgi_conn_send_get_values(struct fcgi_conn *fconn) +{ + struct buffer outbuf; + struct buffer *mbuf; + struct fcgi_param max_reqs = { .n = ist("FCGI_MAX_REQS"), .v = ist("")}; + struct fcgi_param mpxs_conns = { .n = ist("FCGI_MPXS_CONNS"), .v = ist("")}; + int ret = 0; + + TRACE_ENTER(FCGI_EV_TX_RECORD|FCGI_EV_TX_GETVAL, fconn->conn); + + mbuf = br_tail(fconn->mbuf); + retry: + if (!fcgi_get_buf(fconn, mbuf)) { + fconn->flags |= FCGI_CF_MUX_MALLOC; + fconn->flags |= FCGI_CF_DEM_MROOM; + TRACE_STATE("waiting for fconn mbuf ring allocation", FCGI_EV_TX_RECORD|FCGI_EV_FCONN_BLK, fconn->conn); + ret = 0; + goto end; + } + + while (1) { + outbuf = b_make(b_tail(mbuf), b_contig_space(mbuf), 0, 0); + if (outbuf.size >= FCGI_RECORD_HEADER_SZ || !b_space_wraps(mbuf)) + break; + realign_again: + b_slow_realign(mbuf, trash.area, b_data(mbuf)); + } + + if (outbuf.size < FCGI_RECORD_HEADER_SZ) + goto full; + + /* vsn: 1(FCGI_VERSION), type: (9)FCGI_GET_VALUES, id: 0x0000, + * len: 0x0000 (fill later), padding: 0x00, rsv: 0x00 */ + memcpy(outbuf.area, "\x01\x09\x00\x00\x00\x00\x00\x00", FCGI_RECORD_HEADER_SZ); + outbuf.data = FCGI_RECORD_HEADER_SZ; + + /* Note: Don't send the param FCGI_MAX_CONNS because its value cannot be + * handled by HAProxy. + */ + if (!fcgi_encode_param(&outbuf, &max_reqs) || !fcgi_encode_param(&outbuf, &mpxs_conns)) + goto full; + + /* update the record's size now */ + TRACE_PROTO("FCGI GET_VALUES record xferred", FCGI_EV_TX_RECORD|FCGI_EV_TX_GETVAL, fconn->conn, 0, 0, (size_t[]){outbuf.data-8}); + fcgi_set_record_size(outbuf.area, outbuf.data - FCGI_RECORD_HEADER_SZ); + b_add(mbuf, outbuf.data); + ret = 1; + + end: + TRACE_LEAVE(FCGI_EV_TX_RECORD|FCGI_EV_TX_GETVAL, fconn->conn); + return ret; + full: + /* Too large to be encoded. For GET_VALUES records, it is an error */ + if (!b_data(mbuf)) { + TRACE_ERROR("GET_VALUES record too large", FCGI_EV_TX_RECORD|FCGI_EV_TX_GETVAL|FCGI_EV_FCONN_ERR, fconn->conn); + goto fail; + } + + if ((mbuf = br_tail_add(fconn->mbuf)) != NULL) + goto retry; + fconn->flags |= FCGI_CF_MUX_MFULL; + fconn->flags |= FCGI_CF_DEM_MROOM; + TRACE_STATE("mbuf ring full", FCGI_EV_TX_RECORD|FCGI_EV_FCONN_BLK, fconn->conn); + ret = 0; + goto end; + fail: + fconn->state = FCGI_CS_CLOSED; + TRACE_STATE("switching to CLOSED", FCGI_EV_TX_RECORD|FCGI_EV_TX_GETVAL|FCGI_EV_FCONN_END, fconn->conn); + TRACE_DEVEL("leaving on error", FCGI_EV_TX_RECORD|FCGI_EV_TX_GETVAL|FCGI_EV_FCONN_ERR, fconn->conn); + return 0; +} + +/* Processes a GET_VALUES_RESULT record. Returns > 0 on success, 0 if it + * couldn't do anything. It is highly unexpected, but if the record is larger + * than a buffer and cannot be decoded in one time, an error is triggered and + * the connection is closed. GET_VALUES_RESULT record cannot be split. + */ +static int fcgi_conn_handle_values_result(struct fcgi_conn *fconn) +{ + struct buffer inbuf; + struct buffer *dbuf; + size_t offset; + + TRACE_ENTER(FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn); + + dbuf = &fconn->dbuf; + + /* Record too large to be fully decoded */ + if (b_size(dbuf) < (fconn->drl + fconn->drp)) + goto fail; + + /* process full record only */ + if (b_data(dbuf) < (fconn->drl + fconn->drp)) { + TRACE_DEVEL("leaving on missing data", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn); + return 0; + } + + if (unlikely(b_contig_data(dbuf, b_head_ofs(dbuf)) < fconn->drl)) { + /* Realign the dmux buffer if the record wraps. It is unexpected + * at this stage because it should be the first record received + * from the FCGI application. + */ + b_slow_realign_ofs(dbuf, trash.area, 0); + } + + inbuf = b_make(b_head(dbuf), b_data(dbuf), 0, fconn->drl); + + for (offset = 0; offset < b_data(&inbuf); ) { + struct fcgi_param p; + size_t ret; + + ret = fcgi_aligned_decode_param(&inbuf, offset, &p); + if (!ret) { + /* name or value too large to be decoded at once */ + TRACE_ERROR("error decoding GET_VALUES_RESULT param", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL|FCGI_EV_FCONN_ERR, fconn->conn); + goto fail; + } + offset += ret; + + if (isteqi(p.n, ist("FCGI_MPXS_CONNS"))) { + if (isteq(p.v, ist("1"))) { + TRACE_STATE("set mpxs param", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn, 0, 0, (size_t[]){1}); + fconn->flags |= FCGI_CF_MPXS_CONNS; + } + else { + TRACE_STATE("set mpxs param", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn, 0, 0, (size_t[]){0}); + fconn->flags &= ~FCGI_CF_MPXS_CONNS; + } + } + else if (isteqi(p.n, ist("FCGI_MAX_REQS"))) { + fconn->streams_limit = strl2ui(p.v.ptr, p.v.len); + TRACE_STATE("set streams_limit", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn, 0, 0, (size_t[]){fconn->streams_limit}); + } + /* + * Ignore all other params + */ + } + + /* Reset the number of concurrent streams supported if the FCGI + * application does not support connection multiplexing + */ + if (!(fconn->flags & FCGI_CF_MPXS_CONNS)) { + fconn->streams_limit = 1; + TRACE_STATE("no mpxs for streams_limit to 1", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn); + } + + /* We must be sure to have read exactly the announced record length, no + * more no less + */ + if (offset != fconn->drl) { + TRACE_ERROR("invalid GET_VALUES_RESULT record length", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL|FCGI_EV_FCONN_ERR, fconn->conn); + goto fail; + } + + TRACE_PROTO("FCGI GET_VALUES_RESULT record rcvd", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn, 0, 0, (size_t[]){fconn->drl}); + b_del(&fconn->dbuf, fconn->drl + fconn->drp); + fconn->drl = 0; + fconn->drp = 0; + fconn->state = FCGI_CS_RECORD_H; + fcgi_wake_unassigned_streams(fconn); + TRACE_STATE("switching to RECORD_H", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn); + TRACE_LEAVE(FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn); + return 1; + fail: + fconn->state = FCGI_CS_CLOSED; + TRACE_STATE("switching to CLOSED", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn); + TRACE_DEVEL("leaving on error", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL|FCGI_EV_FCONN_ERR, fconn->conn); + return 0; +} + +/* Sends an ABORT_REQUEST record for each active streams. Closed streams are + * excluded, as the streams which already received the end-of-stream. It returns + * > 0 if the record was sent tp all streams. Otherwise it returns 0. + */ +static int fcgi_conn_send_aborts(struct fcgi_conn *fconn) +{ + struct eb32_node *node; + struct fcgi_strm *fstrm; + + TRACE_ENTER(FCGI_EV_TX_RECORD, fconn->conn); + + node = eb32_lookup_ge(&fconn->streams_by_id, 1); + while (node) { + fstrm = container_of(node, struct fcgi_strm, by_id); + node = eb32_next(node); + if (fstrm->state != FCGI_SS_CLOSED && + !(fstrm->flags & (FCGI_SF_ES_RCVD|FCGI_SF_ABRT_SENT)) && + !fcgi_strm_send_abort(fconn, fstrm)) + return 0; + } + fconn->flags |= FCGI_CF_ABRTS_SENT; + TRACE_STATE("aborts sent to all fstrms", FCGI_EV_TX_RECORD, fconn->conn); + TRACE_LEAVE(FCGI_EV_TX_RECORD, fconn->conn); + return 1; +} + +/* Sends a BEGIN_REQUEST record. It returns > 0 on success, 0 if it couldn't do + * anything. BEGIN_REQUEST record cannot be split. So we wait to have enough + * space to proceed. It is small enough to be encoded in an empty buffer. + */ +static int fcgi_strm_send_begin_request(struct fcgi_conn *fconn, struct fcgi_strm *fstrm) +{ + struct buffer outbuf; + struct buffer *mbuf; + struct fcgi_begin_request rec = { .role = FCGI_RESPONDER, .flags = 0}; + int ret; + + TRACE_ENTER(FCGI_EV_TX_RECORD|FCGI_EV_TX_BEGREQ, fconn->conn, fstrm); + + mbuf = br_tail(fconn->mbuf); + retry: + if (!fcgi_get_buf(fconn, mbuf)) { + fconn->flags |= FCGI_CF_MUX_MALLOC; + fstrm->flags |= FCGI_SF_BLK_MROOM; + TRACE_STATE("waiting for fconn mbuf ring allocation", FCGI_EV_TX_RECORD|FCGI_EV_FSTRM_BLK|FCGI_EV_FCONN_BLK, fconn->conn, fstrm); + ret = 0; + goto end; + } + + while (1) { + outbuf = b_make(b_tail(mbuf), b_contig_space(mbuf), 0, 0); + if (outbuf.size >= FCGI_RECORD_HEADER_SZ || !b_space_wraps(mbuf)) + break; + realign_again: + b_slow_realign(mbuf, trash.area, b_data(mbuf)); + } + + if (outbuf.size < FCGI_RECORD_HEADER_SZ) + goto full; + + /* vsn: 1(FCGI_VERSION), type: (1)FCGI_BEGIN_REQUEST, id: fstrm->id, + * len: 0x0008, padding: 0x00, rsv: 0x00 */ + memcpy(outbuf.area, "\x01\x01\x00\x00\x00\x08\x00\x00", FCGI_RECORD_HEADER_SZ); + fcgi_set_record_id(outbuf.area, fstrm->id); + outbuf.data = FCGI_RECORD_HEADER_SZ; + + if (fconn->flags & FCGI_CF_KEEP_CONN) { + TRACE_STATE("keep connection opened", FCGI_EV_TX_RECORD|FCGI_EV_TX_BEGREQ, fconn->conn, fstrm); + rec.flags |= FCGI_KEEP_CONN; + } + if (!fcgi_encode_begin_request(&outbuf, &rec)) + goto full; + + /* commit the record */ + TRACE_PROTO("FCGI BEGIN_REQUEST record xferred", FCGI_EV_TX_RECORD|FCGI_EV_TX_BEGREQ, fconn->conn, fstrm, 0, (size_t[]){0}); + b_add(mbuf, outbuf.data); + fstrm->flags |= FCGI_SF_BEGIN_SENT; + fstrm->state = FCGI_SS_OPEN; + TRACE_STATE("switching to OPEN", FCGI_EV_TX_RECORD|FCGI_EV_TX_BEGREQ, fconn->conn, fstrm); + ret = 1; + + end: + TRACE_LEAVE(FCGI_EV_TX_RECORD|FCGI_EV_TX_BEGREQ, fconn->conn, fstrm); + return ret; + full: + if ((mbuf = br_tail_add(fconn->mbuf)) != NULL) + goto retry; + fconn->flags |= FCGI_CF_MUX_MFULL; + fstrm->flags |= FCGI_SF_BLK_MROOM; + TRACE_STATE("mbuf ring full", FCGI_EV_TX_RECORD|FCGI_EV_FSTRM_BLK|FCGI_EV_FCONN_BLK, fconn->conn); + ret = 0; + goto end; +} + +/* Sends an empty record of type <rtype>. It returns > 0 on success, 0 if it + * couldn't do anything. Empty record cannot be split. So we wait to have enough + * space to proceed. It is small enough to be encoded in an empty buffer. + */ +static int fcgi_strm_send_empty_record(struct fcgi_conn *fconn, struct fcgi_strm *fstrm, + enum fcgi_record_type rtype) +{ + struct buffer outbuf; + struct buffer *mbuf; + int ret; + + TRACE_ENTER(FCGI_EV_TX_RECORD, fconn->conn, fstrm); + mbuf = br_tail(fconn->mbuf); + retry: + if (!fcgi_get_buf(fconn, mbuf)) { + fconn->flags |= FCGI_CF_MUX_MALLOC; + fstrm->flags |= FCGI_SF_BLK_MROOM; + TRACE_STATE("waiting for fconn mbuf ring allocation", FCGI_EV_TX_RECORD|FCGI_EV_FSTRM_BLK|FCGI_EV_FCONN_BLK, fconn->conn, fstrm); + ret = 0; + goto end; + } + + while (1) { + outbuf = b_make(b_tail(mbuf), b_contig_space(mbuf), 0, 0); + if (outbuf.size >= FCGI_RECORD_HEADER_SZ || !b_space_wraps(mbuf)) + break; + realign_again: + b_slow_realign(mbuf, trash.area, b_data(mbuf)); + } + + if (outbuf.size < FCGI_RECORD_HEADER_SZ) + goto full; + + /* vsn: 1(FCGI_VERSION), type: rtype, id: fstrm->id, + * len: 0x0000, padding: 0x00, rsv: 0x00 */ + memcpy(outbuf.area, "\x01\x05\x00\x00\x00\x00\x00\x00", FCGI_RECORD_HEADER_SZ); + outbuf.area[1] = rtype; + fcgi_set_record_id(outbuf.area, fstrm->id); + outbuf.data = FCGI_RECORD_HEADER_SZ; + + /* commit the record */ + b_add(mbuf, outbuf.data); + ret = 1; + + end: + TRACE_LEAVE(FCGI_EV_TX_RECORD, fconn->conn, fstrm); + return ret; + full: + if ((mbuf = br_tail_add(fconn->mbuf)) != NULL) + goto retry; + fconn->flags |= FCGI_CF_MUX_MFULL; + fstrm->flags |= FCGI_SF_BLK_MROOM; + TRACE_STATE("mbuf ring full", FCGI_EV_TX_RECORD|FCGI_EV_FSTRM_BLK|FCGI_EV_FCONN_BLK, fconn->conn, fstrm); + ret = 0; + goto end; +} + + +/* Sends an empty PARAMS record. It relies on fcgi_strm_send_empty_record(). It + * marks the end of params. + */ +static int fcgi_strm_send_empty_params(struct fcgi_conn *fconn, struct fcgi_strm *fstrm) +{ + int ret; + + TRACE_POINT(FCGI_EV_TX_RECORD|FCGI_EV_TX_PARAMS, fconn->conn, fstrm); + ret = fcgi_strm_send_empty_record(fconn, fstrm, FCGI_PARAMS); + if (ret) { + fstrm->flags |= FCGI_SF_EP_SENT; + TRACE_PROTO("FCGI PARAMS record xferred", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, 0, (size_t[]){0}); + } + return ret; +} + +/* Sends an empty STDIN record. It relies on fcgi_strm_send_empty_record(). It + * marks the end of input. On success, all the request was successfully sent. + */ +static int fcgi_strm_send_empty_stdin(struct fcgi_conn *fconn, struct fcgi_strm *fstrm) +{ + int ret; + + TRACE_POINT(FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN|FCGI_EV_TX_EOI, fconn->conn, fstrm); + ret = fcgi_strm_send_empty_record(fconn, fstrm, FCGI_STDIN); + if (ret) { + fstrm->flags |= FCGI_SF_ES_SENT; + TRACE_PROTO("FCGI STDIN record xferred", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, 0, (size_t[]){0}); + TRACE_USER("FCGI request fully xferred", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN|FCGI_EV_TX_EOI, fconn->conn, fstrm); + TRACE_STATE("stdin data fully sent", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN|FCGI_EV_TX_EOI, fconn->conn, fstrm); + } + return ret; +} + +/* Sends an ABORT_REQUEST record. It relies on fcgi_strm_send_empty_record(). It + * stops the request processing. + */ +static int fcgi_strm_send_abort(struct fcgi_conn *fconn, struct fcgi_strm *fstrm) +{ + int ret; + + TRACE_POINT(FCGI_EV_TX_RECORD|FCGI_EV_TX_ABORT, fconn->conn, fstrm); + ret = fcgi_strm_send_empty_record(fconn, fstrm, FCGI_ABORT_REQUEST); + if (ret) { + fstrm->flags |= FCGI_SF_ABRT_SENT; + TRACE_PROTO("FCGI ABORT record xferred", FCGI_EV_TX_RECORD|FCGI_EV_TX_ABORT, fconn->conn, fstrm, 0, (size_t[]){0}); + TRACE_USER("FCGI request aborted", FCGI_EV_TX_RECORD|FCGI_EV_TX_ABORT, fconn->conn, fstrm); + TRACE_STATE("abort sent", FCGI_EV_TX_RECORD|FCGI_EV_TX_ABORT, fconn->conn, fstrm); + } + return ret; +} + +/* Sends a PARAMS record. Returns > 0 on success, 0 if it couldn't do + * anything. If there are too much K/V params to be encoded in a PARAMS record, + * several records are sent. However, a K/V param cannot be split between 2 + * records. + */ +static size_t fcgi_strm_send_params(struct fcgi_conn *fconn, struct fcgi_strm *fstrm, + struct htx *htx) +{ + struct buffer outbuf; + struct buffer *mbuf; + struct htx_blk *blk; + struct htx_sl *sl = NULL; + struct fcgi_strm_params params; + size_t total = 0; + + TRACE_ENTER(FCGI_EV_TX_RECORD|FCGI_EV_TX_PARAMS, fconn->conn, fstrm, htx); + + memset(¶ms, 0, sizeof(params)); + params.p = get_trash_chunk(); + + mbuf = br_tail(fconn->mbuf); + retry: + if (!fcgi_get_buf(fconn, mbuf)) { + fconn->flags |= FCGI_CF_MUX_MALLOC; + fstrm->flags |= FCGI_SF_BLK_MROOM; + TRACE_STATE("waiting for fconn mbuf ring allocation", FCGI_EV_TX_RECORD|FCGI_EV_FSTRM_BLK|FCGI_EV_FCONN_BLK, fconn->conn, fstrm); + goto end; + } + + while (1) { + outbuf = b_make(b_tail(mbuf), b_contig_space(mbuf), 0, 0); + if (outbuf.size >= FCGI_RECORD_HEADER_SZ || !b_space_wraps(mbuf)) + break; + realign_again: + b_slow_realign(mbuf, trash.area, b_data(mbuf)); + } + + if (outbuf.size < FCGI_RECORD_HEADER_SZ) + goto full; + + /* vsn: 1(FCGI_VERSION), type: (4)FCGI_PARAMS, id: fstrm->id, + * len: 0x0000 (fill later), padding: 0x00, rsv: 0x00 */ + memcpy(outbuf.area, "\x01\x04\x00\x00\x00\x00\x00\x00", FCGI_RECORD_HEADER_SZ); + fcgi_set_record_id(outbuf.area, fstrm->id); + outbuf.data = FCGI_RECORD_HEADER_SZ; + + blk = htx_get_head_blk(htx); + while (blk) { + enum htx_blk_type type; + uint32_t size = htx_get_blksz(blk); + struct fcgi_param p; + + type = htx_get_blk_type(blk); + switch (type) { + case HTX_BLK_REQ_SL: + sl = htx_get_blk_ptr(htx, blk); + if (sl->info.req.meth == HTTP_METH_HEAD) + fstrm->h1m.flags |= H1_MF_METH_HEAD; + if (sl->flags & HTX_SL_F_VER_11) + fstrm->h1m.flags |= H1_MF_VER_11; + break; + + case HTX_BLK_HDR: + p.n = htx_get_blk_name(htx, blk); + p.v = htx_get_blk_value(htx, blk); + + if (istmatch(p.n, ist(":fcgi-"))) { + p.n = istadv(p.n, 6); + if (isteq(p.n, ist("gateway_interface"))) + params.mask |= FCGI_SP_CGI_GATEWAY; + else if (isteq(p.n, ist("document_root"))) { + params.mask |= FCGI_SP_DOC_ROOT; + params.docroot = p.v; + } + else if (isteq(p.n, ist("script_name"))) { + params.mask |= FCGI_SP_SCRIPT_NAME; + params.scriptname = p.v; + } + else if (isteq(p.n, ist("path_info"))) { + params.mask |= FCGI_SP_PATH_INFO; + params.pathinfo = p.v; + } + else if (isteq(p.n, ist("request_uri"))) { + params.mask |= FCGI_SP_REQ_URI; + params.uri = p.v; + } + else if (isteq(p.n, ist("request_meth"))) + params.mask |= FCGI_SP_REQ_METH; + else if (isteq(p.n, ist("query_string"))) + params.mask |= FCGI_SP_REQ_QS; + else if (isteq(p.n, ist("server_name"))) + params.mask |= FCGI_SP_SRV_NAME; + else if (isteq(p.n, ist("server_port"))) + params.mask |= FCGI_SP_SRV_PORT; + else if (isteq(p.n, ist("server_protocol"))) + params.mask |= FCGI_SP_SRV_PROTO; + else if (isteq(p.n, ist("remote_addr"))) + params.mask |= FCGI_SP_REM_ADDR; + else if (isteq(p.n, ist("remote_port"))) + params.mask |= FCGI_SP_REM_PORT; + else if (isteq(p.n, ist("script_filename"))) + params.mask |= FCGI_SP_SCRIPT_FILE; + else if (isteq(p.n, ist("path_translated"))) + params.mask |= FCGI_SP_PATH_TRANS; + else if (isteq(p.n, ist("https"))) + params.mask |= FCGI_SP_HTTPS; + else if (isteq(p.n, ist("server_software"))) + params.mask |= FCGI_SP_SRV_SOFT; + } + else if (isteq(p.n, ist("content-length"))) { + p.n = ist("CONTENT_LENGTH"); + params.mask |= FCGI_SP_CONT_LEN; + } + else if (isteq(p.n, ist("content-type"))) + p.n = ist("CONTENT_TYPE"); + else { + struct ist n; + + if (isteq(p.n, ist("host"))) + params.srv_name = p.v; + else if (isteq(p.n, ist("te"))) { + /* "te" may only be sent with "trailers" if this value + * is present, otherwise it must be deleted. + */ + p.v = istist(p.v, ist("trailers")); + if (!isttest(p.v) || (p.v.len > 8 && p.v.ptr[8] != ',')) + break; + p.v = ist("trailers"); + } + + /* Skip header if same name is used to add the server name */ + if (isttest(fconn->proxy->server_id_hdr_name) && isteq(p.n, fconn->proxy->server_id_hdr_name)) + break; + + n = ist2(trash.area, 0); + istcat(&n, ist("http_"), trash.size); + istcat(&n, p.n, trash.size); + p.n = n; + } + + if (!fcgi_encode_param(&outbuf, &p)) { + if (b_space_wraps(mbuf)) + goto realign_again; + if (outbuf.data == FCGI_RECORD_HEADER_SZ) + goto full; + goto done; + } + break; + + case HTX_BLK_EOH: + if (isttest(fconn->proxy->server_id_hdr_name)) { + struct server *srv = objt_server(fconn->conn->target); + + if (!srv) + goto done; + + p.n = ist2(trash.area, 0); + istcat(&p.n, ist("http_"), trash.size); + istcat(&p.n, fconn->proxy->server_id_hdr_name, trash.size); + p.v = ist(srv->id); + + if (!fcgi_encode_param(&outbuf, &p)) { + if (b_space_wraps(mbuf)) + goto realign_again; + if (outbuf.data == FCGI_RECORD_HEADER_SZ) + goto full; + } + TRACE_STATE("add server name header", FCGI_EV_TX_RECORD|FCGI_EV_TX_PARAMS, fconn->conn, fstrm); + } + goto done; + + default: + break; + } + total += size; + blk = htx_remove_blk(htx, blk); + } + + done: + if (!fcgi_set_default_param(fconn, fstrm, htx, sl, ¶ms)) { + TRACE_ERROR("error setting default params", FCGI_EV_TX_RECORD|FCGI_EV_STRM_ERR, fconn->conn, fstrm); + goto error; + } + + if (!fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_CGI_GATEWAY) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_DOC_ROOT) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_SCRIPT_NAME) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_PATH_INFO) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_REQ_URI) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_REQ_METH) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_REQ_QS) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_SRV_NAME) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_SRV_PORT) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_SRV_PROTO) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_REM_ADDR) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_REM_PORT) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_SCRIPT_FILE) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_PATH_TRANS) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_CONT_LEN) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_SRV_SOFT) || + !fcgi_encode_default_param(fconn, fstrm, ¶ms, &outbuf, FCGI_SP_HTTPS)) { + TRACE_ERROR("error encoding default params", FCGI_EV_TX_RECORD|FCGI_EV_STRM_ERR, fconn->conn, fstrm); + goto error; + } + + /* update the record's size */ + TRACE_PROTO("FCGI PARAMS record xferred", FCGI_EV_TX_RECORD|FCGI_EV_TX_PARAMS, fconn->conn, fstrm, 0, (size_t[]){outbuf.data - FCGI_RECORD_HEADER_SZ}); + fcgi_set_record_size(outbuf.area, outbuf.data - FCGI_RECORD_HEADER_SZ); + b_add(mbuf, outbuf.data); + + end: + TRACE_LEAVE(FCGI_EV_TX_RECORD|FCGI_EV_TX_PARAMS, fconn->conn, fstrm, htx, (size_t[]){total}); + return total; + full: + if ((mbuf = br_tail_add(fconn->mbuf)) != NULL) + goto retry; + fconn->flags |= FCGI_CF_MUX_MFULL; + fstrm->flags |= FCGI_SF_BLK_MROOM; + TRACE_STATE("mbuf ring full", FCGI_EV_TX_RECORD|FCGI_EV_FSTRM_BLK|FCGI_EV_FCONN_BLK, fconn->conn, fstrm); + if (total) + goto error; + goto end; + + error: + htx->flags |= HTX_FL_PROCESSING_ERROR; + TRACE_ERROR("processing error sending PARAMS record", FCGI_EV_TX_RECORD|FCGI_EV_STRM_ERR, fconn->conn, fstrm); + fcgi_strm_error(fstrm); + goto end; +} + +/* Sends a STDIN record. Returns > 0 on success, 0 if it couldn't do + * anything. STDIN records contain the request body. + */ +static size_t fcgi_strm_send_stdin(struct fcgi_conn *fconn, struct fcgi_strm *fstrm, + struct htx *htx, size_t count, struct buffer *buf) +{ + struct buffer outbuf; + struct buffer *mbuf; + struct htx_blk *blk; + enum htx_blk_type type; + uint32_t size, extra_bytes; + size_t total = 0; + + extra_bytes = 0; + + TRACE_ENTER(FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, htx, (size_t[]){count}); + if (!count) + goto end; + + mbuf = br_tail(fconn->mbuf); + retry: + if (!fcgi_get_buf(fconn, mbuf)) { + fconn->flags |= FCGI_CF_MUX_MALLOC; + fstrm->flags |= FCGI_SF_BLK_MROOM; + TRACE_STATE("waiting for fconn mbuf ring allocation", FCGI_EV_TX_RECORD|FCGI_EV_FSTRM_BLK|FCGI_EV_FCONN_BLK, fconn->conn, fstrm); + goto end; + } + + /* Perform some optimizations to reduce the number of buffer copies. + * First, if the mux's buffer is empty and the htx area contains exactly + * one data block of the same size as the requested count, and this + * count fits within the record size, then it's possible to simply swap + * the caller's buffer with the mux's output buffer and adjust offsets + * and length to match the entire DATA HTX block in the middle. In this + * case we perform a true zero-copy operation from end-to-end. This is + * the situation that happens all the time with large files. Second, if + * this is not possible, but the mux's output buffer is empty, we still + * have an opportunity to avoid the copy to the intermediary buffer, by + * making the intermediary buffer's area point to the output buffer's + * area. In this case we want to skip the HTX header to make sure that + * copies remain aligned and that this operation remains possible all + * the time. This goes for headers, data blocks and any data extracted + * from the HTX blocks. + */ + blk = htx_get_head_blk(htx); + if (!blk) + goto end; + type = htx_get_blk_type(blk); + size = htx_get_blksz(blk); + if (unlikely(size == count && htx_nbblks(htx) == 1 && type == HTX_BLK_DATA)) { + void *old_area = mbuf->area; + int eom = (htx->flags & HTX_FL_EOM); + + /* Last block of the message: Reserve the size for the empty stdin record */ + if (eom) + extra_bytes = FCGI_RECORD_HEADER_SZ; + + if (b_data(mbuf)) { + /* Too bad there are data left there. We're willing to memcpy/memmove + * up to 1/4 of the buffer, which means that it's OK to copy a large + * record into a buffer containing few data if it needs to be realigned, + * and that it's also OK to copy few data without realigning. Otherwise + * we'll pretend the mbuf is full and wait for it to become empty. + */ + if (size + FCGI_RECORD_HEADER_SZ + extra_bytes <= b_room(mbuf) && + (b_data(mbuf) <= b_size(mbuf) / 4 || + (size <= b_size(mbuf) / 4 && size + FCGI_RECORD_HEADER_SZ + extra_bytes <= b_contig_space(mbuf)))) + goto copy; + goto full; + } + + TRACE_PROTO("sending stding data (zero-copy)", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, htx, (size_t[]){size}); + /* map a FCGI record to the HTX block so that we can put the + * record header there. + */ + *mbuf = b_make(buf->area, buf->size, sizeof(struct htx) + blk->addr - FCGI_RECORD_HEADER_SZ, size + FCGI_RECORD_HEADER_SZ); + outbuf.area = b_head(mbuf); + + /* prepend a FCGI record header just before the DATA block */ + memcpy(outbuf.area, "\x01\x05\x00\x00\x00\x00\x00\x00", FCGI_RECORD_HEADER_SZ); + fcgi_set_record_id(outbuf.area, fstrm->id); + fcgi_set_record_size(outbuf.area, size); + + /* and exchange with our old area */ + buf->area = old_area; + buf->data = buf->head = 0; + total += size; + + htx = (struct htx *)buf->area; + htx_reset(htx); + if (eom) + goto empty_stdin; + goto end; + } + + copy: + while (1) { + outbuf = b_make(b_tail(mbuf), b_contig_space(mbuf), 0, 0); + if (outbuf.size >= FCGI_RECORD_HEADER_SZ + extra_bytes || !b_space_wraps(mbuf)) + break; + realign_again: + b_slow_realign(mbuf, trash.area, b_data(mbuf)); + } + + if (outbuf.size < FCGI_RECORD_HEADER_SZ + extra_bytes) + goto full; + + /* vsn: 1(FCGI_VERSION), type: (5)FCGI_STDIN, id: fstrm->id, + * len: 0x0000 (fill later), padding: 0x00, rsv: 0x00 */ + memcpy(outbuf.area, "\x01\x05\x00\x00\x00\x00\x00\x00", FCGI_RECORD_HEADER_SZ); + fcgi_set_record_id(outbuf.area, fstrm->id); + outbuf.data = FCGI_RECORD_HEADER_SZ; + + blk = htx_get_head_blk(htx); + while (blk && count) { + enum htx_blk_type type = htx_get_blk_type(blk); + uint32_t size = htx_get_blksz(blk); + struct ist v; + + switch (type) { + case HTX_BLK_DATA: + TRACE_PROTO("sending stding data", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, htx, (size_t[]){size}); + v = htx_get_blk_value(htx, blk); + + if (htx_is_unique_blk(htx, blk) && (htx->flags & HTX_FL_EOM)) + extra_bytes = FCGI_RECORD_HEADER_SZ; /* Last block of the message */ + + if (v.len > count) { + v.len = count; + extra_bytes = 0; + } + + if (v.len + FCGI_RECORD_HEADER_SZ + extra_bytes > b_room(&outbuf)) { + /* It doesn't fit at once. If it at least fits once split and + * the amount of data to move is low, let's defragment the + * buffer now. + */ + if (b_space_wraps(mbuf) && + b_data(&outbuf) + v.len + extra_bytes <= b_room(mbuf) && + b_data(mbuf) <= MAX_DATA_REALIGN) + goto realign_again; + v.len = (FCGI_RECORD_HEADER_SZ + extra_bytes > b_room(&outbuf) + ? 0 + : b_room(&outbuf) - FCGI_RECORD_HEADER_SZ - extra_bytes); + } + if (!v.len || !chunk_memcat(&outbuf, v.ptr, v.len)) { + if (outbuf.data == FCGI_RECORD_HEADER_SZ) + goto full; + goto done; + } + if (v.len != size) { + total += v.len; + count -= v.len; + htx_cut_data_blk(htx, blk, v.len); + goto done; + } + break; + + default: + break; + } + total += size; + count -= size; + blk = htx_remove_blk(htx, blk); + } + + done: + /* update the record's size */ + TRACE_PROTO("FCGI STDIN record xferred", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, 0, (size_t[]){outbuf.data - FCGI_RECORD_HEADER_SZ}); + fcgi_set_record_size(outbuf.area, outbuf.data - FCGI_RECORD_HEADER_SZ); + b_add(mbuf, outbuf.data); + + /* Send the empty stding here to finish the message */ + if (htx_is_empty(htx) && (htx->flags & HTX_FL_EOM)) { + empty_stdin: + TRACE_PROTO("sending FCGI STDIN record", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, htx); + if (!fcgi_strm_send_empty_stdin(fconn, fstrm)) { + /* bytes already reserved for this record. It should not fail */ + htx->flags |= HTX_FL_PROCESSING_ERROR; + TRACE_ERROR("processing error sending empty STDIN record", FCGI_EV_TX_RECORD|FCGI_EV_STRM_ERR, fconn->conn, fstrm); + fcgi_strm_error(fstrm); + } + } + + end: + TRACE_LEAVE(FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, htx, (size_t[]){total}); + return total; + full: + if ((mbuf = br_tail_add(fconn->mbuf)) != NULL) + goto retry; + fconn->flags |= FCGI_CF_MUX_MFULL; + fstrm->flags |= FCGI_SF_BLK_MROOM; + TRACE_STATE("mbuf ring full", FCGI_EV_TX_RECORD|FCGI_EV_FSTRM_BLK|FCGI_EV_FCONN_BLK, fconn->conn, fstrm); + goto end; +} + +/* Processes a STDOUT record. Returns > 0 on success, 0 if it couldn't do + * anything. STDOUT records contain the entire response. All the content is + * copied in the stream's rxbuf. The parsing will be handled in fcgi_rcv_buf(). + */ +static int fcgi_strm_handle_stdout(struct fcgi_conn *fconn, struct fcgi_strm *fstrm) +{ + struct buffer *dbuf; + size_t ret; + size_t max; + + TRACE_ENTER(FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm); + + dbuf = &fconn->dbuf; + + /* Only padding remains */ + if (fconn->state == FCGI_CS_RECORD_P) + goto end_transfer; + + if (b_data(dbuf) < (fconn->drl + fconn->drp) && + b_size(dbuf) > (fconn->drl + fconn->drp) && + buf_room_for_htx_data(dbuf)) + goto fail; // incomplete record + + if (!fcgi_get_buf(fconn, &fstrm->rxbuf)) { + fconn->flags |= FCGI_CF_DEM_SALLOC; + TRACE_STATE("waiting for fstrm rxbuf allocation", FCGI_EV_RX_RECORD|FCGI_EV_FSTRM_BLK, fconn->conn, fstrm); + goto fail; + } + + /*max = MIN(b_room(&fstrm->rxbuf), fconn->drl);*/ + max = buf_room_for_htx_data(&fstrm->rxbuf); + if (!b_data(&fstrm->rxbuf)) + fstrm->rxbuf.head = sizeof(struct htx); + if (max > fconn->drl) + max = fconn->drl; + + ret = b_xfer(&fstrm->rxbuf, dbuf, max); + if (!ret) + goto fail; + fconn->drl -= ret; + TRACE_DATA("move some data to fstrm rxbuf", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm, 0, (size_t[]){ret}); + TRACE_PROTO("FCGI STDOUT record rcvd", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm, 0, (size_t[]){ret}); + + if (!buf_room_for_htx_data(&fstrm->rxbuf)) { + fconn->flags |= FCGI_CF_DEM_SFULL; + TRACE_STATE("fstrm rxbuf full", FCGI_EV_RX_RECORD|FCGI_EV_FSTRM_BLK, fconn->conn, fstrm); + } + + if (fconn->drl) + goto fail; + + end_transfer: + fconn->state = FCGI_CS_RECORD_P; + fconn->drl += fconn->drp; + fconn->drp = 0; + ret = MIN(b_data(&fconn->dbuf), fconn->drl); + b_del(&fconn->dbuf, ret); + fconn->drl -= ret; + if (fconn->drl) + goto fail; + + fconn->state = FCGI_CS_RECORD_H; + TRACE_STATE("switching to RECORD_H", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn, fstrm); + TRACE_LEAVE(FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm); + return 1; + fail: + TRACE_DEVEL("leaving on missing data or error", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm); + return 0; +} + + +/* Processes an empty STDOUT. Returns > 0 on success, 0 if it couldn't do + * anything. It only skip the padding in fact, there is no payload for such + * records. It marks the end of the response. + */ +static int fcgi_strm_handle_empty_stdout(struct fcgi_conn *fconn, struct fcgi_strm *fstrm) +{ + int ret; + + TRACE_ENTER(FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm); + + fconn->state = FCGI_CS_RECORD_P; + TRACE_STATE("switching to RECORD_P", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm); + fconn->drl += fconn->drp; + fconn->drp = 0; + ret = MIN(b_data(&fconn->dbuf), fconn->drl); + b_del(&fconn->dbuf, ret); + fconn->drl -= ret; + if (fconn->drl) { + TRACE_DEVEL("leaving on missing data or error", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm); + return 0; + } + fconn->state = FCGI_CS_RECORD_H; + fstrm->flags |= FCGI_SF_ES_RCVD; + TRACE_PROTO("FCGI STDOUT record rcvd", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm, 0, (size_t[]){0}); + TRACE_STATE("stdout data fully send, switching to RECORD_H", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR|FCGI_EV_RX_EOI, fconn->conn, fstrm); + TRACE_LEAVE(FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm); + return 1; +} + +/* Processes a STDERR record. Returns > 0 on success, 0 if it couldn't do + * anything. + */ +static int fcgi_strm_handle_stderr(struct fcgi_conn *fconn, struct fcgi_strm *fstrm) +{ + struct buffer *dbuf; + struct buffer tag; + size_t ret; + + TRACE_ENTER(FCGI_EV_RX_RECORD|FCGI_EV_RX_STDERR, fconn->conn, fstrm); + dbuf = &fconn->dbuf; + + /* Only padding remains */ + if (fconn->state == FCGI_CS_RECORD_P || !fconn->drl) + goto end_transfer; + + if (b_data(dbuf) < (fconn->drl + fconn->drp) && + b_size(dbuf) > (fconn->drl + fconn->drp) && + buf_room_for_htx_data(dbuf)) + goto fail; // incomplete record + + chunk_reset(&trash); + ret = b_force_xfer(&trash, dbuf, MIN(b_room(&trash), fconn->drl)); + if (!ret) + goto fail; + fconn->drl -= ret; + TRACE_PROTO("FCGI STDERR record rcvd", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDERR, fconn->conn, fstrm, 0, (size_t[]){ret}); + + trash.area[ret] = '\n'; + trash.area[ret+1] = '\0'; + tag.area = fconn->app->name; tag.data = strlen(fconn->app->name); + app_log(&fconn->app->loggers, &tag, LOG_ERR, "%s", trash.area); + + if (fconn->drl) + goto fail; + + end_transfer: + fconn->state = FCGI_CS_RECORD_P; + fconn->drl += fconn->drp; + fconn->drp = 0; + ret = MIN(b_data(&fconn->dbuf), fconn->drl); + b_del(&fconn->dbuf, ret); + fconn->drl -= ret; + if (fconn->drl) + goto fail; + fconn->state = FCGI_CS_RECORD_H; + TRACE_STATE("switching to RECORD_H", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn, fstrm); + TRACE_LEAVE(FCGI_EV_RX_RECORD|FCGI_EV_RX_STDERR, fconn->conn, fstrm); + return 1; + fail: + TRACE_DEVEL("leaving on missing data or error", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDERR, fconn->conn, fstrm); + return 0; +} + +/* Processes an END_REQUEST record. Returns > 0 on success, 0 if it couldn't do + * anything. If the empty STDOUT record is not already received, this one marks + * the end of the response. It is highly unexpected, but if the record is larger + * than a buffer and cannot be decoded in one time, an error is triggered and + * the connection is closed. END_REQUEST record cannot be split. + */ +static int fcgi_strm_handle_end_request(struct fcgi_conn *fconn, struct fcgi_strm *fstrm) +{ + struct buffer inbuf; + struct buffer *dbuf; + struct fcgi_end_request endreq; + + TRACE_ENTER(FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ, fconn->conn, fstrm); + dbuf = &fconn->dbuf; + + /* Record too large to be fully decoded */ + if (b_size(dbuf) < (fconn->drl + fconn->drp)) { + TRACE_ERROR("END_REQUEST record too large", FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ|FCGI_EV_FSTRM_ERR, fconn->conn, fstrm); + goto fail; + } + + /* process full record only */ + if (b_data(dbuf) < (fconn->drl + fconn->drp)) { + TRACE_DEVEL("leaving on missing data", FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ, fconn->conn); + return 0; + } + + if (unlikely(b_contig_data(dbuf, b_head_ofs(dbuf)) < fconn->drl)) { + /* Realign the dmux buffer if the record wraps. It is unexpected + * at this stage because it should be the first record received + * from the FCGI application. + */ + b_slow_realign_ofs(dbuf, trash.area, 0); + } + + inbuf = b_make(b_head(dbuf), b_data(dbuf), 0, fconn->drl); + + if (!fcgi_decode_end_request(&inbuf, 0, &endreq)) { + TRACE_ERROR("END_REQUEST record decoding failure", FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ|FCGI_EV_FSTRM_ERR, fconn->conn, fstrm); + goto fail; + } + + fstrm->flags |= FCGI_SF_ES_RCVD; + TRACE_STATE("end of script reported", FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ|FCGI_EV_RX_EOI, fconn->conn, fstrm); + TRACE_PROTO("FCGI END_REQUEST record rcvd", FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ, fconn->conn, fstrm, 0, (size_t[]){fconn->drl}); + fstrm->proto_status = endreq.errcode; + fcgi_strm_close(fstrm); + + b_del(&fconn->dbuf, fconn->drl + fconn->drp); + fconn->drl = 0; + fconn->drp = 0; + fconn->state = FCGI_CS_RECORD_H; + TRACE_STATE("switching to RECORD_H", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn, fstrm); + TRACE_LEAVE(FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ, fconn->conn, fstrm); + return 1; + + fail: + fcgi_strm_error(fstrm); + TRACE_DEVEL("leaving on error", FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ|FCGI_EV_FSTRM_ERR, fconn->conn, fstrm); + return 0; +} + +/* process Rx records to be demultiplexed */ +static void fcgi_process_demux(struct fcgi_conn *fconn) +{ + struct fcgi_strm *fstrm = NULL, *tmp_fstrm; + struct fcgi_header hdr; + int ret; + + TRACE_ENTER(FCGI_EV_FCONN_WAKE, fconn->conn); + + if (fconn->state == FCGI_CS_CLOSED) + return; + + if (unlikely(fconn->state < FCGI_CS_RECORD_H)) { + if (fconn->state == FCGI_CS_INIT) { + TRACE_STATE("waiting FCGI GET_VALUES to be sent", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR|FCGI_EV_RX_GETVAL, fconn->conn); + return; + } + if (fconn->state == FCGI_CS_SETTINGS) { + /* ensure that what is pending is a valid GET_VALUES_RESULT record. */ + TRACE_STATE("receiving FCGI record header", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn); + ret = fcgi_decode_record_hdr(&fconn->dbuf, 0, &hdr); + if (!ret) { + TRACE_ERROR("header record decoding failure", FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ|FCGI_EV_FSTRM_ERR, fconn->conn, fstrm); + goto fail; + } + b_del(&fconn->dbuf, ret); + + if (hdr.id || (hdr.type != FCGI_GET_VALUES_RESULT && hdr.type != FCGI_UNKNOWN_TYPE)) { + fconn->state = FCGI_CS_CLOSED; + TRACE_ERROR("unexpected record type or flags", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR|FCGI_EV_RX_GETVAL|FCGI_EV_FCONN_ERR, fconn->conn); + TRACE_STATE("switching to CLOSED", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR|FCGI_EV_RX_GETVAL|FCGI_EV_FCONN_ERR, fconn->conn); + goto fail; + } + goto new_record; + } + } + + /* process as many incoming records as possible below */ + while (1) { + if (!b_data(&fconn->dbuf)) { + TRACE_DEVEL("no more Rx data", FCGI_EV_RX_RECORD, fconn->conn); + break; + } + + if (fconn->state == FCGI_CS_CLOSED) { + TRACE_STATE("end of connection reported", FCGI_EV_RX_RECORD|FCGI_EV_RX_EOI, fconn->conn); + break; + } + + if (fconn->state == FCGI_CS_RECORD_H) { + TRACE_PROTO("receiving FCGI record header", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn); + ret = fcgi_decode_record_hdr(&fconn->dbuf, 0, &hdr); + if (!ret) + break; + b_del(&fconn->dbuf, ret); + + new_record: + fconn->dsi = hdr.id; + fconn->drt = hdr.type; + fconn->drl = hdr.len; + fconn->drp = hdr.padding; + fconn->state = FCGI_CS_RECORD_D; + TRACE_STATE("FCGI record header rcvd, switching to RECORD_D", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn); + } + + /* Only FCGI_CS_RECORD_D or FCGI_CS_RECORD_P */ + tmp_fstrm = fcgi_conn_st_by_id(fconn, fconn->dsi); + + if (tmp_fstrm != fstrm && fstrm && fcgi_strm_sc(fstrm) && + (b_data(&fstrm->rxbuf) || + fcgi_conn_read0_pending(fconn) || + fstrm->state == FCGI_SS_CLOSED || + (fstrm->flags & FCGI_SF_ES_RCVD) || + se_fl_test(fstrm->sd, SE_FL_ERROR | SE_FL_ERR_PENDING | SE_FL_EOS))) { + /* we may have to signal the upper layers */ + TRACE_DEVEL("notifying stream before switching SID", FCGI_EV_RX_RECORD|FCGI_EV_STRM_WAKE, fconn->conn, fstrm); + se_fl_set(fstrm->sd, SE_FL_RCV_MORE); + fcgi_strm_notify_recv(fstrm); + } + fstrm = tmp_fstrm; + + if (fstrm->state == FCGI_SS_CLOSED && fconn->dsi != 0) { + /* ignore all record for closed streams */ + goto ignore_record; + } + if (fstrm->state == FCGI_SS_IDLE) { + /* ignore all record for unknown streams */ + goto ignore_record; + } + + switch (fconn->drt) { + case FCGI_GET_VALUES_RESULT: + TRACE_PROTO("receiving FCGI GET_VALUES_RESULT record", FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn); + ret = fcgi_conn_handle_values_result(fconn); + break; + + case FCGI_STDOUT: + if (fstrm->flags & FCGI_SF_ES_RCVD) + goto ignore_record; + + TRACE_PROTO("receiving FCGI STDOUT record", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDOUT, fconn->conn, fstrm); + if (fconn->drl) + ret = fcgi_strm_handle_stdout(fconn, fstrm); + else + ret = fcgi_strm_handle_empty_stdout(fconn, fstrm); + break; + + case FCGI_STDERR: + TRACE_PROTO("receiving FCGI STDERR record", FCGI_EV_RX_RECORD|FCGI_EV_RX_STDERR, fconn->conn, fstrm); + ret = fcgi_strm_handle_stderr(fconn, fstrm); + break; + + case FCGI_END_REQUEST: + TRACE_PROTO("receiving FCGI END_REQUEST record", FCGI_EV_RX_RECORD|FCGI_EV_RX_ENDREQ, fconn->conn, fstrm); + ret = fcgi_strm_handle_end_request(fconn, fstrm); + break; + + /* implement all extra record types here */ + default: + ignore_record: + /* drop records that we ignore. They may be + * larger than the buffer so we drain all of + * their contents until we reach the end. + */ + fconn->state = FCGI_CS_RECORD_P; + fconn->drl += fconn->drp; + fconn->drp = 0; + ret = MIN(b_data(&fconn->dbuf), fconn->drl); + TRACE_PROTO("receiving FCGI ignored record", FCGI_EV_RX_RECORD, fconn->conn, fstrm, 0, (size_t[]){ret}); + TRACE_STATE("switching to RECORD_P", FCGI_EV_RX_RECORD, fconn->conn, fstrm); + b_del(&fconn->dbuf, ret); + fconn->drl -= ret; + ret = (fconn->drl == 0); + } + + /* error or missing data condition met above ? */ + if (ret <= 0) { + TRACE_DEVEL("insufficient data to proceed", FCGI_EV_RX_RECORD, fconn->conn, fstrm); + break; + } + + if (fconn->state != FCGI_CS_RECORD_H && !(fconn->drl+fconn->drp)) { + fconn->state = FCGI_CS_RECORD_H; + TRACE_STATE("switching to RECORD_H", FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn); + } + } + + fail: + /* we can go here on missing data, blocked response or error */ + if (fstrm && fcgi_strm_sc(fstrm) && + (b_data(&fstrm->rxbuf) || + fcgi_conn_read0_pending(fconn) || + fstrm->state == FCGI_SS_CLOSED || + (fstrm->flags & FCGI_SF_ES_RCVD) || + se_fl_test(fstrm->sd, SE_FL_ERROR | SE_FL_ERR_PENDING | SE_FL_EOS))) { + /* we may have to signal the upper layers */ + TRACE_DEVEL("notifying stream before switching SID", FCGI_EV_RX_RECORD|FCGI_EV_STRM_WAKE, fconn->conn, fstrm); + se_fl_set(fstrm->sd, SE_FL_RCV_MORE); + fcgi_strm_notify_recv(fstrm); + } + + fcgi_conn_restart_reading(fconn, 0); +} + +/* process Tx records from streams to be multiplexed. Returns > 0 if it reached + * the end. + */ +static int fcgi_process_mux(struct fcgi_conn *fconn) +{ + struct fcgi_strm *fstrm, *fstrm_back; + + TRACE_ENTER(FCGI_EV_FCONN_WAKE, fconn->conn); + + if (unlikely(fconn->state < FCGI_CS_RECORD_H)) { + if (unlikely(fconn->state == FCGI_CS_INIT)) { + if (!(fconn->flags & FCGI_CF_GET_VALUES)) { + fconn->state = FCGI_CS_RECORD_H; + TRACE_STATE("switching to RECORD_H", FCGI_EV_TX_RECORD|FCGI_EV_RX_RECORD|FCGI_EV_RX_FHDR, fconn->conn); + fcgi_wake_unassigned_streams(fconn); + goto mux; + } + TRACE_PROTO("sending FCGI GET_VALUES record", FCGI_EV_TX_RECORD|FCGI_EV_TX_GETVAL, fconn->conn); + if (unlikely(!fcgi_conn_send_get_values(fconn))) + goto fail; + fconn->state = FCGI_CS_SETTINGS; + TRACE_STATE("switching to SETTINGS", FCGI_EV_TX_RECORD|FCGI_EV_RX_RECORD|FCGI_EV_RX_GETVAL, fconn->conn); + } + /* need to wait for the other side */ + if (fconn->state < FCGI_CS_RECORD_H) + goto done; + } + + mux: + list_for_each_entry_safe(fstrm, fstrm_back, &fconn->send_list, send_list) { + if (fconn->state == FCGI_CS_CLOSED || fconn->flags & FCGI_CF_MUX_BLOCK_ANY) + break; + + if (fstrm->flags & FCGI_SF_NOTIFIED) + continue; + + /* If the sender changed his mind and unsubscribed, let's just + * remove the stream from the send_list. + */ + if (!(fstrm->flags & (FCGI_SF_WANT_SHUTR|FCGI_SF_WANT_SHUTW)) && + (!fstrm->subs || !(fstrm->subs->events & SUB_RETRY_SEND))) { + LIST_DEL_INIT(&fstrm->send_list); + continue; + } + + if (fstrm->subs && fstrm->subs->events & SUB_RETRY_SEND) { + TRACE_POINT(FCGI_EV_STRM_WAKE, fconn->conn, fstrm); + fstrm->flags &= ~FCGI_SF_BLK_ANY; + fstrm->flags |= FCGI_SF_NOTIFIED; + tasklet_wakeup(fstrm->subs->tasklet); + fstrm->subs->events &= ~SUB_RETRY_SEND; + if (!fstrm->subs->events) + fstrm->subs = NULL; + } else { + /* it's the shut request that was queued */ + TRACE_POINT(FCGI_EV_STRM_WAKE, fconn->conn, fstrm); + tasklet_wakeup(fstrm->shut_tl); + } + } + + fail: + if (fconn->state == FCGI_CS_CLOSED) { + if (fconn->stream_cnt - fconn->nb_reserved > 0) { + fcgi_conn_send_aborts(fconn); + if (fconn->flags & FCGI_CF_MUX_BLOCK_ANY) { + TRACE_DEVEL("leaving in blocked situation", FCGI_EV_FCONN_WAKE|FCGI_EV_FCONN_BLK, fconn->conn); + return 0; + } + } + } + + done: + TRACE_LEAVE(FCGI_EV_FCONN_WAKE, fconn->conn); + return 1; +} + + +/* Attempt to read data, and subscribe if none available. + * The function returns 1 if data has been received, otherwise zero. + */ +static int fcgi_recv(struct fcgi_conn *fconn) +{ + struct connection *conn = fconn->conn; + struct buffer *buf; + int max; + size_t ret; + + TRACE_ENTER(FCGI_EV_FCONN_RECV, conn); + + if (fconn->wait_event.events & SUB_RETRY_RECV) { + TRACE_DEVEL("leaving on sub_recv", FCGI_EV_FCONN_RECV, conn); + return (b_data(&fconn->dbuf)); + } + + if (!fcgi_recv_allowed(fconn)) { + TRACE_DEVEL("leaving on !recv_allowed", FCGI_EV_FCONN_RECV, conn); + return 1; + } + + buf = fcgi_get_buf(fconn, &fconn->dbuf); + if (!buf) { + TRACE_DEVEL("waiting for fconn dbuf allocation", FCGI_EV_FCONN_RECV|FCGI_EV_FCONN_BLK, conn); + fconn->flags |= FCGI_CF_DEM_DALLOC; + return 0; + } + + if (!b_data(buf)) { + /* try to pre-align the buffer like the + * rxbufs will be to optimize memory copies. We'll make + * sure that the record header lands at the end of the + * HTX block to alias it upon recv. We cannot use the + * head because rcv_buf() will realign the buffer if + * it's empty. Thus we cheat and pretend we already + * have a few bytes there. + */ + max = buf_room_for_htx_data(buf) + (fconn->state == FCGI_CS_RECORD_H ? FCGI_RECORD_HEADER_SZ : 0); + buf->head = sizeof(struct htx) - (fconn->state == FCGI_CS_RECORD_H ? FCGI_RECORD_HEADER_SZ : 0); + } + else + max = buf_room_for_htx_data(buf); + + ret = max ? conn->xprt->rcv_buf(conn, conn->xprt_ctx, buf, max, 0) : 0; + + if (max && !ret && fcgi_recv_allowed(fconn)) { + TRACE_DATA("failed to receive data, subscribing", FCGI_EV_FCONN_RECV, conn); + conn->xprt->subscribe(conn, conn->xprt_ctx, SUB_RETRY_RECV, &fconn->wait_event); + } + else + TRACE_DATA("recv data", FCGI_EV_FCONN_RECV, conn, 0, 0, (size_t[]){ret}); + + if (conn_xprt_read0_pending(conn)) { + TRACE_DATA("received read0", FCGI_EV_FCONN_RECV, conn); + fconn->flags |= FCGI_CF_EOS; + } + if (conn->flags & CO_FL_ERROR) { + TRACE_DATA("connection error", FCGI_EV_FCONN_RECV, conn); + fconn->flags |= FCGI_CF_ERROR; + } + + if (!b_data(buf)) { + fcgi_release_buf(fconn, &fconn->dbuf); + goto end; + } + + if (ret == max) { + TRACE_DEVEL("fconn dbuf full", FCGI_EV_FCONN_RECV|FCGI_EV_FCONN_BLK, conn); + fconn->flags |= FCGI_CF_DEM_DFULL; + } + +end: + TRACE_LEAVE(FCGI_EV_FCONN_RECV, conn); + return !!ret || (fconn->flags & (FCGI_CF_EOS|FCGI_CF_ERROR)); +} + + +/* Try to send data if possible. + * The function returns 1 if data have been sent, otherwise zero. + */ +static int fcgi_send(struct fcgi_conn *fconn) +{ + struct connection *conn = fconn->conn; + int done; + int sent = 0; + + TRACE_ENTER(FCGI_EV_FCONN_SEND, conn); + + if (fconn->flags & (FCGI_CF_ERROR|FCGI_CF_ERR_PENDING)) { + TRACE_DEVEL("leaving on connection error", FCGI_EV_FCONN_SEND, conn); + if (fconn->flags & FCGI_CF_EOS) + fconn->flags |= FCGI_CF_ERROR; + b_reset(br_tail(fconn->mbuf)); + return 1; + } + + + if (conn->flags & CO_FL_WAIT_XPRT) { + /* a handshake was requested */ + goto schedule; + } + + /* This loop is quite simple : it tries to fill as much as it can from + * pending streams into the existing buffer until it's reportedly full + * or the end of send requests is reached. Then it tries to send this + * buffer's contents out, marks it not full if at least one byte could + * be sent, and tries again. + * + * The snd_buf() function normally takes a "flags" argument which may + * be made of a combination of CO_SFL_MSG_MORE to indicate that more + * data immediately comes and CO_SFL_STREAMER to indicate that the + * connection is streaming lots of data (used to increase TLS record + * size at the expense of latency). The former can be sent any time + * there's a buffer full flag, as it indicates at least one stream + * attempted to send and failed so there are pending data. An + * alternative would be to set it as long as there's an active stream + * but that would be problematic for ACKs until we have an absolute + * guarantee that all waiters have at least one byte to send. The + * latter should possibly not be set for now. + */ + + done = 0; + while (!done) { + unsigned int flags = 0; + unsigned int released = 0; + struct buffer *buf; + + /* fill as much as we can into the current buffer */ + while (((fconn->flags & (FCGI_CF_MUX_MFULL|FCGI_CF_MUX_MALLOC)) == 0) && !done) + done = fcgi_process_mux(fconn); + + if (fconn->flags & FCGI_CF_MUX_MALLOC) + done = 1; // we won't go further without extra buffers + + if (conn->flags & CO_FL_ERROR) + break; + + if (fconn->flags & (FCGI_CF_MUX_MFULL | FCGI_CF_DEM_MROOM)) + flags |= CO_SFL_MSG_MORE; + + for (buf = br_head(fconn->mbuf); b_size(buf); buf = br_del_head(fconn->mbuf)) { + if (b_data(buf)) { + int ret; + + ret = conn->xprt->snd_buf(conn, conn->xprt_ctx, buf, b_data(buf), flags); + if (!ret) { + done = 1; + break; + } + sent = 1; + TRACE_DATA("send data", FCGI_EV_FCONN_SEND, conn, 0, 0, (size_t[]){ret}); + b_del(buf, ret); + if (b_data(buf)) { + done = 1; + break; + } + } + b_free(buf); + released++; + } + + if (released) + offer_buffers(NULL, released); + + /* wrote at least one byte, the buffer is not full anymore */ + if (fconn->flags & (FCGI_CF_MUX_MFULL | FCGI_CF_DEM_MROOM)) + TRACE_STATE("fconn mbuf ring not fill anymore", FCGI_EV_FCONN_SEND|FCGI_EV_FCONN_BLK, conn); + fconn->flags &= ~(FCGI_CF_MUX_MFULL | FCGI_CF_DEM_MROOM); + } + + if (conn->flags & CO_FL_ERROR) { + fconn->flags |= FCGI_CF_ERR_PENDING; + if (fconn->flags & FCGI_CF_EOS) + fconn->flags |= FCGI_CF_ERROR; + b_reset(br_tail(fconn->mbuf)); + } + + /* We're not full anymore, so we can wake any task that are waiting + * for us. + */ + if (!(fconn->flags & (FCGI_CF_MUX_MFULL | FCGI_CF_DEM_MROOM)) && fconn->state >= FCGI_CS_RECORD_H) { + struct fcgi_strm *fstrm; + + list_for_each_entry(fstrm, &fconn->send_list, send_list) { + if (fconn->state == FCGI_CS_CLOSED || fconn->flags & FCGI_CF_MUX_BLOCK_ANY) + break; + + if (fstrm->flags & FCGI_SF_NOTIFIED) + continue; + + /* If the sender changed his mind and unsubscribed, let's just + * remove the stream from the send_list. + */ + if (!(fstrm->flags & (FCGI_SF_WANT_SHUTR|FCGI_SF_WANT_SHUTW)) && + (!fstrm->subs || !(fstrm->subs->events & SUB_RETRY_SEND))) { + LIST_DEL_INIT(&fstrm->send_list); + continue; + } + + if (fstrm->subs && fstrm->subs->events & SUB_RETRY_SEND) { + TRACE_DEVEL("waking up pending stream", FCGI_EV_FCONN_SEND|FCGI_EV_STRM_WAKE, conn, fstrm); + fstrm->flags &= ~FCGI_SF_BLK_ANY; + fstrm->flags |= FCGI_SF_NOTIFIED; + tasklet_wakeup(fstrm->subs->tasklet); + fstrm->subs->events &= ~SUB_RETRY_SEND; + if (!fstrm->subs->events) + fstrm->subs = NULL; + } else { + /* it's the shut request that was queued */ + TRACE_POINT(FCGI_EV_STRM_WAKE, fconn->conn, fstrm); + tasklet_wakeup(fstrm->shut_tl); + } + } + } + /* We're done, no more to send */ + if (!br_data(fconn->mbuf)) { + TRACE_DEVEL("leaving with everything sent", FCGI_EV_FCONN_SEND, conn); + goto end; + } +schedule: + if (!(conn->flags & CO_FL_ERROR) && !(fconn->wait_event.events & SUB_RETRY_SEND)) { + TRACE_STATE("more data to send, subscribing", FCGI_EV_FCONN_SEND, conn); + conn->xprt->subscribe(conn, conn->xprt_ctx, SUB_RETRY_SEND, &fconn->wait_event); + } + + TRACE_DEVEL("leaving with some data left to send", FCGI_EV_FCONN_SEND, conn); +end: + return sent || (fconn->flags & (FCGI_CF_ERR_PENDING|FCGI_CF_ERROR)); +} + +/* this is the tasklet referenced in fconn->wait_event.tasklet */ +struct task *fcgi_io_cb(struct task *t, void *ctx, unsigned int state) +{ + struct connection *conn; + struct fcgi_conn *fconn = ctx; + struct tasklet *tl = (struct tasklet *)t; + int conn_in_list; + int ret = 0; + + if (state & TASK_F_USR1) { + /* the tasklet was idling on an idle connection, it might have + * been stolen, let's be careful! + */ + HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + if (tl->context == NULL) { + /* The connection has been taken over by another thread, + * we're no longer responsible for it, so just free the + * tasklet, and do nothing. + */ + HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + tasklet_free(tl); + return NULL; + } + conn = fconn->conn; + TRACE_POINT(FCGI_EV_FCONN_WAKE, conn); + + conn_in_list = conn->flags & CO_FL_LIST_MASK; + if (conn_in_list) + conn_delete_from_tree(conn); + + HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + } else { + /* we're certain the connection was not in an idle list */ + conn = fconn->conn; + TRACE_ENTER(FCGI_EV_FCONN_WAKE, conn); + conn_in_list = 0; + } + + if (!(fconn->wait_event.events & SUB_RETRY_SEND)) + ret = fcgi_send(fconn); + if (!(fconn->wait_event.events & SUB_RETRY_RECV)) + ret |= fcgi_recv(fconn); + if (ret || b_data(&fconn->dbuf)) + ret = fcgi_process(fconn); + + /* If we were in an idle list, we want to add it back into it, + * unless fcgi_process() returned -1, which mean it has destroyed + * the connection (testing !ret is enough, if fcgi_process() wasn't + * called then ret will be 0 anyway. + */ + if (ret < 0) + t = NULL; + + if (!ret && conn_in_list) { + struct server *srv = objt_server(conn->target); + + HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + _srv_add_idle(srv, conn, conn_in_list == CO_FL_SAFE_LIST); + HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + } + return t; +} + +/* callback called on any event by the connection handler. + * It applies changes and returns zero, or < 0 if it wants immediate + * destruction of the connection (which normally doesn not happen in FCGI). + */ +static int fcgi_process(struct fcgi_conn *fconn) +{ + struct connection *conn = fconn->conn; + + TRACE_POINT(FCGI_EV_FCONN_WAKE, conn); + + if (b_data(&fconn->dbuf) && !(fconn->flags & FCGI_CF_DEM_BLOCK_ANY)) { + fcgi_process_demux(fconn); + + if (fconn->state == FCGI_CS_CLOSED || (fconn->flags & FCGI_CF_ERROR)) + b_reset(&fconn->dbuf); + + if (buf_room_for_htx_data(&fconn->dbuf)) + fconn->flags &= ~FCGI_CF_DEM_DFULL; + } + fcgi_send(fconn); + + if (unlikely(fconn->proxy->flags & (PR_FL_DISABLED|PR_FL_STOPPED))) { + /* frontend is stopping, reload likely in progress, let's try + * to announce a graceful shutdown if not yet done. We don't + * care if it fails, it will be tried again later. + */ + TRACE_STATE("proxy stopped, sending ABORT to all streams", FCGI_EV_FCONN_WAKE|FCGI_EV_TX_RECORD, conn); + if (!(fconn->flags & (FCGI_CF_ABRTS_SENT|FCGI_CF_ABRTS_FAILED))) { + if (fconn->stream_cnt - fconn->nb_reserved > 0) + fcgi_conn_send_aborts(fconn); + } + } + + /* + * If we received early data, and the handshake is done, wake + * any stream that was waiting for it. + */ + if (!(fconn->flags & FCGI_CF_WAIT_FOR_HS) && + (conn->flags & (CO_FL_EARLY_SSL_HS | CO_FL_WAIT_XPRT | CO_FL_EARLY_DATA)) == CO_FL_EARLY_DATA) { + struct eb32_node *node; + struct fcgi_strm *fstrm; + + fconn->flags |= FCGI_CF_WAIT_FOR_HS; + node = eb32_lookup_ge(&fconn->streams_by_id, 1); + + while (node) { + fstrm = container_of(node, struct fcgi_strm, by_id); + if (fcgi_strm_sc(fstrm) && se_fl_test(fstrm->sd, SE_FL_WAIT_FOR_HS)) + fcgi_strm_notify_recv(fstrm); + node = eb32_next(node); + } + } + + if ((fconn->flags & FCGI_CF_ERROR) || fcgi_conn_read0_pending(fconn) || + fconn->state == FCGI_CS_CLOSED || (fconn->flags & FCGI_CF_ABRTS_FAILED) || + eb_is_empty(&fconn->streams_by_id)) { + fcgi_wake_some_streams(fconn, 0); + + if (eb_is_empty(&fconn->streams_by_id)) { + /* no more stream, kill the connection now */ + fcgi_release(fconn); + TRACE_DEVEL("leaving after releasing the connection", FCGI_EV_FCONN_WAKE); + return -1; + } + } + + if (!b_data(&fconn->dbuf)) + fcgi_release_buf(fconn, &fconn->dbuf); + + if (fconn->state == FCGI_CS_CLOSED || (fconn->flags & FCGI_CF_ABRTS_FAILED) || + (!br_data(fconn->mbuf) && ((fconn->flags & FCGI_CF_MUX_BLOCK_ANY) || LIST_ISEMPTY(&fconn->send_list)))) + fcgi_release_mbuf(fconn); + + if (fconn->task) { + fconn->task->expire = tick_add(now_ms, (fconn->state == FCGI_CS_CLOSED ? fconn->shut_timeout : fconn->timeout)); + task_queue(fconn->task); + } + + fcgi_send(fconn); + TRACE_LEAVE(FCGI_EV_FCONN_WAKE, conn); + return 0; +} + + +/* wake-up function called by the connection layer (mux_ops.wake) */ +static int fcgi_wake(struct connection *conn) +{ + struct fcgi_conn *fconn = conn->ctx; + + TRACE_POINT(FCGI_EV_FCONN_WAKE, conn); + return (fcgi_process(fconn)); +} + + +static int fcgi_ctl(struct connection *conn, enum mux_ctl_type mux_ctl, void *output) +{ + int ret = 0; + switch (mux_ctl) { + case MUX_CTL_STATUS: + if (!(conn->flags & CO_FL_WAIT_XPRT)) + ret |= MUX_STATUS_READY; + return ret; + case MUX_CTL_EXIT_STATUS: + return MUX_ES_UNKNOWN; + default: + return -1; + } +} + +static int fcgi_sctl(struct stconn *sc, enum mux_sctl_type mux_sctl, void *output) +{ + int ret = 0; + struct fcgi_strm *fstrm = __sc_mux_strm(sc); + + switch (mux_sctl) { + case MUX_SCTL_SID: + if (output) + *((int64_t *)output) = fstrm->id; + return ret; + + default: + return -1; + } +} + +/* Connection timeout management. The principle is that if there's no receipt + * nor sending for a certain amount of time, the connection is closed. If the + * MUX buffer still has lying data or is not allocatable, the connection is + * immediately killed. If it's allocatable and empty, we attempt to send a + * ABORT records. + */ +struct task *fcgi_timeout_task(struct task *t, void *context, unsigned int state) +{ + struct fcgi_conn *fconn = context; + int expired = tick_is_expired(t->expire, now_ms); + + TRACE_ENTER(FCGI_EV_FCONN_WAKE, (fconn ? fconn->conn : NULL)); + + if (fconn) { + HA_SPIN_LOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + + /* Somebody already stole the connection from us, so we should not + * free it, we just have to free the task. + */ + if (!t->context) { + HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + fconn = NULL; + goto do_leave; + } + + if (!expired) { + HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + TRACE_DEVEL("leaving (not expired)", FCGI_EV_FCONN_WAKE, fconn->conn); + return t; + } + + /* We're about to destroy the connection, so make sure nobody attempts + * to steal it from us. + */ + if (fconn->conn->flags & CO_FL_LIST_MASK) + conn_delete_from_tree(fconn->conn); + + HA_SPIN_UNLOCK(IDLE_CONNS_LOCK, &idle_conns[tid].idle_conns_lock); + } + +do_leave: + task_destroy(t); + + if (!fconn) { + /* resources were already deleted */ + TRACE_DEVEL("leaving (not more fconn)", FCGI_EV_FCONN_WAKE); + return NULL; + } + + fconn->task = NULL; + fconn->state = FCGI_CS_CLOSED; + fcgi_wake_some_streams(fconn, 0); + + if (br_data(fconn->mbuf)) { + /* don't even try to send aborts, the buffer is stuck */ + fconn->flags |= FCGI_CF_ABRTS_FAILED; + goto end; + } + + /* try to send but no need to insist */ + if (!fcgi_conn_send_aborts(fconn)) + fconn->flags |= FCGI_CF_ABRTS_FAILED; + + if (br_data(fconn->mbuf) && !(fconn->flags & FCGI_CF_ABRTS_FAILED) && + conn_xprt_ready(fconn->conn)) { + unsigned int released = 0; + struct buffer *buf; + + for (buf = br_head(fconn->mbuf); b_size(buf); buf = br_del_head(fconn->mbuf)) { + if (b_data(buf)) { + int ret = fconn->conn->xprt->snd_buf(fconn->conn, fconn->conn->xprt_ctx, + buf, b_data(buf), 0); + if (!ret) + break; + b_del(buf, ret); + if (b_data(buf)) + break; + b_free(buf); + released++; + } + } + + if (released) + offer_buffers(NULL, released); + } + + end: + /* either we can release everything now or it will be done later once + * the last stream closes. + */ + if (eb_is_empty(&fconn->streams_by_id)) + fcgi_release(fconn); + + TRACE_LEAVE(FCGI_EV_FCONN_WAKE); + return NULL; +} + + +/*******************************************/ +/* functions below are used by the streams */ +/*******************************************/ + +/* Append the description of what is present in error snapshot <es> into <out>. + * The description must be small enough to always fit in a buffer. The output + * buffer may be the trash so the trash must not be used inside this function. + */ +static void fcgi_show_error_snapshot(struct buffer *out, const struct error_snapshot *es) +{ + chunk_appendf(out, + " FCGI connection flags 0x%08x, FCGI stream flags 0x%08x\n" + " H1 msg state %s(%d), H1 msg flags 0x%08x\n" + " H1 chunk len %lld bytes, H1 body len %lld bytes :\n", + es->ctx.h1.c_flags, es->ctx.h1.s_flags, + h1m_state_str(es->ctx.h1.state), es->ctx.h1.state, + es->ctx.h1.m_flags, es->ctx.h1.m_clen, es->ctx.h1.m_blen); +} +/* + * Capture a bad response and archive it in the proxy's structure. By default + * it tries to report the error position as h1m->err_pos. However if this one is + * not set, it will then report h1m->next, which is the last known parsing + * point. The function is able to deal with wrapping buffers. It always displays + * buffers as a contiguous area starting at buf->p. The direction is determined + * thanks to the h1m's flags. + */ +static void fcgi_strm_capture_bad_message(struct fcgi_conn *fconn, struct fcgi_strm *fstrm, + struct h1m *h1m, struct buffer *buf) +{ + struct session *sess = fstrm->sess; + struct proxy *proxy = fconn->proxy; + struct proxy *other_end; + union error_snapshot_ctx ctx; + + if (fcgi_strm_sc(fstrm) && sc_strm(fcgi_strm_sc(fstrm))) { + if (sess == NULL) + sess = __sc_strm(fcgi_strm_sc(fstrm))->sess; + if (!(h1m->flags & H1_MF_RESP)) + other_end = __sc_strm(fcgi_strm_sc(fstrm))->be; + else + other_end = sess->fe; + } else + other_end = NULL; + /* http-specific part now */ + ctx.h1.state = h1m->state; + ctx.h1.c_flags = fconn->flags; + ctx.h1.s_flags = fstrm->flags; + ctx.h1.m_flags = h1m->flags; + ctx.h1.m_clen = h1m->curr_len; + ctx.h1.m_blen = h1m->body_len; + + proxy_capture_error(proxy, 1, other_end, fconn->conn->target, sess, buf, 0, 0, + (h1m->err_pos >= 0) ? h1m->err_pos : h1m->next, + &ctx, fcgi_show_error_snapshot); +} + +static size_t fcgi_strm_parse_headers(struct fcgi_strm *fstrm, struct h1m *h1m, struct htx *htx, + struct buffer *buf, size_t *ofs, size_t max) +{ + int ret; + + TRACE_ENTER(FCGI_EV_RSP_DATA|FCGI_EV_RSP_HDRS, fstrm->fconn->conn, fstrm, 0, (size_t[]){max}); + ret = h1_parse_msg_hdrs(h1m, NULL, htx, buf, *ofs, max); + if (ret <= 0) { + TRACE_DEVEL("leaving on missing data or error", FCGI_EV_RSP_DATA|FCGI_EV_RSP_HDRS, fstrm->fconn->conn, fstrm); + if (htx->flags & HTX_FL_PARSING_ERROR) { + TRACE_ERROR("parsing error, reject H1 response", FCGI_EV_RSP_DATA|FCGI_EV_RSP_HDRS|FCGI_EV_FSTRM_ERR, fstrm->fconn->conn, fstrm); + fcgi_strm_error(fstrm); + fcgi_strm_capture_bad_message(fstrm->fconn, fstrm, h1m, buf); + } + ret = 0; + goto end; + } + + /* Reject any message with an unknown transfer-encoding. In fact if any + * encoding other than "chunked". A 422-Unprocessable-Content is + * returned for an invalid request, a 502-Bad-Gateway for an invalid + * response. + */ + if (h1m->flags & H1_MF_TE_OTHER) { + htx->flags |= HTX_FL_PARSING_ERROR; + TRACE_ERROR("Unknown transfer-encoding", FCGI_EV_RSP_DATA|FCGI_EV_RSP_HDRS|FCGI_EV_FSTRM_ERR, fstrm->fconn->conn, fstrm); + fcgi_strm_error(fstrm); + fcgi_strm_capture_bad_message(fstrm->fconn, fstrm, h1m, buf); + ret = 0; + goto end; + } + + *ofs += ret; + end: + TRACE_LEAVE(FCGI_EV_RSP_DATA|FCGI_EV_RSP_HDRS, fstrm->fconn->conn, fstrm, 0, (size_t[]){ret}); + return ret; + +} + +static size_t fcgi_strm_parse_data(struct fcgi_strm *fstrm, struct h1m *h1m, struct htx **htx, + struct buffer *buf, size_t *ofs, size_t max, struct buffer *htxbuf) +{ + size_t ret; + + TRACE_ENTER(FCGI_EV_RSP_DATA|FCGI_EV_RSP_BODY, fstrm->fconn->conn, fstrm, 0, (size_t[]){max}); + ret = h1_parse_msg_data(h1m, htx, buf, *ofs, max, htxbuf); + if (!ret) { + TRACE_DEVEL("leaving on missing data or error", FCGI_EV_RSP_DATA|FCGI_EV_RSP_BODY, fstrm->fconn->conn, fstrm); + if ((*htx)->flags & HTX_FL_PARSING_ERROR) { + TRACE_ERROR("parsing error, reject H1 response", FCGI_EV_RSP_DATA|FCGI_EV_RSP_BODY|FCGI_EV_FSTRM_ERR, fstrm->fconn->conn, fstrm); + fcgi_strm_error(fstrm); + fcgi_strm_capture_bad_message(fstrm->fconn, fstrm, h1m, buf); + } + goto end; + } + *ofs += ret; + end: + TRACE_LEAVE(FCGI_EV_RSP_DATA|FCGI_EV_RSP_BODY, fstrm->fconn->conn, fstrm, 0, (size_t[]){ret}); + return ret; +} + +static size_t fcgi_strm_parse_trailers(struct fcgi_strm *fstrm, struct h1m *h1m, struct htx *htx, + struct buffer *buf, size_t *ofs, size_t max) +{ + int ret; + + TRACE_ENTER(FCGI_EV_RSP_DATA|FCGI_EV_RSP_TLRS, fstrm->fconn->conn, fstrm, 0, (size_t[]){max}); + ret = h1_parse_msg_tlrs(h1m, htx, buf, *ofs, max); + if (ret <= 0) { + TRACE_DEVEL("leaving on missing data or error", FCGI_EV_RSP_DATA|FCGI_EV_RSP_TLRS, fstrm->fconn->conn, fstrm); + if (htx->flags & HTX_FL_PARSING_ERROR) { + TRACE_ERROR("parsing error, reject H1 response", FCGI_EV_RSP_DATA|FCGI_EV_RSP_TLRS|FCGI_EV_FSTRM_ERR, fstrm->fconn->conn, fstrm); + fcgi_strm_error(fstrm); + fcgi_strm_capture_bad_message(fstrm->fconn, fstrm, h1m, buf); + } + ret = 0; + goto end; + } + *ofs += ret; + end: + TRACE_LEAVE(FCGI_EV_RSP_DATA|FCGI_EV_RSP_TLRS, fstrm->fconn->conn, fstrm, 0, (size_t[]){ret}); + return ret; +} + +static size_t fcgi_strm_parse_response(struct fcgi_strm *fstrm, struct buffer *buf, size_t count) +{ + struct fcgi_conn *fconn = fstrm->fconn; + struct htx *htx; + struct h1m *h1m = &fstrm->h1m; + size_t ret, data, total = 0; + + htx = htx_from_buf(buf); + TRACE_ENTER(FCGI_EV_RSP_DATA, fconn->conn, fstrm, htx, (size_t[]){count}); + + data = htx->data; + if (fstrm->state == FCGI_SS_ERROR) + goto end; + + do { + size_t used = htx_used_space(htx); + + if (h1m->state <= H1_MSG_LAST_LF) { + TRACE_PROTO("parsing response headers", FCGI_EV_RSP_DATA|FCGI_EV_RSP_HDRS, fconn->conn, fstrm); + ret = fcgi_strm_parse_headers(fstrm, h1m, htx, &fstrm->rxbuf, &total, count); + if (!ret) + break; + + TRACE_USER("rcvd H1 response headers", FCGI_EV_RSP_DATA|FCGI_EV_RSP_HDRS, fconn->conn, fstrm, htx); + + if ((h1m->flags & (H1_MF_VER_11|H1_MF_XFER_LEN)) == H1_MF_VER_11) { + struct htx_blk *blk = htx_get_head_blk(htx); + struct htx_sl *sl; + + if (!blk) + break; + sl = htx_get_blk_ptr(htx, blk); + sl->flags |= HTX_SL_F_XFER_LEN; + htx->extra = 0; + } + } + else if (h1m->state < H1_MSG_TRAILERS) { + TRACE_PROTO("parsing response payload", FCGI_EV_RSP_DATA|FCGI_EV_RSP_BODY, fconn->conn, fstrm); + fcgi_strm_parse_data(fstrm, h1m, &htx, &fstrm->rxbuf, &total, count, buf); + + if (!(h1m->flags & H1_MF_XFER_LEN) && fstrm->state != FCGI_SS_ERROR && + (fstrm->flags & FCGI_SF_ES_RCVD) && b_data(&fstrm->rxbuf) == total) { + TRACE_DEVEL("end of data", FCGI_EV_RSP_DATA, fconn->conn, fstrm); + if (htx_is_empty(htx) && !htx_add_endof(htx, HTX_BLK_EOT)) + break; + htx->flags |= HTX_FL_EOM; + h1m->state = H1_MSG_DONE; + TRACE_USER("H1 response fully rcvd", FCGI_EV_RSP_DATA|FCGI_EV_RSP_EOM, fconn->conn, fstrm, htx); + } + + if (h1m->state < H1_MSG_TRAILERS) + break; + + TRACE_PROTO("rcvd response payload data", FCGI_EV_RSP_DATA|FCGI_EV_RSP_BODY, fconn->conn, fstrm, htx); + } + else if (h1m->state == H1_MSG_TRAILERS) { + TRACE_PROTO("parsing response trailers", FCGI_EV_RSP_DATA|FCGI_EV_RSP_TLRS, fconn->conn, fstrm); + fcgi_strm_parse_trailers(fstrm, h1m, htx, &fstrm->rxbuf, &total, count); + if (h1m->state != H1_MSG_DONE) + break; + + TRACE_PROTO("rcvd H1 response trailers", FCGI_EV_RSP_DATA|FCGI_EV_RSP_TLRS, fconn->conn, fstrm, htx); + } + else if (h1m->state == H1_MSG_DONE) { + TRACE_USER("H1 response fully rcvd", FCGI_EV_RSP_DATA|FCGI_EV_RSP_EOM, fconn->conn, fstrm, htx); + if (b_data(&fstrm->rxbuf) > total) { + htx->flags |= HTX_FL_PARSING_ERROR; + TRACE_PROTO("too much data, parsing error", FCGI_EV_RSP_DATA, fconn->conn, fstrm); + fcgi_strm_error(fstrm); + } + break; + } + else { + htx->flags |= HTX_FL_PROCESSING_ERROR; + TRACE_ERROR("unexpected processing error", FCGI_EV_RSP_DATA|FCGI_EV_STRM_ERR, fconn->conn, fstrm); + fcgi_strm_error(fstrm); + break; + } + + count -= htx_used_space(htx) - used; + } while (fstrm->state != FCGI_SS_ERROR); + + if (fstrm->state == FCGI_SS_ERROR) { + b_reset(&fstrm->rxbuf); + htx_to_buf(htx, buf); + TRACE_DEVEL("leaving on error", FCGI_EV_RSP_DATA|FCGI_EV_STRM_ERR, fconn->conn, fstrm); + return 0; + } + + b_del(&fstrm->rxbuf, total); + + end: + htx_to_buf(htx, buf); + ret = htx->data - data; + TRACE_LEAVE(FCGI_EV_RSP_DATA, fconn->conn, fstrm, htx, (size_t[]){ret}); + return ret; +} + +/* + * Attach a new stream to a connection + * (Used for outgoing connections) + */ +static int fcgi_attach(struct connection *conn, struct sedesc *sd, struct session *sess) +{ + struct fcgi_strm *fstrm; + struct fcgi_conn *fconn = conn->ctx; + + TRACE_ENTER(FCGI_EV_FSTRM_NEW, conn); + fstrm = fcgi_stconn_new(fconn, sd->sc, sess); + if (!fstrm) + goto err; + + /* the connection is not idle anymore, let's mark this */ + HA_ATOMIC_AND(&fconn->wait_event.tasklet->state, ~TASK_F_USR1); + xprt_set_used(conn, conn->xprt, conn->xprt_ctx); + + TRACE_LEAVE(FCGI_EV_FSTRM_NEW, conn, fstrm); + return 0; + + err: + TRACE_DEVEL("leaving on error", FCGI_EV_FSTRM_NEW|FCGI_EV_FSTRM_ERR, conn); + return -1; +} + +/* Retrieves the first valid stream connector from this connection, or returns NULL. + * We have to scan because we may have some orphan streams. It might be + * beneficial to scan backwards from the end to reduce the likeliness to find + * orphans. + */ +static struct stconn *fcgi_get_first_sc(const struct connection *conn) +{ + struct fcgi_conn *fconn = conn->ctx; + struct fcgi_strm *fstrm; + struct eb32_node *node; + + node = eb32_first(&fconn->streams_by_id); + while (node) { + fstrm = container_of(node, struct fcgi_strm, by_id); + if (fcgi_strm_sc(fstrm)) + return fcgi_strm_sc(fstrm); + node = eb32_next(node); + } + return NULL; +} + +/* + * Destroy the mux and the associated connection, if it is no longer used + */ +static void fcgi_destroy(void *ctx) +{ + struct fcgi_conn *fconn = ctx; + + TRACE_POINT(FCGI_EV_FCONN_END, fconn->conn); + if (eb_is_empty(&fconn->streams_by_id)) { + BUG_ON(fconn->conn->ctx != fconn); + fcgi_release(fconn); + } +} + +/* + * Detach the stream from the connection and possibly release the connection. + */ +static void fcgi_detach(struct sedesc *sd) +{ + struct fcgi_strm *fstrm = sd->se; + struct fcgi_conn *fconn; + struct session *sess; + + TRACE_ENTER(FCGI_EV_STRM_END, (fstrm ? fstrm->fconn->conn : NULL), fstrm); + + if (!fstrm) { + TRACE_LEAVE(FCGI_EV_STRM_END); + return; + } + + /* there's no txbuf so we're certain no to be able to send anything */ + fstrm->flags &= ~FCGI_SF_NOTIFIED; + + sess = fstrm->sess; + fconn = fstrm->fconn; + fconn->nb_sc--; + + if (fstrm->proto_status == FCGI_PS_CANT_MPX_CONN) { + fconn->flags &= ~FCGI_CF_MPXS_CONNS; + fconn->streams_limit = 1; + } + else if (fstrm->proto_status == FCGI_PS_OVERLOADED || + fstrm->proto_status == FCGI_PS_UNKNOWN_ROLE) { + fconn->flags &= ~FCGI_CF_KEEP_CONN; + fconn->state = FCGI_CS_CLOSED; + } + + /* this stream may be blocked waiting for some data to leave, so orphan + * it in this case. + */ + if (!(fconn->flags & (FCGI_CF_ERR_PENDING|FCGI_CF_ERROR)) && // FIXME: Be sure for ERR_PENDING + (fconn->state != FCGI_CS_CLOSED) && + (fstrm->flags & (FCGI_SF_BLK_MBUSY|FCGI_SF_BLK_MROOM)) && + (fstrm->subs || (fstrm->flags & (FCGI_SF_WANT_SHUTR|FCGI_SF_WANT_SHUTW)))) { + TRACE_DEVEL("leaving on stream blocked", FCGI_EV_STRM_END|FCGI_EV_FSTRM_BLK, fconn->conn, fstrm); + return; + } + + if ((fconn->flags & FCGI_CF_DEM_BLOCK_ANY && fstrm->id == fconn->dsi)) { + /* unblock the connection if it was blocked on this stream. */ + fconn->flags &= ~FCGI_CF_DEM_BLOCK_ANY; + fcgi_conn_restart_reading(fconn, 1); + } + + fcgi_strm_destroy(fstrm); + + if (!(fconn->flags & (FCGI_CF_EOS|FCGI_CF_ERR_PENDING|FCGI_CF_ERROR)) && + (fconn->flags & FCGI_CF_KEEP_CONN)) { + if (fconn->conn->flags & CO_FL_PRIVATE) { + /* Add the connection in the session serverlist, if not already done */ + if (!session_add_conn(sess, fconn->conn, fconn->conn->target)) { + fconn->conn->owner = NULL; + if (eb_is_empty(&fconn->streams_by_id)) { + /* let's kill the connection right away */ + fconn->conn->mux->destroy(fconn); + TRACE_DEVEL("outgoing connection killed", FCGI_EV_STRM_END|FCGI_EV_FCONN_ERR); + return; + } + } + if (eb_is_empty(&fconn->streams_by_id)) { + if (session_check_idle_conn(fconn->conn->owner, fconn->conn) != 0) { + /* The connection is destroyed, let's leave */ + TRACE_DEVEL("outgoing connection killed", FCGI_EV_STRM_END|FCGI_EV_FCONN_ERR); + return; + } + } + } + else { + if (eb_is_empty(&fconn->streams_by_id)) { + /* If the connection is owned by the session, first remove it + * from its list + */ + if (fconn->conn->owner) { + session_unown_conn(fconn->conn->owner, fconn->conn); + fconn->conn->owner = NULL; + } + + /* mark that the tasklet may lose its context to another thread and + * that the handler needs to check it under the idle conns lock. + */ + HA_ATOMIC_OR(&fconn->wait_event.tasklet->state, TASK_F_USR1); + xprt_set_idle(fconn->conn, fconn->conn->xprt, fconn->conn->xprt_ctx); + + if (!srv_add_to_idle_list(objt_server(fconn->conn->target), fconn->conn, 1)) { + /* The server doesn't want it, let's kill the connection right away */ + fconn->conn->mux->destroy(fconn); + TRACE_DEVEL("outgoing connection killed", FCGI_EV_STRM_END|FCGI_EV_FCONN_ERR); + return; + } + /* At this point, the connection has been added to the + * server idle list, so another thread may already have + * hijacked it, so we can't do anything with it. + */ + TRACE_DEVEL("reusable idle connection", FCGI_EV_STRM_END, fconn->conn); + return; + } + else if (!fconn->conn->hash_node->node.node.leaf_p && + fcgi_avail_streams(fconn->conn) > 0 && objt_server(fconn->conn->target) && + !LIST_INLIST(&fconn->conn->session_list)) { + srv_add_to_avail_list(__objt_server(fconn->conn->target), fconn->conn); + } + } + } + + /* We don't want to close right now unless we're removing the last + * stream and the connection is in error. + */ + if (fcgi_conn_is_dead(fconn)) { + /* no more stream will come, kill it now */ + TRACE_DEVEL("leaving, killing dead connection", FCGI_EV_STRM_END, fconn->conn); + fcgi_release(fconn); + } + else if (fconn->task) { + fconn->task->expire = tick_add(now_ms, (fconn->state == FCGI_CS_CLOSED ? fconn->shut_timeout : fconn->timeout)); + task_queue(fconn->task); + TRACE_DEVEL("leaving, refreshing connection's timeout", FCGI_EV_STRM_END, fconn->conn); + } + else + TRACE_DEVEL("leaving", FCGI_EV_STRM_END, fconn->conn); +} + + +/* Performs a synchronous or asynchronous shutr(). */ +static void fcgi_do_shutr(struct fcgi_strm *fstrm) +{ + struct fcgi_conn *fconn = fstrm->fconn; + + TRACE_ENTER(FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + + if (fstrm->state == FCGI_SS_CLOSED) + goto done; + + /* a connstream may require us to immediately kill the whole connection + * for example because of a "tcp-request content reject" rule that is + * normally used to limit abuse. + */ + if (se_fl_test(fstrm->sd, SE_FL_KILL_CONN) && + !(fconn->flags & (FCGI_CF_ABRTS_SENT|FCGI_CF_ABRTS_FAILED))) { + TRACE_STATE("stream wants to kill the connection", FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + fconn->state = FCGI_CS_CLOSED; + } + else if (fstrm->flags & FCGI_SF_BEGIN_SENT) { + TRACE_STATE("no headers sent yet, trying a retryable abort", FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + if (!(fstrm->flags & (FCGI_SF_ES_SENT|FCGI_SF_ABRT_SENT)) && + !fcgi_strm_send_abort(fconn, fstrm)) + goto add_to_list; + } + + fcgi_strm_close(fstrm); + + if (!(fconn->wait_event.events & SUB_RETRY_SEND)) + tasklet_wakeup(fconn->wait_event.tasklet); + done: + fstrm->flags &= ~FCGI_SF_WANT_SHUTR; + TRACE_LEAVE(FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + return; + + add_to_list: + /* Let the handler know we want to shutr, and add ourselves to the + * send list if not yet done. fcgi_deferred_shut() will be + * automatically called via the shut_tl tasklet when there's room + * again. + */ + if (!LIST_INLIST(&fstrm->send_list)) { + if (fstrm->flags & (FCGI_SF_BLK_MBUSY|FCGI_SF_BLK_MROOM)) { + LIST_APPEND(&fconn->send_list, &fstrm->send_list); + } + } + fstrm->flags |= FCGI_SF_WANT_SHUTR; + TRACE_LEAVE(FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + return; +} + +/* Performs a synchronous or asynchronous shutw(). */ +static void fcgi_do_shutw(struct fcgi_strm *fstrm) +{ + struct fcgi_conn *fconn = fstrm->fconn; + + TRACE_ENTER(FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + + if (fstrm->state != FCGI_SS_HLOC || fstrm->state == FCGI_SS_CLOSED) + goto done; + + if (fstrm->state != FCGI_SS_ERROR && (fstrm->flags & FCGI_SF_BEGIN_SENT)) { + if (!(fstrm->flags & (FCGI_SF_ES_SENT|FCGI_SF_ABRT_SENT)) && + !fcgi_strm_send_abort(fconn, fstrm)) + goto add_to_list; + + if (fstrm->state == FCGI_SS_HREM) + fcgi_strm_close(fstrm); + else + fstrm->state = FCGI_SS_HLOC; + } else { + /* a connstream may require us to immediately kill the whole connection + * for example because of a "tcp-request content reject" rule that is + * normally used to limit abuse. + */ + if (se_fl_test(fstrm->sd, SE_FL_KILL_CONN) && + !(fconn->flags & (FCGI_CF_ABRTS_SENT|FCGI_CF_ABRTS_FAILED))) { + TRACE_STATE("stream wants to kill the connection", FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + fconn->state = FCGI_CS_CLOSED; + } + + fcgi_strm_close(fstrm); + } + + if (!(fconn->wait_event.events & SUB_RETRY_SEND)) + tasklet_wakeup(fconn->wait_event.tasklet); + done: + fstrm->flags &= ~FCGI_SF_WANT_SHUTW; + TRACE_LEAVE(FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + return; + + add_to_list: + /* Let the handler know we want to shutr, and add ourselves to the + * send list if not yet done. fcgi_deferred_shut() will be + * automatically called via the shut_tl tasklet when there's room + * again. + */ + if (!LIST_INLIST(&fstrm->send_list)) { + if (fstrm->flags & (FCGI_SF_BLK_MBUSY|FCGI_SF_BLK_MROOM)) { + LIST_APPEND(&fconn->send_list, &fstrm->send_list); + } + } + fstrm->flags |= FCGI_SF_WANT_SHUTW; + TRACE_LEAVE(FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + return; +} + +/* This is the tasklet referenced in fstrm->shut_tl, it is used for + * deferred shutdowns when the fcgi_detach() was done but the mux buffer was full + * and prevented the last record from being emitted. + */ +struct task *fcgi_deferred_shut(struct task *t, void *ctx, unsigned int state) +{ + struct fcgi_strm *fstrm = ctx; + struct fcgi_conn *fconn = fstrm->fconn; + + TRACE_ENTER(FCGI_EV_STRM_SHUT, fconn->conn, fstrm); + + if (fstrm->flags & FCGI_SF_NOTIFIED) { + /* some data processing remains to be done first */ + goto end; + } + + if (fstrm->flags & FCGI_SF_WANT_SHUTW) + fcgi_do_shutw(fstrm); + + if (fstrm->flags & FCGI_SF_WANT_SHUTR) + fcgi_do_shutr(fstrm); + + if (!(fstrm->flags & (FCGI_SF_WANT_SHUTR|FCGI_SF_WANT_SHUTW))) { + /* We're done trying to send, remove ourself from the send_list */ + LIST_DEL_INIT(&fstrm->send_list); + + if (!fcgi_strm_sc(fstrm)) { + fcgi_strm_destroy(fstrm); + if (fcgi_conn_is_dead(fconn)) + fcgi_release(fconn); + } + } + end: + TRACE_LEAVE(FCGI_EV_STRM_SHUT); + return NULL; +} + +/* shutr() called by the stream connector (mux_ops.shutr) */ +static void fcgi_shutr(struct stconn *sc, enum co_shr_mode mode) +{ + struct fcgi_strm *fstrm = __sc_mux_strm(sc); + + TRACE_POINT(FCGI_EV_STRM_SHUT, fstrm->fconn->conn, fstrm); + if (!mode) + return; + fcgi_do_shutr(fstrm); +} + +/* shutw() called by the stream connector (mux_ops.shutw) */ +static void fcgi_shutw(struct stconn *sc, enum co_shw_mode mode) +{ + struct fcgi_strm *fstrm = __sc_mux_strm(sc); + + TRACE_POINT(FCGI_EV_STRM_SHUT, fstrm->fconn->conn, fstrm); + fcgi_do_shutw(fstrm); +} + +/* Called from the upper layer, to subscribe <es> to events <event_type>. The + * event subscriber <es> is not allowed to change from a previous call as long + * as at least one event is still subscribed. The <event_type> must only be a + * combination of SUB_RETRY_RECV and SUB_RETRY_SEND. It always returns 0. + */ +static int fcgi_subscribe(struct stconn *sc, int event_type, struct wait_event *es) +{ + struct fcgi_strm *fstrm = __sc_mux_strm(sc); + struct fcgi_conn *fconn = fstrm->fconn; + + BUG_ON(event_type & ~(SUB_RETRY_SEND|SUB_RETRY_RECV)); + BUG_ON(fstrm->subs && fstrm->subs != es); + + es->events |= event_type; + fstrm->subs = es; + + if (event_type & SUB_RETRY_RECV) + TRACE_DEVEL("unsubscribe(recv)", FCGI_EV_STRM_RECV, fconn->conn, fstrm); + + if (event_type & SUB_RETRY_SEND) { + TRACE_DEVEL("unsubscribe(send)", FCGI_EV_STRM_SEND, fconn->conn, fstrm); + if (!LIST_INLIST(&fstrm->send_list)) + LIST_APPEND(&fconn->send_list, &fstrm->send_list); + } + return 0; +} + +/* Called from the upper layer, to unsubscribe <es> from events <event_type> + * (undo fcgi_subscribe). The <es> pointer is not allowed to differ from the one + * passed to the subscribe() call. It always returns zero. + */ +static int fcgi_unsubscribe(struct stconn *sc, int event_type, struct wait_event *es) +{ + struct fcgi_strm *fstrm = __sc_mux_strm(sc); + struct fcgi_conn *fconn = fstrm->fconn; + + BUG_ON(event_type & ~(SUB_RETRY_SEND|SUB_RETRY_RECV)); + BUG_ON(fstrm->subs && fstrm->subs != es); + + es->events &= ~event_type; + if (!es->events) + fstrm->subs = NULL; + + if (event_type & SUB_RETRY_RECV) + TRACE_DEVEL("subscribe(recv)", FCGI_EV_STRM_RECV, fconn->conn, fstrm); + + if (event_type & SUB_RETRY_SEND) { + TRACE_DEVEL("subscribe(send)", FCGI_EV_STRM_SEND, fconn->conn, fstrm); + fstrm->flags &= ~FCGI_SF_NOTIFIED; + if (!(fstrm->flags & (FCGI_SF_WANT_SHUTR|FCGI_SF_WANT_SHUTW))) + LIST_DEL_INIT(&fstrm->send_list); + } + return 0; +} + +/* Called from the upper layer, to receive data + * + * The caller is responsible for defragmenting <buf> if necessary. But <flags> + * must be tested to know the calling context. If CO_RFL_BUF_FLUSH is set, it + * means the caller wants to flush input data (from the mux buffer and the + * channel buffer) to be able to use kernel splicing or any kind of mux-to-mux + * xfer. If CO_RFL_KEEP_RECV is set, the mux must always subscribe for read + * events before giving back. CO_RFL_BUF_WET is set if <buf> is congested with + * data scheduled for leaving soon. CO_RFL_BUF_NOT_STUCK is set to instruct the + * mux it may optimize the data copy to <buf> if necessary. Otherwise, it should + * copy as much data as possible. + */ +static size_t fcgi_rcv_buf(struct stconn *sc, struct buffer *buf, size_t count, int flags) +{ + struct fcgi_strm *fstrm = __sc_mux_strm(sc); + struct fcgi_conn *fconn = fstrm->fconn; + size_t ret = 0; + + TRACE_ENTER(FCGI_EV_STRM_RECV, fconn->conn, fstrm); + + if (!(fconn->flags & FCGI_CF_DEM_SALLOC)) + ret = fcgi_strm_parse_response(fstrm, buf, count); + else + TRACE_STATE("fstrm rxbuf not allocated", FCGI_EV_STRM_RECV|FCGI_EV_FSTRM_BLK, fconn->conn, fstrm); + + if (b_data(&fstrm->rxbuf)) { + /* If the channel buffer is not empty, consider the mux is + * blocked because it needs more room. But if the channel buffer + * is empty, it means partial data were received and the mux + * needs to receive more data to be able to parse it. + */ + if (b_data(buf)) + se_fl_set(fstrm->sd, SE_FL_RCV_MORE | SE_FL_WANT_ROOM); + } + else { + se_fl_clr(fstrm->sd, SE_FL_RCV_MORE | SE_FL_WANT_ROOM); + if (fstrm->state == FCGI_SS_ERROR || (fstrm->h1m.state == H1_MSG_DONE)) { + se_fl_set(fstrm->sd, SE_FL_EOI); + if (!(fstrm->h1m.flags & (H1_MF_VER_11|H1_MF_XFER_LEN))) + se_fl_set(fstrm->sd, SE_FL_EOS); + } + if (fcgi_conn_read0_pending(fconn)) { + se_fl_set(fstrm->sd, SE_FL_EOS); + if (!se_fl_test(fstrm->sd, SE_FL_EOI)) + se_fl_set(fstrm->sd, SE_FL_ERROR); + } + if (se_fl_test(fstrm->sd, SE_FL_ERR_PENDING)) + se_fl_set(fstrm->sd, SE_FL_ERROR); + fcgi_release_buf(fconn, &fstrm->rxbuf); + } + + if (ret && fconn->dsi == fstrm->id) { + /* demux is blocking on this stream's buffer */ + fconn->flags &= ~FCGI_CF_DEM_SFULL; + fcgi_conn_restart_reading(fconn, 1); + } + + TRACE_LEAVE(FCGI_EV_STRM_RECV, fconn->conn, fstrm); + return ret; +} + + +/* Called from the upper layer, to send data from buffer <buf> for no more than + * <count> bytes. Returns the number of bytes effectively sent. Some status + * flags may be updated on the stream connector. + */ +static size_t fcgi_snd_buf(struct stconn *sc, struct buffer *buf, size_t count, int flags) +{ + struct fcgi_strm *fstrm = __sc_mux_strm(sc); + struct fcgi_conn *fconn = fstrm->fconn; + size_t total = 0; + size_t ret; + struct htx *htx = NULL; + struct htx_sl *sl; + struct htx_blk *blk; + uint32_t bsize; + + TRACE_ENTER(FCGI_EV_STRM_SEND, fconn->conn, fstrm, 0, (size_t[]){count}); + + /* If we were not just woken because we wanted to send but couldn't, + * and there's somebody else that is waiting to send, do nothing, + * we will subscribe later and be put at the end of the list + */ + if (!(fstrm->flags & FCGI_SF_NOTIFIED) && !LIST_ISEMPTY(&fconn->send_list)) { + TRACE_STATE("other streams already waiting, going to the queue and leaving", FCGI_EV_STRM_SEND|FCGI_EV_FSTRM_BLK, fconn->conn, fstrm); + return 0; + } + fstrm->flags &= ~FCGI_SF_NOTIFIED; + + if (fconn->state < FCGI_CS_RECORD_H) { + TRACE_STATE("connection not ready, leaving", FCGI_EV_STRM_SEND|FCGI_EV_FSTRM_BLK, fconn->conn, fstrm); + return 0; + } + + htx = htxbuf(buf); + if (fstrm->id == 0) { + int32_t id = fcgi_conn_get_next_sid(fconn); + + if (id < 0) { + fcgi_strm_close(fstrm); + se_fl_set(fstrm->sd, SE_FL_ERROR); + TRACE_DEVEL("couldn't get a stream ID, leaving in error", FCGI_EV_STRM_SEND|FCGI_EV_FSTRM_ERR|FCGI_EV_STRM_ERR, fconn->conn, fstrm); + return 0; + } + + eb32_delete(&fstrm->by_id); + fstrm->by_id.key = fstrm->id = id; + fconn->max_id = id; + fconn->nb_reserved--; + eb32_insert(&fconn->streams_by_id, &fstrm->by_id); + + + /* Check if length of the body is known or if the message is + * full. Otherwise, the request is invalid. + */ + sl = http_get_stline(htx); + if (!sl || (!(sl->flags & HTX_SL_F_CLEN) && !(htx->flags & HTX_FL_EOM))) { + htx->flags |= HTX_FL_PARSING_ERROR; + fcgi_strm_error(fstrm); + goto done; + } + } + + if (!(fstrm->flags & FCGI_SF_BEGIN_SENT)) { + TRACE_PROTO("sending FCGI BEGIN_REQUEST record", FCGI_EV_TX_RECORD|FCGI_EV_TX_BEGREQ, fconn->conn, fstrm); + if (!fcgi_strm_send_begin_request(fconn, fstrm)) + goto done; + } + + if (!(fstrm->flags & FCGI_SF_OUTGOING_DATA) && count) + fstrm->flags |= FCGI_SF_OUTGOING_DATA; + + while (fstrm->state < FCGI_SS_HLOC && !(fstrm->flags & FCGI_SF_BLK_ANY) && + count && !htx_is_empty(htx)) { + blk = htx_get_head_blk(htx); + ALREADY_CHECKED(blk); + bsize = htx_get_blksz(blk); + + switch (htx_get_blk_type(blk)) { + case HTX_BLK_REQ_SL: + case HTX_BLK_HDR: + TRACE_USER("sending FCGI PARAMS record", FCGI_EV_TX_RECORD|FCGI_EV_TX_PARAMS, fconn->conn, fstrm, htx); + ret = fcgi_strm_send_params(fconn, fstrm, htx); + if (!ret) { + goto done; + } + total += ret; + count -= ret; + break; + + case HTX_BLK_EOH: + if (!(fstrm->flags & FCGI_SF_EP_SENT)) { + TRACE_PROTO("sending FCGI PARAMS record", FCGI_EV_TX_RECORD|FCGI_EV_TX_PARAMS, fconn->conn, fstrm, htx); + ret = fcgi_strm_send_empty_params(fconn, fstrm); + if (!ret) + goto done; + } + if (htx_is_unique_blk(htx, blk) && (htx->flags & HTX_FL_EOM)) { + TRACE_PROTO("sending FCGI STDIN record", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, htx); + ret = fcgi_strm_send_empty_stdin(fconn, fstrm); + if (!ret) + goto done; + } + goto remove_blk; + + case HTX_BLK_DATA: + TRACE_PROTO("sending FCGI STDIN record", FCGI_EV_TX_RECORD|FCGI_EV_TX_STDIN, fconn->conn, fstrm, htx); + ret = fcgi_strm_send_stdin(fconn, fstrm, htx, count, buf); + if (ret > 0) { + htx = htx_from_buf(buf); + total += ret; + count -= ret; + if (ret < bsize) + goto done; + } + break; + + default: + remove_blk: + htx_remove_blk(htx, blk); + total += bsize; + count -= bsize; + break; + } + } + + done: + if (fstrm->state >= FCGI_SS_HLOC) { + /* trim any possibly pending data after we close (extra CR-LF, + * unprocessed trailers, abnormal extra data, ...) + */ + total += count; + count = 0; + } + + if (fstrm->state == FCGI_SS_ERROR) { + TRACE_DEVEL("reporting error to the app-layer stream", FCGI_EV_STRM_SEND|FCGI_EV_FSTRM_ERR|FCGI_EV_STRM_ERR, fconn->conn, fstrm); + se_fl_set_error(fstrm->sd); + if (!(fstrm->flags & FCGI_SF_BEGIN_SENT) || fcgi_strm_send_abort(fconn, fstrm)) + fcgi_strm_close(fstrm); + } + + if (htx) + htx_to_buf(htx, buf); + + if (total > 0) { + if (!(fconn->wait_event.events & SUB_RETRY_SEND)) { + TRACE_DEVEL("data queued, waking up fconn sender", FCGI_EV_STRM_SEND|FCGI_EV_FCONN_SEND|FCGI_EV_FCONN_WAKE, fconn->conn, fstrm); + tasklet_wakeup(fconn->wait_event.tasklet); + } + + /* Ok we managed to send something, leave the send_list */ + if (!(fstrm->flags & (FCGI_SF_WANT_SHUTR|FCGI_SF_WANT_SHUTW))) + LIST_DEL_INIT(&fstrm->send_list); + } + + TRACE_LEAVE(FCGI_EV_STRM_SEND, fconn->conn, fstrm, htx, (size_t[]){total}); + return total; +} + +/* for debugging with CLI's "show fd" command */ +static int fcgi_show_fd(struct buffer *msg, struct connection *conn) +{ + struct fcgi_conn *fconn = conn->ctx; + struct fcgi_strm *fstrm = NULL; + struct eb32_node *node; + int send_cnt = 0; + int tree_cnt = 0; + int orph_cnt = 0; + struct buffer *hmbuf, *tmbuf; + + if (!fconn) + return 0; + + list_for_each_entry(fstrm, &fconn->send_list, send_list) + send_cnt++; + + fstrm = NULL; + node = eb32_first(&fconn->streams_by_id); + while (node) { + fstrm = container_of(node, struct fcgi_strm, by_id); + tree_cnt++; + if (!fcgi_strm_sc(fstrm)) + orph_cnt++; + node = eb32_next(node); + } + + hmbuf = br_head(fconn->mbuf); + tmbuf = br_tail(fconn->mbuf); + chunk_appendf(msg, " fconn.st0=%d .maxid=%d .flg=0x%04x .nbst=%u" + " .nbcs=%u .send_cnt=%d .tree_cnt=%d .orph_cnt=%d .sub=%d " + ".dsi=%d .dbuf=%u@%p+%u/%u .mbuf=[%u..%u|%u],h=[%u@%p+%u/%u],t=[%u@%p+%u/%u]", + fconn->state, fconn->max_id, fconn->flags, + fconn->nb_streams, fconn->nb_sc, send_cnt, tree_cnt, orph_cnt, + fconn->wait_event.events, fconn->dsi, + (unsigned int)b_data(&fconn->dbuf), b_orig(&fconn->dbuf), + (unsigned int)b_head_ofs(&fconn->dbuf), (unsigned int)b_size(&fconn->dbuf), + br_head_idx(fconn->mbuf), br_tail_idx(fconn->mbuf), br_size(fconn->mbuf), + (unsigned int)b_data(hmbuf), b_orig(hmbuf), + (unsigned int)b_head_ofs(hmbuf), (unsigned int)b_size(hmbuf), + (unsigned int)b_data(tmbuf), b_orig(tmbuf), + (unsigned int)b_head_ofs(tmbuf), (unsigned int)b_size(tmbuf)); + + if (fstrm) { + chunk_appendf(msg, " last_fstrm=%p .id=%d .flg=0x%04x .rxbuf=%u@%p+%u/%u .sc=%p", + fstrm, fstrm->id, fstrm->flags, + (unsigned int)b_data(&fstrm->rxbuf), b_orig(&fstrm->rxbuf), + (unsigned int)b_head_ofs(&fstrm->rxbuf), (unsigned int)b_size(&fstrm->rxbuf), + fcgi_strm_sc(fstrm)); + + chunk_appendf(msg, " .sd.flg=0x%08x", se_fl_get(fstrm->sd)); + if (!se_fl_test(fstrm->sd, SE_FL_ORPHAN)) + chunk_appendf(msg, " .sc.flg=0x%08x .sc.app=%p", + fcgi_strm_sc(fstrm)->flags, fcgi_strm_sc(fstrm)->app); + + chunk_appendf(msg, " .subs=%p", fstrm->subs); + if (fstrm->subs) { + chunk_appendf(msg, "(ev=%d tl=%p", fstrm->subs->events, fstrm->subs->tasklet); + chunk_appendf(msg, " tl.calls=%d tl.ctx=%p tl.fct=", + fstrm->subs->tasklet->calls, + fstrm->subs->tasklet->context); + resolve_sym_name(msg, NULL, fstrm->subs->tasklet->process); + chunk_appendf(msg, ")"); + } + } + return 0; +} + +/* Migrate the the connection to the current thread. + * Return 0 if successful, non-zero otherwise. + * Expected to be called with the old thread lock held. + */ +static int fcgi_takeover(struct connection *conn, int orig_tid) +{ + struct fcgi_conn *fcgi = conn->ctx; + struct task *task; + struct task *new_task; + struct tasklet *new_tasklet; + + /* Pre-allocate tasks so that we don't have to roll back after the xprt + * has been migrated. + */ + new_task = task_new_here(); + new_tasklet = tasklet_new(); + if (!new_task || !new_tasklet) + goto fail; + + if (fd_takeover(conn->handle.fd, conn) != 0) + goto fail; + + if (conn->xprt->takeover && conn->xprt->takeover(conn, conn->xprt_ctx, orig_tid) != 0) { + /* We failed to takeover the xprt, even if the connection may + * still be valid, flag it as error'd, as we have already + * taken over the fd, and wake the tasklet, so that it will + * destroy it. + */ + conn->flags |= CO_FL_ERROR; + tasklet_wakeup_on(fcgi->wait_event.tasklet, orig_tid); + goto fail; + } + + if (fcgi->wait_event.events) + fcgi->conn->xprt->unsubscribe(fcgi->conn, fcgi->conn->xprt_ctx, + fcgi->wait_event.events, &fcgi->wait_event); + + task = fcgi->task; + if (task) { + /* only assign a task if there was already one, otherwise + * the preallocated new task will be released. + */ + task->context = NULL; + fcgi->task = NULL; + __ha_barrier_store(); + task_kill(task); + + fcgi->task = new_task; + new_task = NULL; + fcgi->task->process = fcgi_timeout_task; + fcgi->task->context = fcgi; + } + + /* To let the tasklet know it should free itself, and do nothing else, + * set its context to NULL; + */ + fcgi->wait_event.tasklet->context = NULL; + tasklet_wakeup_on(fcgi->wait_event.tasklet, orig_tid); + + fcgi->wait_event.tasklet = new_tasklet; + fcgi->wait_event.tasklet->process = fcgi_io_cb; + fcgi->wait_event.tasklet->context = fcgi; + fcgi->conn->xprt->subscribe(fcgi->conn, fcgi->conn->xprt_ctx, + SUB_RETRY_RECV, &fcgi->wait_event); + + if (new_task) + __task_free(new_task); + return 0; + fail: + if (new_task) + __task_free(new_task); + tasklet_free(new_tasklet); + return -1; +} + +/****************************************/ +/* MUX initialization and instantiation */ +/****************************************/ + +/* The mux operations */ +static const struct mux_ops mux_fcgi_ops = { + .init = fcgi_init, + .wake = fcgi_wake, + .attach = fcgi_attach, + .get_first_sc = fcgi_get_first_sc, + .detach = fcgi_detach, + .destroy = fcgi_destroy, + .avail_streams = fcgi_avail_streams, + .used_streams = fcgi_used_streams, + .rcv_buf = fcgi_rcv_buf, + .snd_buf = fcgi_snd_buf, + .subscribe = fcgi_subscribe, + .unsubscribe = fcgi_unsubscribe, + .shutr = fcgi_shutr, + .shutw = fcgi_shutw, + .ctl = fcgi_ctl, + .sctl = fcgi_sctl, + .show_fd = fcgi_show_fd, + .takeover = fcgi_takeover, + .flags = MX_FL_HTX|MX_FL_HOL_RISK|MX_FL_NO_UPG, + .name = "FCGI", +}; + + +/* this mux registers FCGI proto */ +static struct mux_proto_list mux_proto_fcgi = +{ .token = IST("fcgi"), .mode = PROTO_MODE_HTTP, .side = PROTO_SIDE_BE, .mux = &mux_fcgi_ops }; + +INITCALL1(STG_REGISTER, register_mux_proto, &mux_proto_fcgi); + +/* + * Local variables: + * c-indent-level: 8 + * c-basic-offset: 8 + * End: + */ |