// SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0-only) /* Copyright(c) 2014 - 2020 Intel Corporation */ #include #include #include "adf_accel_devices.h" #include "adf_common_drv.h" #include "adf_transport.h" #include "adf_transport_access_macros.h" #include "adf_cfg.h" #include "adf_cfg_strings.h" #include "qat_crypto.h" #include "icp_qat_fw.h" #define SEC ADF_KERNEL_SEC static struct service_hndl qat_crypto; void qat_crypto_put_instance(struct qat_crypto_instance *inst) { atomic_dec(&inst->refctr); adf_dev_put(inst->accel_dev); } static int qat_crypto_free_instances(struct adf_accel_dev *accel_dev) { struct qat_crypto_instance *inst, *tmp; int i; list_for_each_entry_safe(inst, tmp, &accel_dev->crypto_list, list) { for (i = 0; i < atomic_read(&inst->refctr); i++) qat_crypto_put_instance(inst); if (inst->sym_tx) adf_remove_ring(inst->sym_tx); if (inst->sym_rx) adf_remove_ring(inst->sym_rx); if (inst->pke_tx) adf_remove_ring(inst->pke_tx); if (inst->pke_rx) adf_remove_ring(inst->pke_rx); list_del(&inst->list); kfree(inst); } return 0; } struct qat_crypto_instance *qat_crypto_get_instance_node(int node) { struct adf_accel_dev *accel_dev = NULL, *tmp_dev; struct qat_crypto_instance *inst = NULL, *tmp_inst; unsigned long best = ~0; list_for_each_entry(tmp_dev, adf_devmgr_get_head(), list) { unsigned long ctr; if ((node == dev_to_node(&GET_DEV(tmp_dev)) || dev_to_node(&GET_DEV(tmp_dev)) < 0) && adf_dev_started(tmp_dev) && !list_empty(&tmp_dev->crypto_list)) { ctr = atomic_read(&tmp_dev->ref_count); if (best > ctr) { accel_dev = tmp_dev; best = ctr; } } } if (!accel_dev) { pr_info("QAT: Could not find a device on node %d\n", node); /* Get any started device */ list_for_each_entry(tmp_dev, adf_devmgr_get_head(), list) { if (adf_dev_started(tmp_dev) && !list_empty(&tmp_dev->crypto_list)) { accel_dev = tmp_dev; break; } } } if (!accel_dev) return NULL; best = ~0; list_for_each_entry(tmp_inst, &accel_dev->crypto_list, list) { unsigned long ctr; ctr = atomic_read(&tmp_inst->refctr); if (best > ctr) { inst = tmp_inst; best = ctr; } } if (inst) { if (adf_dev_get(accel_dev)) { dev_err(&GET_DEV(accel_dev), "Could not increment dev refctr\n"); return NULL; } atomic_inc(&inst->refctr); } return inst; } /** * qat_crypto_dev_config() - create dev config required to create crypto inst. * * @accel_dev: Pointer to acceleration device. * * Function creates device configuration required to create crypto instances * * Return: 0 on success, error code otherwise. */ int qat_crypto_dev_config(struct adf_accel_dev *accel_dev) { int cpus = num_online_cpus(); int banks = GET_MAX_BANKS(accel_dev); int instances = min(cpus, banks); char key[ADF_CFG_MAX_KEY_LEN_IN_BYTES]; int i; unsigned long val; if (adf_cfg_section_add(accel_dev, ADF_KERNEL_SEC)) goto err; if (adf_cfg_section_add(accel_dev, "Accelerator0")) goto err; /* Temporarily set the number of crypto instances to zero to avoid * registering the crypto algorithms. * This will be removed when the algorithms will support the * CRYPTO_TFM_REQ_MAY_BACKLOG flag */ instances = 0; for (i = 0; i < instances; i++) { val = i; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_BANK_NUM, i); if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, (void *)&val, ADF_DEC)) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_ETRMGR_CORE_AFFINITY, i); if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, (void *)&val, ADF_DEC)) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_SIZE, i); val = 128; if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, (void *)&val, ADF_DEC)) goto err; val = 512; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_SIZE, i); if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, (void *)&val, ADF_DEC)) goto err; val = 0; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_TX, i); if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, (void *)&val, ADF_DEC)) goto err; val = 2; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_TX, i); if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, (void *)&val, ADF_DEC)) goto err; val = 8; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_RX, i); if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, (void *)&val, ADF_DEC)) goto err; val = 10; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_RX, i); if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, key, (void *)&val, ADF_DEC)) goto err; val = ADF_COALESCING_DEF_TIME; snprintf(key, sizeof(key), ADF_ETRMGR_COALESCE_TIMER_FORMAT, i); if (adf_cfg_add_key_value_param(accel_dev, "Accelerator0", key, (void *)&val, ADF_DEC)) goto err; } val = i; if (adf_cfg_add_key_value_param(accel_dev, ADF_KERNEL_SEC, ADF_NUM_CY, (void *)&val, ADF_DEC)) goto err; set_bit(ADF_STATUS_CONFIGURED, &accel_dev->status); return 0; err: dev_err(&GET_DEV(accel_dev), "Failed to start QAT accel dev\n"); return -EINVAL; } EXPORT_SYMBOL_GPL(qat_crypto_dev_config); static int qat_crypto_create_instances(struct adf_accel_dev *accel_dev) { int i; unsigned long bank; unsigned long num_inst, num_msg_sym, num_msg_asym; int msg_size; struct qat_crypto_instance *inst; char key[ADF_CFG_MAX_KEY_LEN_IN_BYTES]; char val[ADF_CFG_MAX_VAL_LEN_IN_BYTES]; INIT_LIST_HEAD(&accel_dev->crypto_list); if (adf_cfg_get_param_value(accel_dev, SEC, ADF_NUM_CY, val)) return -EFAULT; if (kstrtoul(val, 0, &num_inst)) return -EFAULT; for (i = 0; i < num_inst; i++) { inst = kzalloc_node(sizeof(*inst), GFP_KERNEL, dev_to_node(&GET_DEV(accel_dev))); if (!inst) goto err; list_add_tail(&inst->list, &accel_dev->crypto_list); inst->id = i; atomic_set(&inst->refctr, 0); inst->accel_dev = accel_dev; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_BANK_NUM, i); if (adf_cfg_get_param_value(accel_dev, SEC, key, val)) goto err; if (kstrtoul(val, 10, &bank)) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_SIZE, i); if (adf_cfg_get_param_value(accel_dev, SEC, key, val)) goto err; if (kstrtoul(val, 10, &num_msg_sym)) goto err; num_msg_sym = num_msg_sym >> 1; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_SIZE, i); if (adf_cfg_get_param_value(accel_dev, SEC, key, val)) goto err; if (kstrtoul(val, 10, &num_msg_asym)) goto err; num_msg_asym = num_msg_asym >> 1; msg_size = ICP_QAT_FW_REQ_DEFAULT_SZ; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_TX, i); if (adf_create_ring(accel_dev, SEC, bank, num_msg_sym, msg_size, key, NULL, 0, &inst->sym_tx)) goto err; msg_size = msg_size >> 1; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_TX, i); if (adf_create_ring(accel_dev, SEC, bank, num_msg_asym, msg_size, key, NULL, 0, &inst->pke_tx)) goto err; msg_size = ICP_QAT_FW_RESP_DEFAULT_SZ; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_SYM_RX, i); if (adf_create_ring(accel_dev, SEC, bank, num_msg_sym, msg_size, key, qat_alg_callback, 0, &inst->sym_rx)) goto err; snprintf(key, sizeof(key), ADF_CY "%d" ADF_RING_ASYM_RX, i); if (adf_create_ring(accel_dev, SEC, bank, num_msg_asym, msg_size, key, qat_alg_asym_callback, 0, &inst->pke_rx)) goto err; } return 0; err: qat_crypto_free_instances(accel_dev); return -ENOMEM; } static int qat_crypto_init(struct adf_accel_dev *accel_dev) { if (qat_crypto_create_instances(accel_dev)) return -EFAULT; return 0; } static int qat_crypto_shutdown(struct adf_accel_dev *accel_dev) { return qat_crypto_free_instances(accel_dev); } static int qat_crypto_event_handler(struct adf_accel_dev *accel_dev, enum adf_event event) { int ret; switch (event) { case ADF_EVENT_INIT: ret = qat_crypto_init(accel_dev); break; case ADF_EVENT_SHUTDOWN: ret = qat_crypto_shutdown(accel_dev); break; case ADF_EVENT_RESTARTING: case ADF_EVENT_RESTARTED: case ADF_EVENT_START: case ADF_EVENT_STOP: default: ret = 0; } return ret; } int qat_crypto_register(void) { memset(&qat_crypto, 0, sizeof(qat_crypto)); qat_crypto.event_hld = qat_crypto_event_handler; qat_crypto.name = "qat_crypto"; return adf_service_register(&qat_crypto); } int qat_crypto_unregister(void) { return adf_service_unregister(&qat_crypto); }