// SPDX-License-Identifier: GPL-2.0 /* * SD/MMC Greybus driver. * * Copyright 2014-2015 Google Inc. * Copyright 2014-2015 Linaro Ltd. */ #include #include #include #include #include #include #include "greybus.h" #include "gbphy.h" struct gb_sdio_host { struct gb_connection *connection; struct gbphy_device *gbphy_dev; struct mmc_host *mmc; struct mmc_request *mrq; struct mutex lock; /* lock for this host */ size_t data_max; spinlock_t xfer; /* lock to cancel ongoing transfer */ bool xfer_stop; struct workqueue_struct *mrq_workqueue; struct work_struct mrqwork; u8 queued_events; bool removed; bool card_present; bool read_only; }; #define GB_SDIO_RSP_R1_R5_R6_R7 (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \ GB_SDIO_RSP_OPCODE) #define GB_SDIO_RSP_R3_R4 (GB_SDIO_RSP_PRESENT) #define GB_SDIO_RSP_R2 (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \ GB_SDIO_RSP_136) #define GB_SDIO_RSP_R1B (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \ GB_SDIO_RSP_OPCODE | GB_SDIO_RSP_BUSY) /* kernel vdd starts at 0x80 and we need to translate to greybus ones 0x01 */ #define GB_SDIO_VDD_SHIFT 8 #ifndef MMC_CAP2_CORE_RUNTIME_PM #define MMC_CAP2_CORE_RUNTIME_PM 0 #endif static inline bool single_op(struct mmc_command *cmd) { u32 opcode = cmd->opcode; return opcode == MMC_WRITE_BLOCK || opcode == MMC_READ_SINGLE_BLOCK; } static void _gb_sdio_set_host_caps(struct gb_sdio_host *host, u32 r) { u32 caps = 0; u32 caps2 = 0; caps = ((r & GB_SDIO_CAP_NONREMOVABLE) ? MMC_CAP_NONREMOVABLE : 0) | ((r & GB_SDIO_CAP_4_BIT_DATA) ? MMC_CAP_4_BIT_DATA : 0) | ((r & GB_SDIO_CAP_8_BIT_DATA) ? MMC_CAP_8_BIT_DATA : 0) | ((r & GB_SDIO_CAP_MMC_HS) ? MMC_CAP_MMC_HIGHSPEED : 0) | ((r & GB_SDIO_CAP_SD_HS) ? MMC_CAP_SD_HIGHSPEED : 0) | ((r & GB_SDIO_CAP_ERASE) ? MMC_CAP_ERASE : 0) | ((r & GB_SDIO_CAP_1_2V_DDR) ? MMC_CAP_1_2V_DDR : 0) | ((r & GB_SDIO_CAP_1_8V_DDR) ? MMC_CAP_1_8V_DDR : 0) | ((r & GB_SDIO_CAP_POWER_OFF_CARD) ? MMC_CAP_POWER_OFF_CARD : 0) | ((r & GB_SDIO_CAP_UHS_SDR12) ? MMC_CAP_UHS_SDR12 : 0) | ((r & GB_SDIO_CAP_UHS_SDR25) ? MMC_CAP_UHS_SDR25 : 0) | ((r & GB_SDIO_CAP_UHS_SDR50) ? MMC_CAP_UHS_SDR50 : 0) | ((r & GB_SDIO_CAP_UHS_SDR104) ? MMC_CAP_UHS_SDR104 : 0) | ((r & GB_SDIO_CAP_UHS_DDR50) ? MMC_CAP_UHS_DDR50 : 0) | ((r & GB_SDIO_CAP_DRIVER_TYPE_A) ? MMC_CAP_DRIVER_TYPE_A : 0) | ((r & GB_SDIO_CAP_DRIVER_TYPE_C) ? MMC_CAP_DRIVER_TYPE_C : 0) | ((r & GB_SDIO_CAP_DRIVER_TYPE_D) ? MMC_CAP_DRIVER_TYPE_D : 0); caps2 = ((r & GB_SDIO_CAP_HS200_1_2V) ? MMC_CAP2_HS200_1_2V_SDR : 0) | ((r & GB_SDIO_CAP_HS400_1_2V) ? MMC_CAP2_HS400_1_2V : 0) | ((r & GB_SDIO_CAP_HS400_1_8V) ? MMC_CAP2_HS400_1_8V : 0) | ((r & GB_SDIO_CAP_HS200_1_8V) ? MMC_CAP2_HS200_1_8V_SDR : 0); host->mmc->caps = caps; host->mmc->caps2 = caps2 | MMC_CAP2_CORE_RUNTIME_PM; if (caps & MMC_CAP_NONREMOVABLE) host->card_present = true; } static u32 _gb_sdio_get_host_ocr(u32 ocr) { return (((ocr & GB_SDIO_VDD_165_195) ? MMC_VDD_165_195 : 0) | ((ocr & GB_SDIO_VDD_20_21) ? MMC_VDD_20_21 : 0) | ((ocr & GB_SDIO_VDD_21_22) ? MMC_VDD_21_22 : 0) | ((ocr & GB_SDIO_VDD_22_23) ? MMC_VDD_22_23 : 0) | ((ocr & GB_SDIO_VDD_23_24) ? MMC_VDD_23_24 : 0) | ((ocr & GB_SDIO_VDD_24_25) ? MMC_VDD_24_25 : 0) | ((ocr & GB_SDIO_VDD_25_26) ? MMC_VDD_25_26 : 0) | ((ocr & GB_SDIO_VDD_26_27) ? MMC_VDD_26_27 : 0) | ((ocr & GB_SDIO_VDD_27_28) ? MMC_VDD_27_28 : 0) | ((ocr & GB_SDIO_VDD_28_29) ? MMC_VDD_28_29 : 0) | ((ocr & GB_SDIO_VDD_29_30) ? MMC_VDD_29_30 : 0) | ((ocr & GB_SDIO_VDD_30_31) ? MMC_VDD_30_31 : 0) | ((ocr & GB_SDIO_VDD_31_32) ? MMC_VDD_31_32 : 0) | ((ocr & GB_SDIO_VDD_32_33) ? MMC_VDD_32_33 : 0) | ((ocr & GB_SDIO_VDD_33_34) ? MMC_VDD_33_34 : 0) | ((ocr & GB_SDIO_VDD_34_35) ? MMC_VDD_34_35 : 0) | ((ocr & GB_SDIO_VDD_35_36) ? MMC_VDD_35_36 : 0) ); } static int gb_sdio_get_caps(struct gb_sdio_host *host) { struct gb_sdio_get_caps_response response; struct mmc_host *mmc = host->mmc; u16 data_max; u32 blksz; u32 ocr; u32 r; int ret; ret = gb_operation_sync(host->connection, GB_SDIO_TYPE_GET_CAPABILITIES, NULL, 0, &response, sizeof(response)); if (ret < 0) return ret; r = le32_to_cpu(response.caps); _gb_sdio_set_host_caps(host, r); /* get the max block size that could fit our payload */ data_max = gb_operation_get_payload_size_max(host->connection); data_max = min(data_max - sizeof(struct gb_sdio_transfer_request), data_max - sizeof(struct gb_sdio_transfer_response)); blksz = min_t(u16, le16_to_cpu(response.max_blk_size), data_max); blksz = max_t(u32, 512, blksz); mmc->max_blk_size = rounddown_pow_of_two(blksz); mmc->max_blk_count = le16_to_cpu(response.max_blk_count); host->data_max = data_max; /* get ocr supported values */ ocr = _gb_sdio_get_host_ocr(le32_to_cpu(response.ocr)); mmc->ocr_avail = ocr; mmc->ocr_avail_sdio = mmc->ocr_avail; mmc->ocr_avail_sd = mmc->ocr_avail; mmc->ocr_avail_mmc = mmc->ocr_avail; /* get frequency range values */ mmc->f_min = le32_to_cpu(response.f_min); mmc->f_max = le32_to_cpu(response.f_max); return 0; } static void _gb_queue_event(struct gb_sdio_host *host, u8 event) { if (event & GB_SDIO_CARD_INSERTED) host->queued_events &= ~GB_SDIO_CARD_REMOVED; else if (event & GB_SDIO_CARD_REMOVED) host->queued_events &= ~GB_SDIO_CARD_INSERTED; host->queued_events |= event; } static int _gb_sdio_process_events(struct gb_sdio_host *host, u8 event) { u8 state_changed = 0; if (event & GB_SDIO_CARD_INSERTED) { if (host->mmc->caps & MMC_CAP_NONREMOVABLE) return 0; if (host->card_present) return 0; host->card_present = true; state_changed = 1; } if (event & GB_SDIO_CARD_REMOVED) { if (host->mmc->caps & MMC_CAP_NONREMOVABLE) return 0; if (!(host->card_present)) return 0; host->card_present = false; state_changed = 1; } if (event & GB_SDIO_WP) host->read_only = true; if (state_changed) { dev_info(mmc_dev(host->mmc), "card %s now event\n", (host->card_present ? "inserted" : "removed")); mmc_detect_change(host->mmc, 0); } return 0; } static int gb_sdio_request_handler(struct gb_operation *op) { struct gb_sdio_host *host = gb_connection_get_data(op->connection); struct gb_message *request; struct gb_sdio_event_request *payload; u8 type = op->type; int ret = 0; u8 event; if (type != GB_SDIO_TYPE_EVENT) { dev_err(mmc_dev(host->mmc), "unsupported unsolicited event: %u\n", type); return -EINVAL; } request = op->request; if (request->payload_size < sizeof(*payload)) { dev_err(mmc_dev(host->mmc), "wrong event size received (%zu < %zu)\n", request->payload_size, sizeof(*payload)); return -EINVAL; } payload = request->payload; event = payload->event; if (host->removed) _gb_queue_event(host, event); else ret = _gb_sdio_process_events(host, event); return ret; } static int gb_sdio_set_ios(struct gb_sdio_host *host, struct gb_sdio_set_ios_request *request) { int ret; ret = gbphy_runtime_get_sync(host->gbphy_dev); if (ret) return ret; ret = gb_operation_sync(host->connection, GB_SDIO_TYPE_SET_IOS, request, sizeof(*request), NULL, 0); gbphy_runtime_put_autosuspend(host->gbphy_dev); return ret; } static int _gb_sdio_send(struct gb_sdio_host *host, struct mmc_data *data, size_t len, u16 nblocks, off_t skip) { struct gb_sdio_transfer_request *request; struct gb_sdio_transfer_response *response; struct gb_operation *operation; struct scatterlist *sg = data->sg; unsigned int sg_len = data->sg_len; size_t copied; u16 send_blksz; u16 send_blocks; int ret; WARN_ON(len > host->data_max); operation = gb_operation_create(host->connection, GB_SDIO_TYPE_TRANSFER, len + sizeof(*request), sizeof(*response), GFP_KERNEL); if (!operation) return -ENOMEM; request = operation->request->payload; request->data_flags = (data->flags >> 8); request->data_blocks = cpu_to_le16(nblocks); request->data_blksz = cpu_to_le16(data->blksz); copied = sg_pcopy_to_buffer(sg, sg_len, &request->data[0], len, skip); if (copied != len) { ret = -EINVAL; goto err_put_operation; } ret = gb_operation_request_send_sync(operation); if (ret < 0) goto err_put_operation; response = operation->response->payload; send_blocks = le16_to_cpu(response->data_blocks); send_blksz = le16_to_cpu(response->data_blksz); if (len != send_blksz * send_blocks) { dev_err(mmc_dev(host->mmc), "send: size received: %zu != %d\n", len, send_blksz * send_blocks); ret = -EINVAL; } err_put_operation: gb_operation_put(operation); return ret; } static int _gb_sdio_recv(struct gb_sdio_host *host, struct mmc_data *data, size_t len, u16 nblocks, off_t skip) { struct gb_sdio_transfer_request *request; struct gb_sdio_transfer_response *response; struct gb_operation *operation; struct scatterlist *sg = data->sg; unsigned int sg_len = data->sg_len; size_t copied; u16 recv_blksz; u16 recv_blocks; int ret; WARN_ON(len > host->data_max); operation = gb_operation_create(host->connection, GB_SDIO_TYPE_TRANSFER, sizeof(*request), len + sizeof(*response), GFP_KERNEL); if (!operation) return -ENOMEM; request = operation->request->payload; request->data_flags = (data->flags >> 8); request->data_blocks = cpu_to_le16(nblocks); request->data_blksz = cpu_to_le16(data->blksz); ret = gb_operation_request_send_sync(operation); if (ret < 0) goto err_put_operation; response = operation->response->payload; recv_blocks = le16_to_cpu(response->data_blocks); recv_blksz = le16_to_cpu(response->data_blksz); if (len != recv_blksz * recv_blocks) { dev_err(mmc_dev(host->mmc), "recv: size received: %d != %zu\n", recv_blksz * recv_blocks, len); ret = -EINVAL; goto err_put_operation; } copied = sg_pcopy_from_buffer(sg, sg_len, &response->data[0], len, skip); if (copied != len) ret = -EINVAL; err_put_operation: gb_operation_put(operation); return ret; } static int gb_sdio_transfer(struct gb_sdio_host *host, struct mmc_data *data) { size_t left, len; off_t skip = 0; int ret = 0; u16 nblocks; if (single_op(data->mrq->cmd) && data->blocks > 1) { ret = -ETIMEDOUT; goto out; } left = data->blksz * data->blocks; while (left) { /* check is a stop transmission is pending */ spin_lock(&host->xfer); if (host->xfer_stop) { host->xfer_stop = false; spin_unlock(&host->xfer); ret = -EINTR; goto out; } spin_unlock(&host->xfer); len = min(left, host->data_max); nblocks = len / data->blksz; len = nblocks * data->blksz; if (data->flags & MMC_DATA_READ) { ret = _gb_sdio_recv(host, data, len, nblocks, skip); if (ret < 0) goto out; } else { ret = _gb_sdio_send(host, data, len, nblocks, skip); if (ret < 0) goto out; } data->bytes_xfered += len; left -= len; skip += len; } out: data->error = ret; return ret; } static int gb_sdio_command(struct gb_sdio_host *host, struct mmc_command *cmd) { struct gb_sdio_command_request request = {0}; struct gb_sdio_command_response response; struct mmc_data *data = host->mrq->data; unsigned int timeout_ms; u8 cmd_flags; u8 cmd_type; int i; int ret; switch (mmc_resp_type(cmd)) { case MMC_RSP_NONE: cmd_flags = GB_SDIO_RSP_NONE; break; case MMC_RSP_R1: cmd_flags = GB_SDIO_RSP_R1_R5_R6_R7; break; case MMC_RSP_R1B: cmd_flags = GB_SDIO_RSP_R1B; break; case MMC_RSP_R2: cmd_flags = GB_SDIO_RSP_R2; break; case MMC_RSP_R3: cmd_flags = GB_SDIO_RSP_R3_R4; break; default: dev_err(mmc_dev(host->mmc), "cmd flag invalid 0x%04x\n", mmc_resp_type(cmd)); ret = -EINVAL; goto out; } switch (mmc_cmd_type(cmd)) { case MMC_CMD_BC: cmd_type = GB_SDIO_CMD_BC; break; case MMC_CMD_BCR: cmd_type = GB_SDIO_CMD_BCR; break; case MMC_CMD_AC: cmd_type = GB_SDIO_CMD_AC; break; case MMC_CMD_ADTC: cmd_type = GB_SDIO_CMD_ADTC; break; default: dev_err(mmc_dev(host->mmc), "cmd type invalid 0x%04x\n", mmc_cmd_type(cmd)); ret = -EINVAL; goto out; } request.cmd = cmd->opcode; request.cmd_flags = cmd_flags; request.cmd_type = cmd_type; request.cmd_arg = cpu_to_le32(cmd->arg); /* some controllers need to know at command time data details */ if (data) { request.data_blocks = cpu_to_le16(data->blocks); request.data_blksz = cpu_to_le16(data->blksz); } timeout_ms = cmd->busy_timeout ? cmd->busy_timeout : GB_OPERATION_TIMEOUT_DEFAULT; ret = gb_operation_sync_timeout(host->connection, GB_SDIO_TYPE_COMMAND, &request, sizeof(request), &response, sizeof(response), timeout_ms); if (ret < 0) goto out; /* no response expected */ if (cmd_flags == GB_SDIO_RSP_NONE) goto out; /* long response expected */ if (cmd_flags & GB_SDIO_RSP_R2) for (i = 0; i < 4; i++) cmd->resp[i] = le32_to_cpu(response.resp[i]); else cmd->resp[0] = le32_to_cpu(response.resp[0]); out: cmd->error = ret; return ret; } static void gb_sdio_mrq_work(struct work_struct *work) { struct gb_sdio_host *host; struct mmc_request *mrq; int ret; host = container_of(work, struct gb_sdio_host, mrqwork); ret = gbphy_runtime_get_sync(host->gbphy_dev); if (ret) return; mutex_lock(&host->lock); mrq = host->mrq; if (!mrq) { mutex_unlock(&host->lock); gbphy_runtime_put_autosuspend(host->gbphy_dev); dev_err(mmc_dev(host->mmc), "mmc request is NULL"); return; } if (host->removed) { mrq->cmd->error = -ESHUTDOWN; goto done; } if (mrq->sbc) { ret = gb_sdio_command(host, mrq->sbc); if (ret < 0) goto done; } ret = gb_sdio_command(host, mrq->cmd); if (ret < 0) goto done; if (mrq->data) { ret = gb_sdio_transfer(host, mrq->data); if (ret < 0) goto done; } if (mrq->stop) { ret = gb_sdio_command(host, mrq->stop); if (ret < 0) goto done; } done: host->mrq = NULL; mutex_unlock(&host->lock); mmc_request_done(host->mmc, mrq); gbphy_runtime_put_autosuspend(host->gbphy_dev); } static void gb_mmc_request(struct mmc_host *mmc, struct mmc_request *mrq) { struct gb_sdio_host *host = mmc_priv(mmc); struct mmc_command *cmd = mrq->cmd; /* Check if it is a cancel to ongoing transfer */ if (cmd->opcode == MMC_STOP_TRANSMISSION) { spin_lock(&host->xfer); host->xfer_stop = true; spin_unlock(&host->xfer); } mutex_lock(&host->lock); WARN_ON(host->mrq); host->mrq = mrq; if (host->removed) { mrq->cmd->error = -ESHUTDOWN; goto out; } if (!host->card_present) { mrq->cmd->error = -ENOMEDIUM; goto out; } queue_work(host->mrq_workqueue, &host->mrqwork); mutex_unlock(&host->lock); return; out: host->mrq = NULL; mutex_unlock(&host->lock); mmc_request_done(mmc, mrq); } static void gb_mmc_set_ios(struct mmc_host *mmc, struct mmc_ios *ios) { struct gb_sdio_host *host = mmc_priv(mmc); struct gb_sdio_set_ios_request request; int ret; u8 power_mode; u8 bus_width; u8 timing; u8 signal_voltage; u8 drv_type; u32 vdd = 0; mutex_lock(&host->lock); request.clock = cpu_to_le32(ios->clock); if (ios->vdd) vdd = 1 << (ios->vdd - GB_SDIO_VDD_SHIFT); request.vdd = cpu_to_le32(vdd); request.bus_mode = (ios->bus_mode == MMC_BUSMODE_OPENDRAIN ? GB_SDIO_BUSMODE_OPENDRAIN : GB_SDIO_BUSMODE_PUSHPULL); switch (ios->power_mode) { case MMC_POWER_OFF: default: power_mode = GB_SDIO_POWER_OFF; break; case MMC_POWER_UP: power_mode = GB_SDIO_POWER_UP; break; case MMC_POWER_ON: power_mode = GB_SDIO_POWER_ON; break; case MMC_POWER_UNDEFINED: power_mode = GB_SDIO_POWER_UNDEFINED; break; } request.power_mode = power_mode; switch (ios->bus_width) { case MMC_BUS_WIDTH_1: bus_width = GB_SDIO_BUS_WIDTH_1; break; case MMC_BUS_WIDTH_4: default: bus_width = GB_SDIO_BUS_WIDTH_4; break; case MMC_BUS_WIDTH_8: bus_width = GB_SDIO_BUS_WIDTH_8; break; } request.bus_width = bus_width; switch (ios->timing) { case MMC_TIMING_LEGACY: default: timing = GB_SDIO_TIMING_LEGACY; break; case MMC_TIMING_MMC_HS: timing = GB_SDIO_TIMING_MMC_HS; break; case MMC_TIMING_SD_HS: timing = GB_SDIO_TIMING_SD_HS; break; case MMC_TIMING_UHS_SDR12: timing = GB_SDIO_TIMING_UHS_SDR12; break; case MMC_TIMING_UHS_SDR25: timing = GB_SDIO_TIMING_UHS_SDR25; break; case MMC_TIMING_UHS_SDR50: timing = GB_SDIO_TIMING_UHS_SDR50; break; case MMC_TIMING_UHS_SDR104: timing = GB_SDIO_TIMING_UHS_SDR104; break; case MMC_TIMING_UHS_DDR50: timing = GB_SDIO_TIMING_UHS_DDR50; break; case MMC_TIMING_MMC_DDR52: timing = GB_SDIO_TIMING_MMC_DDR52; break; case MMC_TIMING_MMC_HS200: timing = GB_SDIO_TIMING_MMC_HS200; break; case MMC_TIMING_MMC_HS400: timing = GB_SDIO_TIMING_MMC_HS400; break; } request.timing = timing; switch (ios->signal_voltage) { case MMC_SIGNAL_VOLTAGE_330: signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_330; break; case MMC_SIGNAL_VOLTAGE_180: default: signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_180; break; case MMC_SIGNAL_VOLTAGE_120: signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_120; break; } request.signal_voltage = signal_voltage; switch (ios->drv_type) { case MMC_SET_DRIVER_TYPE_A: drv_type = GB_SDIO_SET_DRIVER_TYPE_A; break; case MMC_SET_DRIVER_TYPE_C: drv_type = GB_SDIO_SET_DRIVER_TYPE_C; break; case MMC_SET_DRIVER_TYPE_D: drv_type = GB_SDIO_SET_DRIVER_TYPE_D; break; case MMC_SET_DRIVER_TYPE_B: default: drv_type = GB_SDIO_SET_DRIVER_TYPE_B; break; } request.drv_type = drv_type; ret = gb_sdio_set_ios(host, &request); if (ret < 0) goto out; memcpy(&mmc->ios, ios, sizeof(mmc->ios)); out: mutex_unlock(&host->lock); } static int gb_mmc_get_ro(struct mmc_host *mmc) { struct gb_sdio_host *host = mmc_priv(mmc); mutex_lock(&host->lock); if (host->removed) { mutex_unlock(&host->lock); return -ESHUTDOWN; } mutex_unlock(&host->lock); return host->read_only; } static int gb_mmc_get_cd(struct mmc_host *mmc) { struct gb_sdio_host *host = mmc_priv(mmc); mutex_lock(&host->lock); if (host->removed) { mutex_unlock(&host->lock); return -ESHUTDOWN; } mutex_unlock(&host->lock); return host->card_present; } static int gb_mmc_switch_voltage(struct mmc_host *mmc, struct mmc_ios *ios) { return 0; } static const struct mmc_host_ops gb_sdio_ops = { .request = gb_mmc_request, .set_ios = gb_mmc_set_ios, .get_ro = gb_mmc_get_ro, .get_cd = gb_mmc_get_cd, .start_signal_voltage_switch = gb_mmc_switch_voltage, }; static int gb_sdio_probe(struct gbphy_device *gbphy_dev, const struct gbphy_device_id *id) { struct gb_connection *connection; struct mmc_host *mmc; struct gb_sdio_host *host; int ret = 0; mmc = mmc_alloc_host(sizeof(*host), &gbphy_dev->dev); if (!mmc) return -ENOMEM; connection = gb_connection_create(gbphy_dev->bundle, le16_to_cpu(gbphy_dev->cport_desc->id), gb_sdio_request_handler); if (IS_ERR(connection)) { ret = PTR_ERR(connection); goto exit_mmc_free; } host = mmc_priv(mmc); host->mmc = mmc; host->removed = true; host->connection = connection; gb_connection_set_data(connection, host); host->gbphy_dev = gbphy_dev; gb_gbphy_set_data(gbphy_dev, host); ret = gb_connection_enable_tx(connection); if (ret) goto exit_connection_destroy; ret = gb_sdio_get_caps(host); if (ret < 0) goto exit_connection_disable; mmc->ops = &gb_sdio_ops; mmc->max_segs = host->mmc->max_blk_count; /* for now we make a map 1:1 between max request and segment size */ mmc->max_req_size = mmc->max_blk_size * mmc->max_blk_count; mmc->max_seg_size = mmc->max_req_size; mutex_init(&host->lock); spin_lock_init(&host->xfer); host->mrq_workqueue = alloc_workqueue("mmc-%s", 0, 1, dev_name(&gbphy_dev->dev)); if (!host->mrq_workqueue) { ret = -ENOMEM; goto exit_connection_disable; } INIT_WORK(&host->mrqwork, gb_sdio_mrq_work); ret = gb_connection_enable(connection); if (ret) goto exit_wq_destroy; ret = mmc_add_host(mmc); if (ret < 0) goto exit_wq_destroy; host->removed = false; ret = _gb_sdio_process_events(host, host->queued_events); host->queued_events = 0; gbphy_runtime_put_autosuspend(gbphy_dev); return ret; exit_wq_destroy: destroy_workqueue(host->mrq_workqueue); exit_connection_disable: gb_connection_disable(connection); exit_connection_destroy: gb_connection_destroy(connection); exit_mmc_free: mmc_free_host(mmc); return ret; } static void gb_sdio_remove(struct gbphy_device *gbphy_dev) { struct gb_sdio_host *host = gb_gbphy_get_data(gbphy_dev); struct gb_connection *connection = host->connection; struct mmc_host *mmc; int ret; ret = gbphy_runtime_get_sync(gbphy_dev); if (ret) gbphy_runtime_get_noresume(gbphy_dev); mutex_lock(&host->lock); host->removed = true; mmc = host->mmc; gb_connection_set_data(connection, NULL); mutex_unlock(&host->lock); flush_workqueue(host->mrq_workqueue); destroy_workqueue(host->mrq_workqueue); gb_connection_disable_rx(connection); mmc_remove_host(mmc); gb_connection_disable(connection); gb_connection_destroy(connection); mmc_free_host(mmc); } static const struct gbphy_device_id gb_sdio_id_table[] = { { GBPHY_PROTOCOL(GREYBUS_PROTOCOL_SDIO) }, { }, }; MODULE_DEVICE_TABLE(gbphy, gb_sdio_id_table); static struct gbphy_driver sdio_driver = { .name = "sdio", .probe = gb_sdio_probe, .remove = gb_sdio_remove, .id_table = gb_sdio_id_table, }; module_gbphy_driver(sdio_driver); MODULE_LICENSE("GPL v2");