diff options
Diffstat (limited to 'greybus_sdio.patch')
| -rw-r--r-- | greybus_sdio.patch | 891 |
1 files changed, 891 insertions, 0 deletions
diff --git a/greybus_sdio.patch b/greybus_sdio.patch new file mode 100644 index 00000000000000..32e0d3af6d5804 --- /dev/null +++ b/greybus_sdio.patch @@ -0,0 +1,891 @@ +--- + drivers/greybus/sdio.c | 884 +++++++++++++++++++++++++++++++++++++++++++++++++ + 1 file changed, 884 insertions(+) + +--- /dev/null ++++ b/drivers/greybus/sdio.c +@@ -0,0 +1,884 @@ ++/* ++ * SD/MMC Greybus driver. ++ * ++ * Copyright 2014-2015 Google Inc. ++ * Copyright 2014-2015 Linaro Ltd. ++ * ++ * Released under the GPLv2 only. ++ */ ++ ++#include <linux/kernel.h> ++#include <linux/mmc/core.h> ++#include <linux/mmc/host.h> ++#include <linux/mmc/mmc.h> ++#include <linux/scatterlist.h> ++#include <linux/workqueue.h> ++ ++#include "greybus.h" ++#include "gbphy.h" ++ ++struct gb_sdio_host { ++ struct gb_connection *connection; ++ struct gbphy_device *gbphy_dev; ++ struct mmc_host *mmc; ++ struct mmc_request *mrq; ++ struct mutex lock; /* lock for this host */ ++ size_t data_max; ++ spinlock_t xfer; /* lock to cancel ongoing transfer */ ++ bool xfer_stop; ++ struct workqueue_struct *mrq_workqueue; ++ struct work_struct mrqwork; ++ u8 queued_events; ++ bool removed; ++ bool card_present; ++ bool read_only; ++}; ++ ++ ++#define GB_SDIO_RSP_R1_R5_R6_R7 (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \ ++ GB_SDIO_RSP_OPCODE) ++#define GB_SDIO_RSP_R3_R4 (GB_SDIO_RSP_PRESENT) ++#define GB_SDIO_RSP_R2 (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \ ++ GB_SDIO_RSP_136) ++#define GB_SDIO_RSP_R1B (GB_SDIO_RSP_PRESENT | GB_SDIO_RSP_CRC | \ ++ GB_SDIO_RSP_OPCODE | GB_SDIO_RSP_BUSY) ++ ++/* kernel vdd starts at 0x80 and we need to translate to greybus ones 0x01 */ ++#define GB_SDIO_VDD_SHIFT 8 ++ ++#ifndef MMC_CAP2_CORE_RUNTIME_PM ++#define MMC_CAP2_CORE_RUNTIME_PM 0 ++#endif ++ ++static inline bool single_op(struct mmc_command *cmd) ++{ ++ uint32_t opcode = cmd->opcode; ++ ++ return opcode == MMC_WRITE_BLOCK || ++ opcode == MMC_READ_SINGLE_BLOCK; ++} ++ ++static void _gb_sdio_set_host_caps(struct gb_sdio_host *host, u32 r) ++{ ++ u32 caps = 0; ++ u32 caps2 = 0; ++ ++ caps = ((r & GB_SDIO_CAP_NONREMOVABLE) ? MMC_CAP_NONREMOVABLE : 0) | ++ ((r & GB_SDIO_CAP_4_BIT_DATA) ? MMC_CAP_4_BIT_DATA : 0) | ++ ((r & GB_SDIO_CAP_8_BIT_DATA) ? MMC_CAP_8_BIT_DATA : 0) | ++ ((r & GB_SDIO_CAP_MMC_HS) ? MMC_CAP_MMC_HIGHSPEED : 0) | ++ ((r & GB_SDIO_CAP_SD_HS) ? MMC_CAP_SD_HIGHSPEED : 0) | ++ ((r & GB_SDIO_CAP_ERASE) ? MMC_CAP_ERASE : 0) | ++ ((r & GB_SDIO_CAP_1_2V_DDR) ? MMC_CAP_1_2V_DDR : 0) | ++ ((r & GB_SDIO_CAP_1_8V_DDR) ? MMC_CAP_1_8V_DDR : 0) | ++ ((r & GB_SDIO_CAP_POWER_OFF_CARD) ? MMC_CAP_POWER_OFF_CARD : 0) | ++ ((r & GB_SDIO_CAP_UHS_SDR12) ? MMC_CAP_UHS_SDR12 : 0) | ++ ((r & GB_SDIO_CAP_UHS_SDR25) ? MMC_CAP_UHS_SDR25 : 0) | ++ ((r & GB_SDIO_CAP_UHS_SDR50) ? MMC_CAP_UHS_SDR50 : 0) | ++ ((r & GB_SDIO_CAP_UHS_SDR104) ? MMC_CAP_UHS_SDR104 : 0) | ++ ((r & GB_SDIO_CAP_UHS_DDR50) ? MMC_CAP_UHS_DDR50 : 0) | ++ ((r & GB_SDIO_CAP_DRIVER_TYPE_A) ? MMC_CAP_DRIVER_TYPE_A : 0) | ++ ((r & GB_SDIO_CAP_DRIVER_TYPE_C) ? MMC_CAP_DRIVER_TYPE_C : 0) | ++ ((r & GB_SDIO_CAP_DRIVER_TYPE_D) ? MMC_CAP_DRIVER_TYPE_D : 0); ++ ++ caps2 = ((r & GB_SDIO_CAP_HS200_1_2V) ? MMC_CAP2_HS200_1_2V_SDR : 0) | ++ ((r & GB_SDIO_CAP_HS400_1_2V) ? MMC_CAP2_HS400_1_2V : 0) | ++ ((r & GB_SDIO_CAP_HS400_1_8V) ? MMC_CAP2_HS400_1_8V : 0) | ++ ((r & GB_SDIO_CAP_HS200_1_8V) ? MMC_CAP2_HS200_1_8V_SDR : 0); ++ ++ host->mmc->caps = caps; ++ host->mmc->caps2 = caps2 | MMC_CAP2_CORE_RUNTIME_PM; ++ ++ if (caps & MMC_CAP_NONREMOVABLE) ++ host->card_present = true; ++} ++ ++static u32 _gb_sdio_get_host_ocr(u32 ocr) ++{ ++ return (((ocr & GB_SDIO_VDD_165_195) ? MMC_VDD_165_195 : 0) | ++ ((ocr & GB_SDIO_VDD_20_21) ? MMC_VDD_20_21 : 0) | ++ ((ocr & GB_SDIO_VDD_21_22) ? MMC_VDD_21_22 : 0) | ++ ((ocr & GB_SDIO_VDD_22_23) ? MMC_VDD_22_23 : 0) | ++ ((ocr & GB_SDIO_VDD_23_24) ? MMC_VDD_23_24 : 0) | ++ ((ocr & GB_SDIO_VDD_24_25) ? MMC_VDD_24_25 : 0) | ++ ((ocr & GB_SDIO_VDD_25_26) ? MMC_VDD_25_26 : 0) | ++ ((ocr & GB_SDIO_VDD_26_27) ? MMC_VDD_26_27 : 0) | ++ ((ocr & GB_SDIO_VDD_27_28) ? MMC_VDD_27_28 : 0) | ++ ((ocr & GB_SDIO_VDD_28_29) ? MMC_VDD_28_29 : 0) | ++ ((ocr & GB_SDIO_VDD_29_30) ? MMC_VDD_29_30 : 0) | ++ ((ocr & GB_SDIO_VDD_30_31) ? MMC_VDD_30_31 : 0) | ++ ((ocr & GB_SDIO_VDD_31_32) ? MMC_VDD_31_32 : 0) | ++ ((ocr & GB_SDIO_VDD_32_33) ? MMC_VDD_32_33 : 0) | ++ ((ocr & GB_SDIO_VDD_33_34) ? MMC_VDD_33_34 : 0) | ++ ((ocr & GB_SDIO_VDD_34_35) ? MMC_VDD_34_35 : 0) | ++ ((ocr & GB_SDIO_VDD_35_36) ? MMC_VDD_35_36 : 0) ++ ); ++} ++ ++static int gb_sdio_get_caps(struct gb_sdio_host *host) ++{ ++ struct gb_sdio_get_caps_response response; ++ struct mmc_host *mmc = host->mmc; ++ u16 data_max; ++ u32 blksz; ++ u32 ocr; ++ u32 r; ++ int ret; ++ ++ ret = gb_operation_sync(host->connection, GB_SDIO_TYPE_GET_CAPABILITIES, ++ NULL, 0, &response, sizeof(response)); ++ if (ret < 0) ++ return ret; ++ r = le32_to_cpu(response.caps); ++ ++ _gb_sdio_set_host_caps(host, r); ++ ++ /* get the max block size that could fit our payload */ ++ data_max = gb_operation_get_payload_size_max(host->connection); ++ data_max = min(data_max - sizeof(struct gb_sdio_transfer_request), ++ data_max - sizeof(struct gb_sdio_transfer_response)); ++ ++ blksz = min_t(u16, le16_to_cpu(response.max_blk_size), data_max); ++ blksz = max_t(u32, 512, blksz); ++ ++ mmc->max_blk_size = rounddown_pow_of_two(blksz); ++ mmc->max_blk_count = le16_to_cpu(response.max_blk_count); ++ host->data_max = data_max; ++ ++ /* get ocr supported values */ ++ ocr = _gb_sdio_get_host_ocr(le32_to_cpu(response.ocr)); ++ mmc->ocr_avail = ocr; ++ mmc->ocr_avail_sdio = mmc->ocr_avail; ++ mmc->ocr_avail_sd = mmc->ocr_avail; ++ mmc->ocr_avail_mmc = mmc->ocr_avail; ++ ++ /* get frequency range values */ ++ mmc->f_min = le32_to_cpu(response.f_min); ++ mmc->f_max = le32_to_cpu(response.f_max); ++ ++ return 0; ++} ++ ++static void _gb_queue_event(struct gb_sdio_host *host, u8 event) ++{ ++ if (event & GB_SDIO_CARD_INSERTED) ++ host->queued_events &= ~GB_SDIO_CARD_REMOVED; ++ else if (event & GB_SDIO_CARD_REMOVED) ++ host->queued_events &= ~GB_SDIO_CARD_INSERTED; ++ ++ host->queued_events |= event; ++} ++ ++static int _gb_sdio_process_events(struct gb_sdio_host *host, u8 event) ++{ ++ u8 state_changed = 0; ++ ++ if (event & GB_SDIO_CARD_INSERTED) { ++ if (host->mmc->caps & MMC_CAP_NONREMOVABLE) ++ return 0; ++ if (host->card_present) ++ return 0; ++ host->card_present = true; ++ state_changed = 1; ++ } ++ ++ if (event & GB_SDIO_CARD_REMOVED) { ++ if (host->mmc->caps & MMC_CAP_NONREMOVABLE) ++ return 0; ++ if (!(host->card_present)) ++ return 0; ++ host->card_present = false; ++ state_changed = 1; ++ } ++ ++ if (event & GB_SDIO_WP) { ++ host->read_only = true; ++ } ++ ++ if (state_changed) { ++ dev_info(mmc_dev(host->mmc), "card %s now event\n", ++ (host->card_present ? "inserted" : "removed")); ++ mmc_detect_change(host->mmc, 0); ++ } ++ ++ return 0; ++} ++ ++static int gb_sdio_request_handler(struct gb_operation *op) ++{ ++ struct gb_sdio_host *host = gb_connection_get_data(op->connection); ++ struct gb_message *request; ++ struct gb_sdio_event_request *payload; ++ u8 type = op->type; ++ int ret = 0; ++ u8 event; ++ ++ if (type != GB_SDIO_TYPE_EVENT) { ++ dev_err(mmc_dev(host->mmc), ++ "unsupported unsolicited event: %u\n", type); ++ return -EINVAL; ++ } ++ ++ request = op->request; ++ ++ if (request->payload_size < sizeof(*payload)) { ++ dev_err(mmc_dev(host->mmc), "wrong event size received (%zu < %zu)\n", ++ request->payload_size, sizeof(*payload)); ++ return -EINVAL; ++ } ++ ++ payload = request->payload; ++ event = payload->event; ++ ++ if (host->removed) ++ _gb_queue_event(host, event); ++ else ++ ret = _gb_sdio_process_events(host, event); ++ ++ return ret; ++} ++ ++static int gb_sdio_set_ios(struct gb_sdio_host *host, ++ struct gb_sdio_set_ios_request *request) ++{ ++ int ret; ++ ++ ret = gbphy_runtime_get_sync(host->gbphy_dev); ++ if (ret) ++ return ret; ++ ++ ret = gb_operation_sync(host->connection, GB_SDIO_TYPE_SET_IOS, request, ++ sizeof(*request), NULL, 0); ++ ++ gbphy_runtime_put_autosuspend(host->gbphy_dev); ++ ++ return ret; ++} ++ ++static int _gb_sdio_send(struct gb_sdio_host *host, struct mmc_data *data, ++ size_t len, u16 nblocks, off_t skip) ++{ ++ struct gb_sdio_transfer_request *request; ++ struct gb_sdio_transfer_response *response; ++ struct gb_operation *operation; ++ struct scatterlist *sg = data->sg; ++ unsigned int sg_len = data->sg_len; ++ size_t copied; ++ u16 send_blksz; ++ u16 send_blocks; ++ int ret; ++ ++ WARN_ON(len > host->data_max); ++ ++ operation = gb_operation_create(host->connection, GB_SDIO_TYPE_TRANSFER, ++ len + sizeof(*request), ++ sizeof(*response), GFP_KERNEL); ++ if (!operation) ++ return -ENOMEM; ++ ++ request = operation->request->payload; ++ request->data_flags = (data->flags >> 8); ++ request->data_blocks = cpu_to_le16(nblocks); ++ request->data_blksz = cpu_to_le16(data->blksz); ++ ++ copied = sg_pcopy_to_buffer(sg, sg_len, &request->data[0], len, skip); ++ ++ if (copied != len) { ++ ret = -EINVAL; ++ goto err_put_operation; ++ } ++ ++ ret = gb_operation_request_send_sync(operation); ++ if (ret < 0) ++ goto err_put_operation; ++ ++ response = operation->response->payload; ++ ++ send_blocks = le16_to_cpu(response->data_blocks); ++ send_blksz = le16_to_cpu(response->data_blksz); ++ ++ if (len != send_blksz * send_blocks) { ++ dev_err(mmc_dev(host->mmc), "send: size received: %zu != %d\n", ++ len, send_blksz * send_blocks); ++ ret = -EINVAL; ++ } ++ ++err_put_operation: ++ gb_operation_put(operation); ++ ++ return ret; ++} ++ ++static int _gb_sdio_recv(struct gb_sdio_host *host, struct mmc_data *data, ++ size_t len, u16 nblocks, off_t skip) ++{ ++ struct gb_sdio_transfer_request *request; ++ struct gb_sdio_transfer_response *response; ++ struct gb_operation *operation; ++ struct scatterlist *sg = data->sg; ++ unsigned int sg_len = data->sg_len; ++ size_t copied; ++ u16 recv_blksz; ++ u16 recv_blocks; ++ int ret; ++ ++ WARN_ON(len > host->data_max); ++ ++ operation = gb_operation_create(host->connection, GB_SDIO_TYPE_TRANSFER, ++ sizeof(*request), ++ len + sizeof(*response), GFP_KERNEL); ++ if (!operation) ++ return -ENOMEM; ++ ++ request = operation->request->payload; ++ request->data_flags = (data->flags >> 8); ++ request->data_blocks = cpu_to_le16(nblocks); ++ request->data_blksz = cpu_to_le16(data->blksz); ++ ++ ret = gb_operation_request_send_sync(operation); ++ if (ret < 0) ++ goto err_put_operation; ++ ++ response = operation->response->payload; ++ recv_blocks = le16_to_cpu(response->data_blocks); ++ recv_blksz = le16_to_cpu(response->data_blksz); ++ ++ if (len != recv_blksz * recv_blocks) { ++ dev_err(mmc_dev(host->mmc), "recv: size received: %d != %zu\n", ++ recv_blksz * recv_blocks, len); ++ ret = -EINVAL; ++ goto err_put_operation; ++ } ++ ++ copied = sg_pcopy_from_buffer(sg, sg_len, &response->data[0], len, ++ skip); ++ if (copied != len) ++ ret = -EINVAL; ++ ++err_put_operation: ++ gb_operation_put(operation); ++ ++ return ret; ++} ++ ++static int gb_sdio_transfer(struct gb_sdio_host *host, struct mmc_data *data) ++{ ++ size_t left, len; ++ off_t skip = 0; ++ int ret = 0; ++ u16 nblocks; ++ ++ if (single_op(data->mrq->cmd) && data->blocks > 1) { ++ ret = -ETIMEDOUT; ++ goto out; ++ } ++ ++ left = data->blksz * data->blocks; ++ ++ while (left) { ++ /* check is a stop transmission is pending */ ++ spin_lock(&host->xfer); ++ if (host->xfer_stop) { ++ host->xfer_stop = false; ++ spin_unlock(&host->xfer); ++ ret = -EINTR; ++ goto out; ++ } ++ spin_unlock(&host->xfer); ++ len = min(left, host->data_max); ++ nblocks = len / data->blksz; ++ len = nblocks * data->blksz; ++ ++ if (data->flags & MMC_DATA_READ) { ++ ret = _gb_sdio_recv(host, data, len, nblocks, skip); ++ if (ret < 0) ++ goto out; ++ } else { ++ ret = _gb_sdio_send(host, data, len, nblocks, skip); ++ if (ret < 0) ++ goto out; ++ } ++ data->bytes_xfered += len; ++ left -= len; ++ skip += len; ++ } ++ ++out: ++ data->error = ret; ++ return ret; ++} ++ ++static int gb_sdio_command(struct gb_sdio_host *host, struct mmc_command *cmd) ++{ ++ struct gb_sdio_command_request request = {0}; ++ struct gb_sdio_command_response response; ++ struct mmc_data *data = host->mrq->data; ++ u8 cmd_flags; ++ u8 cmd_type; ++ int i; ++ int ret; ++ ++ switch (mmc_resp_type(cmd)) { ++ case MMC_RSP_NONE: ++ cmd_flags = GB_SDIO_RSP_NONE; ++ break; ++ case MMC_RSP_R1: ++ cmd_flags = GB_SDIO_RSP_R1_R5_R6_R7; ++ break; ++ case MMC_RSP_R1B: ++ cmd_flags = GB_SDIO_RSP_R1B; ++ break; ++ case MMC_RSP_R2: ++ cmd_flags = GB_SDIO_RSP_R2; ++ break; ++ case MMC_RSP_R3: ++ cmd_flags = GB_SDIO_RSP_R3_R4; ++ break; ++ default: ++ dev_err(mmc_dev(host->mmc), "cmd flag invalid 0x%04x\n", ++ mmc_resp_type(cmd)); ++ ret = -EINVAL; ++ goto out; ++ } ++ ++ switch (mmc_cmd_type(cmd)) { ++ case MMC_CMD_BC: ++ cmd_type = GB_SDIO_CMD_BC; ++ break; ++ case MMC_CMD_BCR: ++ cmd_type = GB_SDIO_CMD_BCR; ++ break; ++ case MMC_CMD_AC: ++ cmd_type = GB_SDIO_CMD_AC; ++ break; ++ case MMC_CMD_ADTC: ++ cmd_type = GB_SDIO_CMD_ADTC; ++ break; ++ default: ++ dev_err(mmc_dev(host->mmc), "cmd type invalid 0x%04x\n", ++ mmc_cmd_type(cmd)); ++ ret = -EINVAL; ++ goto out; ++ } ++ ++ request.cmd = cmd->opcode; ++ request.cmd_flags = cmd_flags; ++ request.cmd_type = cmd_type; ++ request.cmd_arg = cpu_to_le32(cmd->arg); ++ /* some controllers need to know at command time data details */ ++ if (data) { ++ request.data_blocks = cpu_to_le16(data->blocks); ++ request.data_blksz = cpu_to_le16(data->blksz); ++ } ++ ++ ret = gb_operation_sync(host->connection, GB_SDIO_TYPE_COMMAND, ++ &request, sizeof(request), &response, ++ sizeof(response)); ++ if (ret < 0) ++ goto out; ++ ++ /* no response expected */ ++ if (cmd_flags & GB_SDIO_RSP_NONE) ++ goto out; ++ ++ /* long response expected */ ++ if (cmd_flags & GB_SDIO_RSP_R2) ++ for (i = 0; i < 4; i++) ++ cmd->resp[i] = le32_to_cpu(response.resp[i]); ++ else ++ cmd->resp[0] = le32_to_cpu(response.resp[0]); ++ ++out: ++ cmd->error = ret; ++ return ret; ++} ++ ++static void gb_sdio_mrq_work(struct work_struct *work) ++{ ++ struct gb_sdio_host *host; ++ struct mmc_request *mrq; ++ int ret; ++ ++ host = container_of(work, struct gb_sdio_host, mrqwork); ++ ++ ret = gbphy_runtime_get_sync(host->gbphy_dev); ++ if (ret) ++ return; ++ ++ mutex_lock(&host->lock); ++ mrq = host->mrq; ++ if (!mrq) { ++ mutex_unlock(&host->lock); ++ gbphy_runtime_put_autosuspend(host->gbphy_dev); ++ dev_err(mmc_dev(host->mmc), "mmc request is NULL"); ++ return; ++ } ++ ++ if (host->removed) { ++ mrq->cmd->error = -ESHUTDOWN; ++ goto done; ++ } ++ ++ if (mrq->sbc) { ++ ret = gb_sdio_command(host, mrq->sbc); ++ if (ret < 0) ++ goto done; ++ } ++ ++ ret = gb_sdio_command(host, mrq->cmd); ++ if (ret < 0) ++ goto done; ++ ++ if (mrq->data) { ++ ret = gb_sdio_transfer(host, mrq->data); ++ if (ret < 0) ++ goto done; ++ } ++ ++ if (mrq->stop) { ++ ret = gb_sdio_command(host, mrq->stop); ++ if (ret < 0) ++ goto done; ++ } ++ ++done: ++ host->mrq = NULL; ++ mutex_unlock(&host->lock); ++ mmc_request_done(host->mmc, mrq); ++ gbphy_runtime_put_autosuspend(host->gbphy_dev); ++} ++ ++static void gb_mmc_request(struct mmc_host *mmc, struct mmc_request *mrq) ++{ ++ struct gb_sdio_host *host = mmc_priv(mmc); ++ struct mmc_command *cmd = mrq->cmd; ++ ++ /* Check if it is a cancel to ongoing transfer */ ++ if (cmd->opcode == MMC_STOP_TRANSMISSION) { ++ spin_lock(&host->xfer); ++ host->xfer_stop = true; ++ spin_unlock(&host->xfer); ++ } ++ ++ mutex_lock(&host->lock); ++ ++ WARN_ON(host->mrq); ++ host->mrq = mrq; ++ ++ if (host->removed) { ++ mrq->cmd->error = -ESHUTDOWN; ++ goto out; ++ } ++ if (!host->card_present) { ++ mrq->cmd->error = -ENOMEDIUM; ++ goto out; ++ } ++ ++ queue_work(host->mrq_workqueue, &host->mrqwork); ++ ++ mutex_unlock(&host->lock); ++ return; ++ ++out: ++ host->mrq = NULL; ++ mutex_unlock(&host->lock); ++ mmc_request_done(mmc, mrq); ++} ++ ++static void gb_mmc_set_ios(struct mmc_host *mmc, struct mmc_ios *ios) ++{ ++ struct gb_sdio_host *host = mmc_priv(mmc); ++ struct gb_sdio_set_ios_request request; ++ int ret; ++ u8 power_mode; ++ u8 bus_width; ++ u8 timing; ++ u8 signal_voltage; ++ u8 drv_type; ++ u32 vdd = 0; ++ ++ mutex_lock(&host->lock); ++ request.clock = cpu_to_le32(ios->clock); ++ ++ if (ios->vdd) ++ vdd = 1 << (ios->vdd - GB_SDIO_VDD_SHIFT); ++ request.vdd = cpu_to_le32(vdd); ++ ++ request.bus_mode = (ios->bus_mode == MMC_BUSMODE_OPENDRAIN ? ++ GB_SDIO_BUSMODE_OPENDRAIN : ++ GB_SDIO_BUSMODE_PUSHPULL); ++ ++ switch (ios->power_mode) { ++ case MMC_POWER_OFF: ++ default: ++ power_mode = GB_SDIO_POWER_OFF; ++ break; ++ case MMC_POWER_UP: ++ power_mode = GB_SDIO_POWER_UP; ++ break; ++ case MMC_POWER_ON: ++ power_mode = GB_SDIO_POWER_ON; ++ break; ++ case MMC_POWER_UNDEFINED: ++ power_mode = GB_SDIO_POWER_UNDEFINED; ++ break; ++ } ++ request.power_mode = power_mode; ++ ++ switch (ios->bus_width) { ++ case MMC_BUS_WIDTH_1: ++ bus_width = GB_SDIO_BUS_WIDTH_1; ++ break; ++ case MMC_BUS_WIDTH_4: ++ default: ++ bus_width = GB_SDIO_BUS_WIDTH_4; ++ break; ++ case MMC_BUS_WIDTH_8: ++ bus_width = GB_SDIO_BUS_WIDTH_8; ++ break; ++ } ++ request.bus_width = bus_width; ++ ++ switch (ios->timing) { ++ case MMC_TIMING_LEGACY: ++ default: ++ timing = GB_SDIO_TIMING_LEGACY; ++ break; ++ case MMC_TIMING_MMC_HS: ++ timing = GB_SDIO_TIMING_MMC_HS; ++ break; ++ case MMC_TIMING_SD_HS: ++ timing = GB_SDIO_TIMING_SD_HS; ++ break; ++ case MMC_TIMING_UHS_SDR12: ++ timing = GB_SDIO_TIMING_UHS_SDR12; ++ break; ++ case MMC_TIMING_UHS_SDR25: ++ timing = GB_SDIO_TIMING_UHS_SDR25; ++ break; ++ case MMC_TIMING_UHS_SDR50: ++ timing = GB_SDIO_TIMING_UHS_SDR50; ++ break; ++ case MMC_TIMING_UHS_SDR104: ++ timing = GB_SDIO_TIMING_UHS_SDR104; ++ break; ++ case MMC_TIMING_UHS_DDR50: ++ timing = GB_SDIO_TIMING_UHS_DDR50; ++ break; ++ case MMC_TIMING_MMC_DDR52: ++ timing = GB_SDIO_TIMING_MMC_DDR52; ++ break; ++ case MMC_TIMING_MMC_HS200: ++ timing = GB_SDIO_TIMING_MMC_HS200; ++ break; ++ case MMC_TIMING_MMC_HS400: ++ timing = GB_SDIO_TIMING_MMC_HS400; ++ break; ++ } ++ request.timing = timing; ++ ++ switch (ios->signal_voltage) { ++ case MMC_SIGNAL_VOLTAGE_330: ++ signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_330; ++ break; ++ case MMC_SIGNAL_VOLTAGE_180: ++ default: ++ signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_180; ++ break; ++ case MMC_SIGNAL_VOLTAGE_120: ++ signal_voltage = GB_SDIO_SIGNAL_VOLTAGE_120; ++ break; ++ } ++ request.signal_voltage = signal_voltage; ++ ++ switch (ios->drv_type) { ++ case MMC_SET_DRIVER_TYPE_A: ++ drv_type = GB_SDIO_SET_DRIVER_TYPE_A; ++ break; ++ case MMC_SET_DRIVER_TYPE_C: ++ drv_type = GB_SDIO_SET_DRIVER_TYPE_C; ++ break; ++ case MMC_SET_DRIVER_TYPE_D: ++ drv_type = GB_SDIO_SET_DRIVER_TYPE_D; ++ break; ++ case MMC_SET_DRIVER_TYPE_B: ++ default: ++ drv_type = GB_SDIO_SET_DRIVER_TYPE_B; ++ break; ++ } ++ request.drv_type = drv_type; ++ ++ ret = gb_sdio_set_ios(host, &request); ++ if (ret < 0) ++ goto out; ++ ++ memcpy(&mmc->ios, ios, sizeof(mmc->ios)); ++ ++out: ++ mutex_unlock(&host->lock); ++} ++ ++static int gb_mmc_get_ro(struct mmc_host *mmc) ++{ ++ struct gb_sdio_host *host = mmc_priv(mmc); ++ ++ mutex_lock(&host->lock); ++ if (host->removed) { ++ mutex_unlock(&host->lock); ++ return -ESHUTDOWN; ++ } ++ mutex_unlock(&host->lock); ++ ++ return host->read_only; ++} ++ ++static int gb_mmc_get_cd(struct mmc_host *mmc) ++{ ++ struct gb_sdio_host *host = mmc_priv(mmc); ++ ++ mutex_lock(&host->lock); ++ if (host->removed) { ++ mutex_unlock(&host->lock); ++ return -ESHUTDOWN; ++ } ++ mutex_unlock(&host->lock); ++ ++ return host->card_present; ++} ++ ++static int gb_mmc_switch_voltage(struct mmc_host *mmc, struct mmc_ios *ios) ++{ ++ return 0; ++} ++ ++static const struct mmc_host_ops gb_sdio_ops = { ++ .request = gb_mmc_request, ++ .set_ios = gb_mmc_set_ios, ++ .get_ro = gb_mmc_get_ro, ++ .get_cd = gb_mmc_get_cd, ++ .start_signal_voltage_switch = gb_mmc_switch_voltage, ++}; ++ ++static int gb_sdio_probe(struct gbphy_device *gbphy_dev, ++ const struct gbphy_device_id *id) ++{ ++ struct gb_connection *connection; ++ struct mmc_host *mmc; ++ struct gb_sdio_host *host; ++ int ret = 0; ++ ++ mmc = mmc_alloc_host(sizeof(*host), &gbphy_dev->dev); ++ if (!mmc) ++ return -ENOMEM; ++ ++ connection = gb_connection_create(gbphy_dev->bundle, ++ le16_to_cpu(gbphy_dev->cport_desc->id), ++ gb_sdio_request_handler); ++ if (IS_ERR(connection)) { ++ ret = PTR_ERR(connection); ++ goto exit_mmc_free; ++ } ++ ++ host = mmc_priv(mmc); ++ host->mmc = mmc; ++ host->removed = true; ++ ++ host->connection = connection; ++ gb_connection_set_data(connection, host); ++ host->gbphy_dev = gbphy_dev; ++ gb_gbphy_set_data(gbphy_dev, host); ++ ++ ret = gb_connection_enable_tx(connection); ++ if (ret) ++ goto exit_connection_destroy; ++ ++ ret = gb_sdio_get_caps(host); ++ if (ret < 0) ++ goto exit_connection_disable; ++ ++ mmc->ops = &gb_sdio_ops; ++ ++ mmc->max_segs = host->mmc->max_blk_count; ++ ++ /* for now we make a map 1:1 between max request and segment size */ ++ mmc->max_req_size = mmc->max_blk_size * mmc->max_blk_count; ++ mmc->max_seg_size = mmc->max_req_size; ++ ++ mutex_init(&host->lock); ++ spin_lock_init(&host->xfer); ++ host->mrq_workqueue = alloc_workqueue("mmc-%s", 0, 1, ++ dev_name(&gbphy_dev->dev)); ++ if (!host->mrq_workqueue) { ++ ret = -ENOMEM; ++ goto exit_connection_disable; ++ } ++ INIT_WORK(&host->mrqwork, gb_sdio_mrq_work); ++ ++ ret = gb_connection_enable(connection); ++ if (ret) ++ goto exit_wq_destroy; ++ ++ ret = mmc_add_host(mmc); ++ if (ret < 0) ++ goto exit_wq_destroy; ++ host->removed = false; ++ ret = _gb_sdio_process_events(host, host->queued_events); ++ host->queued_events = 0; ++ ++ gbphy_runtime_put_autosuspend(gbphy_dev); ++ ++ return ret; ++ ++exit_wq_destroy: ++ destroy_workqueue(host->mrq_workqueue); ++exit_connection_disable: ++ gb_connection_disable(connection); ++exit_connection_destroy: ++ gb_connection_destroy(connection); ++exit_mmc_free: ++ mmc_free_host(mmc); ++ ++ return ret; ++} ++ ++static void gb_sdio_remove(struct gbphy_device *gbphy_dev) ++{ ++ struct gb_sdio_host *host = gb_gbphy_get_data(gbphy_dev); ++ struct gb_connection *connection = host->connection; ++ struct mmc_host *mmc; ++ int ret; ++ ++ ret = gbphy_runtime_get_sync(gbphy_dev); ++ if (ret) ++ gbphy_runtime_get_noresume(gbphy_dev); ++ ++ mutex_lock(&host->lock); ++ host->removed = true; ++ mmc = host->mmc; ++ gb_connection_set_data(connection, NULL); ++ mutex_unlock(&host->lock); ++ ++ flush_workqueue(host->mrq_workqueue); ++ destroy_workqueue(host->mrq_workqueue); ++ gb_connection_disable_rx(connection); ++ mmc_remove_host(mmc); ++ gb_connection_disable(connection); ++ gb_connection_destroy(connection); ++ mmc_free_host(mmc); ++} ++ ++static const struct gbphy_device_id gb_sdio_id_table[] = { ++ { GBPHY_PROTOCOL(GREYBUS_PROTOCOL_SDIO) }, ++ { }, ++}; ++MODULE_DEVICE_TABLE(gbphy, gb_sdio_id_table); ++ ++static struct gbphy_driver sdio_driver = { ++ .name = "sdio", ++ .probe = gb_sdio_probe, ++ .remove = gb_sdio_remove, ++ .id_table = gb_sdio_id_table, ++}; ++ ++module_gbphy_driver(sdio_driver); ++MODULE_LICENSE("GPL v2"); |
