aboutsummaryrefslogtreecommitdiffstats
path: root/greybus_gpio.patch
diff options
authorGreg Kroah-Hartman <gregkh@linuxfoundation.org>2016-09-15 13:54:11 +0200
committerGreg Kroah-Hartman <gregkh@linuxfoundation.org>2016-09-15 13:54:11 +0200
commitc1aa01c96e590714d99c5f17cfae1b14dec8bdee (patch)
tree5811422f65efdff3fd89fdaf9a8c6d7d1cc4d06b /greybus_gpio.patch
parent4758ee8bc89a86c2110b9e85878538ced8045ef5 (diff)
downloadpatches-c1aa01c96e590714d99c5f17cfae1b14dec8bdee.tar.gz
greybus patches
Diffstat (limited to 'greybus_gpio.patch')
-rw-r--r--greybus_gpio.patch774
1 files changed, 774 insertions, 0 deletions
diff --git a/greybus_gpio.patch b/greybus_gpio.patch
new file mode 100644
index 00000000000000..743837044b239d
--- /dev/null
+++ b/greybus_gpio.patch
@@ -0,0 +1,774 @@
+---
+ drivers/greybus/gpio.c | 767 +++++++++++++++++++++++++++++++++++++++++++++++++
+ 1 file changed, 767 insertions(+)
+
+--- /dev/null
++++ b/drivers/greybus/gpio.c
+@@ -0,0 +1,767 @@
++/*
++ * GPIO Greybus driver.
++ *
++ * Copyright 2014 Google Inc.
++ * Copyright 2014 Linaro Ltd.
++ *
++ * Released under the GPLv2 only.
++ */
++
++#include <linux/kernel.h>
++#include <linux/module.h>
++#include <linux/slab.h>
++#include <linux/gpio.h>
++#include <linux/irq.h>
++#include <linux/irqdomain.h>
++#include <linux/mutex.h>
++
++#include "greybus.h"
++#include "gbphy.h"
++
++struct gb_gpio_line {
++ /* The following has to be an array of line_max entries */
++ /* --> make them just a flags field */
++ u8 active: 1,
++ direction: 1, /* 0 = output, 1 = input */
++ value: 1; /* 0 = low, 1 = high */
++ u16 debounce_usec;
++
++ u8 irq_type;
++ bool irq_type_pending;
++ bool masked;
++ bool masked_pending;
++};
++
++struct gb_gpio_controller {
++ struct gbphy_device *gbphy_dev;
++ struct gb_connection *connection;
++ u8 line_max; /* max line number */
++ struct gb_gpio_line *lines;
++
++ struct gpio_chip chip;
++ struct irq_chip irqc;
++ struct irq_chip *irqchip;
++ struct irq_domain *irqdomain;
++ unsigned int irq_base;
++ irq_flow_handler_t irq_handler;
++ unsigned int irq_default_type;
++ struct mutex irq_lock;
++};
++#define gpio_chip_to_gb_gpio_controller(chip) \
++ container_of(chip, struct gb_gpio_controller, chip)
++#define irq_data_to_gpio_chip(d) (d->domain->host_data)
++
++static int gb_gpio_line_count_operation(struct gb_gpio_controller *ggc)
++{
++ struct gb_gpio_line_count_response response;
++ int ret;
++
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_LINE_COUNT,
++ NULL, 0, &response, sizeof(response));
++ if (!ret)
++ ggc->line_max = response.count;
++ return ret;
++}
++
++static int gb_gpio_activate_operation(struct gb_gpio_controller *ggc, u8 which)
++{
++ struct gb_gpio_activate_request request;
++ struct gbphy_device *gbphy_dev = ggc->gbphy_dev;
++ int ret;
++
++ ret = gbphy_runtime_get_sync(gbphy_dev);
++ if (ret)
++ return ret;
++
++ request.which = which;
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_ACTIVATE,
++ &request, sizeof(request), NULL, 0);
++ if (ret) {
++ gbphy_runtime_put_autosuspend(gbphy_dev);
++ return ret;
++ }
++
++ ggc->lines[which].active = true;
++
++ return 0;
++}
++
++static void gb_gpio_deactivate_operation(struct gb_gpio_controller *ggc,
++ u8 which)
++{
++ struct gbphy_device *gbphy_dev = ggc->gbphy_dev;
++ struct device *dev = &gbphy_dev->dev;
++ struct gb_gpio_deactivate_request request;
++ int ret;
++
++ request.which = which;
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_DEACTIVATE,
++ &request, sizeof(request), NULL, 0);
++ if (ret) {
++ dev_err(dev, "failed to deactivate gpio %u\n", which);
++ goto out_pm_put;
++ }
++
++ ggc->lines[which].active = false;
++
++out_pm_put:
++ gbphy_runtime_put_autosuspend(gbphy_dev);
++}
++
++static int gb_gpio_get_direction_operation(struct gb_gpio_controller *ggc,
++ u8 which)
++{
++ struct device *dev = &ggc->gbphy_dev->dev;
++ struct gb_gpio_get_direction_request request;
++ struct gb_gpio_get_direction_response response;
++ int ret;
++ u8 direction;
++
++ request.which = which;
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_GET_DIRECTION,
++ &request, sizeof(request),
++ &response, sizeof(response));
++ if (ret)
++ return ret;
++
++ direction = response.direction;
++ if (direction && direction != 1) {
++ dev_warn(dev, "gpio %u direction was %u (should be 0 or 1)\n",
++ which, direction);
++ }
++ ggc->lines[which].direction = direction ? 1 : 0;
++ return 0;
++}
++
++static int gb_gpio_direction_in_operation(struct gb_gpio_controller *ggc,
++ u8 which)
++{
++ struct gb_gpio_direction_in_request request;
++ int ret;
++
++ request.which = which;
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_DIRECTION_IN,
++ &request, sizeof(request), NULL, 0);
++ if (!ret)
++ ggc->lines[which].direction = 1;
++ return ret;
++}
++
++static int gb_gpio_direction_out_operation(struct gb_gpio_controller *ggc,
++ u8 which, bool value_high)
++{
++ struct gb_gpio_direction_out_request request;
++ int ret;
++
++ request.which = which;
++ request.value = value_high ? 1 : 0;
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_DIRECTION_OUT,
++ &request, sizeof(request), NULL, 0);
++ if (!ret)
++ ggc->lines[which].direction = 0;
++ return ret;
++}
++
++static int gb_gpio_get_value_operation(struct gb_gpio_controller *ggc,
++ u8 which)
++{
++ struct device *dev = &ggc->gbphy_dev->dev;
++ struct gb_gpio_get_value_request request;
++ struct gb_gpio_get_value_response response;
++ int ret;
++ u8 value;
++
++ request.which = which;
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_GET_VALUE,
++ &request, sizeof(request),
++ &response, sizeof(response));
++ if (ret) {
++ dev_err(dev, "failed to get value of gpio %u\n", which);
++ return ret;
++ }
++
++ value = response.value;
++ if (value && value != 1) {
++ dev_warn(dev, "gpio %u value was %u (should be 0 or 1)\n",
++ which, value);
++ }
++ ggc->lines[which].value = value ? 1 : 0;
++ return 0;
++}
++
++static void gb_gpio_set_value_operation(struct gb_gpio_controller *ggc,
++ u8 which, bool value_high)
++{
++ struct device *dev = &ggc->gbphy_dev->dev;
++ struct gb_gpio_set_value_request request;
++ int ret;
++
++ if (ggc->lines[which].direction == 1) {
++ dev_warn(dev, "refusing to set value of input gpio %u\n",
++ which);
++ return;
++ }
++
++ request.which = which;
++ request.value = value_high ? 1 : 0;
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_SET_VALUE,
++ &request, sizeof(request), NULL, 0);
++ if (ret) {
++ dev_err(dev, "failed to set value of gpio %u\n", which);
++ return;
++ }
++
++ ggc->lines[which].value = request.value;
++}
++
++static int gb_gpio_set_debounce_operation(struct gb_gpio_controller *ggc,
++ u8 which, u16 debounce_usec)
++{
++ struct gb_gpio_set_debounce_request request;
++ int ret;
++
++ request.which = which;
++ request.usec = cpu_to_le16(debounce_usec);
++ ret = gb_operation_sync(ggc->connection, GB_GPIO_TYPE_SET_DEBOUNCE,
++ &request, sizeof(request), NULL, 0);
++ if (!ret)
++ ggc->lines[which].debounce_usec = debounce_usec;
++ return ret;
++}
++
++static void _gb_gpio_irq_mask(struct gb_gpio_controller *ggc, u8 hwirq)
++{
++ struct device *dev = &ggc->gbphy_dev->dev;
++ struct gb_gpio_irq_mask_request request;
++ int ret;
++
++ request.which = hwirq;
++ ret = gb_operation_sync(ggc->connection,
++ GB_GPIO_TYPE_IRQ_MASK,
++ &request, sizeof(request), NULL, 0);
++ if (ret)
++ dev_err(dev, "failed to mask irq: %d\n", ret);
++}
++
++static void _gb_gpio_irq_unmask(struct gb_gpio_controller *ggc, u8 hwirq)
++{
++ struct device *dev = &ggc->gbphy_dev->dev;
++ struct gb_gpio_irq_unmask_request request;
++ int ret;
++
++ request.which = hwirq;
++ ret = gb_operation_sync(ggc->connection,
++ GB_GPIO_TYPE_IRQ_UNMASK,
++ &request, sizeof(request), NULL, 0);
++ if (ret)
++ dev_err(dev, "failed to unmask irq: %d\n", ret);
++}
++
++static void _gb_gpio_irq_set_type(struct gb_gpio_controller *ggc,
++ u8 hwirq, u8 type)
++{
++ struct device *dev = &ggc->gbphy_dev->dev;
++ struct gb_gpio_irq_type_request request;
++ int ret;
++
++ request.which = hwirq;
++ request.type = type;
++
++ ret = gb_operation_sync(ggc->connection,
++ GB_GPIO_TYPE_IRQ_TYPE,
++ &request, sizeof(request), NULL, 0);
++ if (ret)
++ dev_err(dev, "failed to set irq type: %d\n", ret);
++}
++
++static void gb_gpio_irq_mask(struct irq_data *d)
++{
++ struct gpio_chip *chip = irq_data_to_gpio_chip(d);
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++ struct gb_gpio_line *line = &ggc->lines[d->hwirq];
++
++ line->masked = true;
++ line->masked_pending = true;
++}
++
++static void gb_gpio_irq_unmask(struct irq_data *d)
++{
++ struct gpio_chip *chip = irq_data_to_gpio_chip(d);
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++ struct gb_gpio_line *line = &ggc->lines[d->hwirq];
++
++ line->masked = false;
++ line->masked_pending = true;
++}
++
++static int gb_gpio_irq_set_type(struct irq_data *d, unsigned int type)
++{
++ struct gpio_chip *chip = irq_data_to_gpio_chip(d);
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++ struct gb_gpio_line *line = &ggc->lines[d->hwirq];
++ struct device *dev = &ggc->gbphy_dev->dev;
++ u8 irq_type;
++
++ switch (type) {
++ case IRQ_TYPE_NONE:
++ irq_type = GB_GPIO_IRQ_TYPE_NONE;
++ break;
++ case IRQ_TYPE_EDGE_RISING:
++ irq_type = GB_GPIO_IRQ_TYPE_EDGE_RISING;
++ break;
++ case IRQ_TYPE_EDGE_FALLING:
++ irq_type = GB_GPIO_IRQ_TYPE_EDGE_FALLING;
++ break;
++ case IRQ_TYPE_EDGE_BOTH:
++ irq_type = GB_GPIO_IRQ_TYPE_EDGE_BOTH;
++ break;
++ case IRQ_TYPE_LEVEL_LOW:
++ irq_type = GB_GPIO_IRQ_TYPE_LEVEL_LOW;
++ break;
++ case IRQ_TYPE_LEVEL_HIGH:
++ irq_type = GB_GPIO_IRQ_TYPE_LEVEL_HIGH;
++ break;
++ default:
++ dev_err(dev, "unsupported irq type: %u\n", type);
++ return -EINVAL;
++ }
++
++ line->irq_type = irq_type;
++ line->irq_type_pending = true;
++
++ return 0;
++}
++
++static void gb_gpio_irq_bus_lock(struct irq_data *d)
++{
++ struct gpio_chip *chip = irq_data_to_gpio_chip(d);
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ mutex_lock(&ggc->irq_lock);
++}
++
++static void gb_gpio_irq_bus_sync_unlock(struct irq_data *d)
++{
++ struct gpio_chip *chip = irq_data_to_gpio_chip(d);
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++ struct gb_gpio_line *line = &ggc->lines[d->hwirq];
++
++ if (line->irq_type_pending) {
++ _gb_gpio_irq_set_type(ggc, d->hwirq, line->irq_type);
++ line->irq_type_pending = false;
++ }
++
++ if (line->masked_pending) {
++ if (line->masked)
++ _gb_gpio_irq_mask(ggc, d->hwirq);
++ else
++ _gb_gpio_irq_unmask(ggc, d->hwirq);
++ line->masked_pending = false;
++ }
++
++ mutex_unlock(&ggc->irq_lock);
++}
++
++static int gb_gpio_request_handler(struct gb_operation *op)
++{
++ struct gb_connection *connection = op->connection;
++ struct gb_gpio_controller *ggc = gb_connection_get_data(connection);
++ struct device *dev = &ggc->gbphy_dev->dev;
++ struct gb_message *request;
++ struct gb_gpio_irq_event_request *event;
++ u8 type = op->type;
++ int irq;
++ struct irq_desc *desc;
++
++ if (type != GB_GPIO_TYPE_IRQ_EVENT) {
++ dev_err(dev, "unsupported unsolicited request: %u\n", type);
++ return -EINVAL;
++ }
++
++ request = op->request;
++
++ if (request->payload_size < sizeof(*event)) {
++ dev_err(dev, "short event received (%zu < %zu)\n",
++ request->payload_size, sizeof(*event));
++ return -EINVAL;
++ }
++
++ event = request->payload;
++ if (event->which > ggc->line_max) {
++ dev_err(dev, "invalid hw irq: %d\n", event->which);
++ return -EINVAL;
++ }
++
++ irq = irq_find_mapping(ggc->irqdomain, event->which);
++ if (!irq) {
++ dev_err(dev, "failed to find IRQ\n");
++ return -EINVAL;
++ }
++ desc = irq_to_desc(irq);
++ if (!desc) {
++ dev_err(dev, "failed to look up irq\n");
++ return -EINVAL;
++ }
++
++ local_irq_disable();
++ generic_handle_irq_desc(desc);
++ local_irq_enable();
++
++ return 0;
++}
++
++static int gb_gpio_request(struct gpio_chip *chip, unsigned offset)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ return gb_gpio_activate_operation(ggc, (u8)offset);
++}
++
++static void gb_gpio_free(struct gpio_chip *chip, unsigned offset)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ gb_gpio_deactivate_operation(ggc, (u8)offset);
++}
++
++static int gb_gpio_get_direction(struct gpio_chip *chip, unsigned offset)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++ u8 which;
++ int ret;
++
++ which = (u8)offset;
++ ret = gb_gpio_get_direction_operation(ggc, which);
++ if (ret)
++ return ret;
++
++ return ggc->lines[which].direction ? 1 : 0;
++}
++
++static int gb_gpio_direction_input(struct gpio_chip *chip, unsigned offset)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ return gb_gpio_direction_in_operation(ggc, (u8)offset);
++}
++
++static int gb_gpio_direction_output(struct gpio_chip *chip, unsigned offset,
++ int value)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ return gb_gpio_direction_out_operation(ggc, (u8)offset, !!value);
++}
++
++static int gb_gpio_get(struct gpio_chip *chip, unsigned offset)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++ u8 which;
++ int ret;
++
++ which = (u8)offset;
++ ret = gb_gpio_get_value_operation(ggc, which);
++ if (ret)
++ return ret;
++
++ return ggc->lines[which].value;
++}
++
++static void gb_gpio_set(struct gpio_chip *chip, unsigned offset, int value)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ gb_gpio_set_value_operation(ggc, (u8)offset, !!value);
++}
++
++static int gb_gpio_set_debounce(struct gpio_chip *chip, unsigned offset,
++ unsigned debounce)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++ u16 usec;
++
++ if (debounce > U16_MAX)
++ return -EINVAL;
++ usec = (u16)debounce;
++
++ return gb_gpio_set_debounce_operation(ggc, (u8)offset, usec);
++}
++
++static int gb_gpio_controller_setup(struct gb_gpio_controller *ggc)
++{
++ int ret;
++
++ /* Now find out how many lines there are */
++ ret = gb_gpio_line_count_operation(ggc);
++ if (ret)
++ return ret;
++
++ ggc->lines = kcalloc(ggc->line_max + 1, sizeof(*ggc->lines),
++ GFP_KERNEL);
++ if (!ggc->lines)
++ return -ENOMEM;
++
++ return ret;
++}
++
++/**
++ * gb_gpio_irq_map() - maps an IRQ into a GB gpio irqchip
++ * @d: the irqdomain used by this irqchip
++ * @irq: the global irq number used by this GB gpio irqchip irq
++ * @hwirq: the local IRQ/GPIO line offset on this GB gpio
++ *
++ * This function will set up the mapping for a certain IRQ line on a
++ * GB gpio by assigning the GB gpio as chip data, and using the irqchip
++ * stored inside the GB gpio.
++ */
++static int gb_gpio_irq_map(struct irq_domain *domain, unsigned int irq,
++ irq_hw_number_t hwirq)
++{
++ struct gpio_chip *chip = domain->host_data;
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ irq_set_chip_data(irq, ggc);
++ irq_set_chip_and_handler(irq, ggc->irqchip, ggc->irq_handler);
++ irq_set_noprobe(irq);
++ /*
++ * No set-up of the hardware will happen if IRQ_TYPE_NONE
++ * is passed as default type.
++ */
++ if (ggc->irq_default_type != IRQ_TYPE_NONE)
++ irq_set_irq_type(irq, ggc->irq_default_type);
++
++ return 0;
++}
++
++static void gb_gpio_irq_unmap(struct irq_domain *d, unsigned int irq)
++{
++ irq_set_chip_and_handler(irq, NULL, NULL);
++ irq_set_chip_data(irq, NULL);
++}
++
++static const struct irq_domain_ops gb_gpio_domain_ops = {
++ .map = gb_gpio_irq_map,
++ .unmap = gb_gpio_irq_unmap,
++};
++
++/**
++ * gb_gpio_irqchip_remove() - removes an irqchip added to a gb_gpio_controller
++ * @ggc: the gb_gpio_controller to remove the irqchip from
++ *
++ * This is called only from gb_gpio_remove()
++ */
++static void gb_gpio_irqchip_remove(struct gb_gpio_controller *ggc)
++{
++ unsigned int offset;
++
++ /* Remove all IRQ mappings and delete the domain */
++ if (ggc->irqdomain) {
++ for (offset = 0; offset < (ggc->line_max + 1); offset++)
++ irq_dispose_mapping(irq_find_mapping(ggc->irqdomain, offset));
++ irq_domain_remove(ggc->irqdomain);
++ }
++
++ if (ggc->irqchip) {
++ ggc->irqchip = NULL;
++ }
++}
++
++/**
++ * gb_gpio_irqchip_add() - adds an irqchip to a gpio chip
++ * @chip: the gpio chip to add the irqchip to
++ * @irqchip: the irqchip to add to the adapter
++ * @first_irq: if not dynamically assigned, the base (first) IRQ to
++ * allocate gpio irqs from
++ * @handler: the irq handler to use (often a predefined irq core function)
++ * @type: the default type for IRQs on this irqchip, pass IRQ_TYPE_NONE
++ * to have the core avoid setting up any default type in the hardware.
++ *
++ * This function closely associates a certain irqchip with a certain
++ * gpio chip, providing an irq domain to translate the local IRQs to
++ * global irqs, and making sure that the gpio chip
++ * is passed as chip data to all related functions. Driver callbacks
++ * need to use container_of() to get their local state containers back
++ * from the gpio chip passed as chip data. An irqdomain will be stored
++ * in the gpio chip that shall be used by the driver to handle IRQ number
++ * translation. The gpio chip will need to be initialized and registered
++ * before calling this function.
++ */
++static int gb_gpio_irqchip_add(struct gpio_chip *chip,
++ struct irq_chip *irqchip,
++ unsigned int first_irq,
++ irq_flow_handler_t handler,
++ unsigned int type)
++{
++ struct gb_gpio_controller *ggc;
++ unsigned int offset;
++ unsigned irq_base;
++
++ if (!chip || !irqchip)
++ return -EINVAL;
++
++ ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ ggc->irqchip = irqchip;
++ ggc->irq_handler = handler;
++ ggc->irq_default_type = type;
++ ggc->irqdomain = irq_domain_add_simple(NULL,
++ ggc->line_max + 1, first_irq,
++ &gb_gpio_domain_ops, chip);
++ if (!ggc->irqdomain) {
++ ggc->irqchip = NULL;
++ return -EINVAL;
++ }
++
++ /*
++ * Prepare the mapping since the irqchip shall be orthogonal to
++ * any gpio calls. If the first_irq was zero, this is
++ * necessary to allocate descriptors for all IRQs.
++ */
++ for (offset = 0; offset < (ggc->line_max + 1); offset++) {
++ irq_base = irq_create_mapping(ggc->irqdomain, offset);
++ if (offset == 0)
++ ggc->irq_base = irq_base;
++ }
++
++ return 0;
++}
++
++static int gb_gpio_to_irq(struct gpio_chip *chip, unsigned offset)
++{
++ struct gb_gpio_controller *ggc = gpio_chip_to_gb_gpio_controller(chip);
++
++ return irq_find_mapping(ggc->irqdomain, offset);
++}
++
++static int gb_gpio_probe(struct gbphy_device *gbphy_dev,
++ const struct gbphy_device_id *id)
++{
++ struct gb_connection *connection;
++ struct gb_gpio_controller *ggc;
++ struct gpio_chip *gpio;
++ struct irq_chip *irqc;
++ int ret;
++
++ ggc = kzalloc(sizeof(*ggc), GFP_KERNEL);
++ if (!ggc)
++ return -ENOMEM;
++
++ connection = gb_connection_create(gbphy_dev->bundle,
++ le16_to_cpu(gbphy_dev->cport_desc->id),
++ gb_gpio_request_handler);
++ if (IS_ERR(connection)) {
++ ret = PTR_ERR(connection);
++ goto exit_ggc_free;
++ }
++
++ ggc->connection = connection;
++ gb_connection_set_data(connection, ggc);
++ ggc->gbphy_dev = gbphy_dev;
++ gb_gbphy_set_data(gbphy_dev, ggc);
++
++ ret = gb_connection_enable_tx(connection);
++ if (ret)
++ goto exit_connection_destroy;
++
++ ret = gb_gpio_controller_setup(ggc);
++ if (ret)
++ goto exit_connection_disable;
++
++ irqc = &ggc->irqc;
++ irqc->irq_mask = gb_gpio_irq_mask;
++ irqc->irq_unmask = gb_gpio_irq_unmask;
++ irqc->irq_set_type = gb_gpio_irq_set_type;
++ irqc->irq_bus_lock = gb_gpio_irq_bus_lock;
++ irqc->irq_bus_sync_unlock = gb_gpio_irq_bus_sync_unlock;
++ irqc->name = "greybus_gpio";
++
++ mutex_init(&ggc->irq_lock);
++
++ gpio = &ggc->chip;
++
++ gpio->label = "greybus_gpio";
++ gpio->parent = &gbphy_dev->dev;
++ gpio->owner = THIS_MODULE;
++
++ gpio->request = gb_gpio_request;
++ gpio->free = gb_gpio_free;
++ gpio->get_direction = gb_gpio_get_direction;
++ gpio->direction_input = gb_gpio_direction_input;
++ gpio->direction_output = gb_gpio_direction_output;
++ gpio->get = gb_gpio_get;
++ gpio->set = gb_gpio_set;
++ gpio->set_debounce = gb_gpio_set_debounce;
++ gpio->to_irq = gb_gpio_to_irq;
++ gpio->base = -1; /* Allocate base dynamically */
++ gpio->ngpio = ggc->line_max + 1;
++ gpio->can_sleep = true;
++
++ ret = gb_connection_enable(connection);
++ if (ret)
++ goto exit_line_free;
++
++ ret = gpiochip_add(gpio);
++ if (ret) {
++ dev_err(&connection->bundle->dev,
++ "failed to add gpio chip: %d\n", ret);
++ goto exit_line_free;
++ }
++
++ ret = gb_gpio_irqchip_add(gpio, irqc, 0,
++ handle_level_irq, IRQ_TYPE_NONE);
++ if (ret) {
++ dev_err(&connection->bundle->dev,
++ "failed to add irq chip: %d\n", ret);
++ goto exit_gpiochip_remove;
++ }
++
++ gbphy_runtime_put_autosuspend(gbphy_dev);
++ return 0;
++
++exit_gpiochip_remove:
++ gpiochip_remove(gpio);
++exit_line_free:
++ kfree(ggc->lines);
++exit_connection_disable:
++ gb_connection_disable(connection);
++exit_connection_destroy:
++ gb_connection_destroy(connection);
++exit_ggc_free:
++ kfree(ggc);
++ return ret;
++}
++
++static void gb_gpio_remove(struct gbphy_device *gbphy_dev)
++{
++ struct gb_gpio_controller *ggc = gb_gbphy_get_data(gbphy_dev);
++ struct gb_connection *connection = ggc->connection;
++ int ret;
++
++ ret = gbphy_runtime_get_sync(gbphy_dev);
++ if (ret)
++ gbphy_runtime_get_noresume(gbphy_dev);
++
++ gb_connection_disable_rx(connection);
++ gb_gpio_irqchip_remove(ggc);
++ gpiochip_remove(&ggc->chip);
++ gb_connection_disable(connection);
++ gb_connection_destroy(connection);
++ kfree(ggc->lines);
++ kfree(ggc);
++}
++
++static const struct gbphy_device_id gb_gpio_id_table[] = {
++ { GBPHY_PROTOCOL(GREYBUS_PROTOCOL_GPIO) },
++ { },
++};
++MODULE_DEVICE_TABLE(gbphy, gb_gpio_id_table);
++
++static struct gbphy_driver gpio_driver = {
++ .name = "gpio",
++ .probe = gb_gpio_probe,
++ .remove = gb_gpio_remove,
++ .id_table = gb_gpio_id_table,
++};
++
++module_gbphy_driver(gpio_driver);
++MODULE_LICENSE("GPL v2");