greybus: Revert "connection: switch to using spin_lock_irqsave/spin_lock_irqrestore...
[cascardo/linux.git] / drivers / staging / greybus / connection.c
index 9b8112b..f261468 100644 (file)
@@ -10,6 +10,7 @@
 #include <linux/workqueue.h>
 
 #include "greybus.h"
+#include "greybus_trace.h"
 
 
 static void gb_connection_kref_release(struct kref *kref);
@@ -20,8 +21,7 @@ static DEFINE_MUTEX(gb_connection_mutex);
 
 
 /* Caller holds gb_connection_mutex. */
-static struct gb_connection *
-gb_connection_intf_find(struct gb_interface *intf, u16 cport_id)
+static bool gb_connection_cport_in_use(struct gb_interface *intf, u16 cport_id)
 {
        struct gb_host_device *hd = intf->hd;
        struct gb_connection *connection;
@@ -29,19 +29,23 @@ gb_connection_intf_find(struct gb_interface *intf, u16 cport_id)
        list_for_each_entry(connection, &hd->connections, hd_links) {
                if (connection->intf == intf &&
                                connection->intf_cport_id == cport_id)
-                       return connection;
+                       return true;
        }
 
-       return NULL;
+       return false;
 }
 
 static void gb_connection_get(struct gb_connection *connection)
 {
        kref_get(&connection->kref);
+
+       trace_gb_connection_get(connection);
 }
 
 static void gb_connection_put(struct gb_connection *connection)
 {
+       trace_gb_connection_put(connection);
+
        kref_put(&connection->kref, gb_connection_kref_release);
 }
 
@@ -76,6 +80,8 @@ void greybus_data_rcvd(struct gb_host_device *hd, u16 cport_id,
 {
        struct gb_connection *connection;
 
+       trace_gb_hd_in(hd);
+
        connection = gb_connection_hd_find(hd, cport_id);
        if (!connection) {
                dev_err(&hd->dev,
@@ -93,6 +99,8 @@ static void gb_connection_kref_release(struct kref *kref)
 
        connection = container_of(kref, struct gb_connection, kref);
 
+       trace_gb_connection_release(connection);
+
        kfree(connection);
 }
 
@@ -112,13 +120,14 @@ static void gb_connection_init_name(struct gb_connection *connection)
 }
 
 /*
- * gb_connection_create() - create a Greybus connection
+ * _gb_connection_create() - create a Greybus connection
  * @hd:                        host device of the connection
  * @hd_cport_id:       host-device cport id, or -1 for dynamic allocation
  * @intf:              remote interface, or NULL for static connections
  * @bundle:            remote-interface bundle (may be NULL)
  * @cport_id:          remote-interface cport id, or 0 for static connections
- * @protocol_id:       protocol id
+ * @handler:           request handler (may be NULL)
+ * @flags:             connection flags
  *
  * Create a Greybus connection, representing the bidirectional link
  * between a CPort on a (local) Greybus host device and a CPort on
@@ -130,52 +139,50 @@ static void gb_connection_init_name(struct gb_connection *connection)
  * Serialised against concurrent create and destroy using the
  * gb_connection_mutex.
  *
- * Return: A pointer to the new connection if successful, or NULL otherwise.
+ * Return: A pointer to the new connection if successful, or an ERR_PTR
+ * otherwise.
  */
 static struct gb_connection *
-gb_connection_create(struct gb_host_device *hd, int hd_cport_id,
+_gb_connection_create(struct gb_host_device *hd, int hd_cport_id,
                                struct gb_interface *intf,
                                struct gb_bundle *bundle, int cport_id,
-                               u8 protocol_id)
+                               gb_request_handler_t handler,
+                               unsigned long flags)
 {
        struct gb_connection *connection;
-       struct ida *id_map = &hd->cport_id_map;
-       int ida_start, ida_end;
-
-       if (hd_cport_id < 0) {
-               ida_start = 0;
-               ida_end = hd->num_cports;
-       } else if (hd_cport_id < hd->num_cports) {
-               ida_start = hd_cport_id;
-               ida_end = hd_cport_id + 1;
-       } else {
-               dev_err(&hd->dev, "cport %d not available\n", hd_cport_id);
-               return NULL;
-       }
+       unsigned long irqflags;
+       int ret;
 
        mutex_lock(&gb_connection_mutex);
 
-       if (intf && gb_connection_intf_find(intf, cport_id)) {
+       if (intf && gb_connection_cport_in_use(intf, cport_id)) {
                dev_err(&intf->dev, "cport %u already in use\n", cport_id);
+               ret = -EBUSY;
                goto err_unlock;
        }
 
-       hd_cport_id = ida_simple_get(id_map, ida_start, ida_end, GFP_KERNEL);
-       if (hd_cport_id < 0)
+       ret = gb_hd_cport_allocate(hd, hd_cport_id, flags);
+       if (ret < 0) {
+               dev_err(&hd->dev, "failed to allocate cport: %d\n", ret);
                goto err_unlock;
+       }
+       hd_cport_id = ret;
 
        connection = kzalloc(sizeof(*connection), GFP_KERNEL);
-       if (!connection)
-               goto err_remove_ida;
+       if (!connection) {
+               ret = -ENOMEM;
+               goto err_hd_cport_release;
+       }
 
        connection->hd_cport_id = hd_cport_id;
        connection->intf_cport_id = cport_id;
        connection->hd = hd;
        connection->intf = intf;
-
-       connection->protocol_id = protocol_id;
-
        connection->bundle = bundle;
+       connection->handler = handler;
+       connection->flags = flags;
+       if (intf && (intf->quirks & GB_INTERFACE_QUIRK_NO_CPORT_FEATURES))
+               connection->flags |= GB_CONNECTION_FLAG_NO_FLOWCTRL;
        connection->state = GB_CONNECTION_STATE_DISABLED;
 
        atomic_set(&connection->op_cycle, 0);
@@ -185,14 +192,16 @@ gb_connection_create(struct gb_host_device *hd, int hd_cport_id,
 
        connection->wq = alloc_workqueue("%s:%d", WQ_UNBOUND, 1,
                                         dev_name(&hd->dev), hd_cport_id);
-       if (!connection->wq)
+       if (!connection->wq) {
+               ret = -ENOMEM;
                goto err_free_connection;
+       }
 
        kref_init(&connection->kref);
 
        gb_connection_init_name(connection);
 
-       spin_lock_irq(&gb_connections_lock);
+       spin_lock_irqsave(&gb_connections_lock, irqflags);
        list_add(&connection->hd_links, &hd->connections);
 
        if (bundle)
@@ -200,37 +209,75 @@ gb_connection_create(struct gb_host_device *hd, int hd_cport_id,
        else
                INIT_LIST_HEAD(&connection->bundle_links);
 
-       spin_unlock_irq(&gb_connections_lock);
+       spin_unlock_irqrestore(&gb_connections_lock, irqflags);
 
        mutex_unlock(&gb_connection_mutex);
 
+       trace_gb_connection_create(connection);
+
        return connection;
 
 err_free_connection:
        kfree(connection);
-err_remove_ida:
-       ida_simple_remove(id_map, hd_cport_id);
+err_hd_cport_release:
+       gb_hd_cport_release(hd, hd_cport_id);
 err_unlock:
        mutex_unlock(&gb_connection_mutex);
 
-       return NULL;
+       return ERR_PTR(ret);
+}
+
+struct gb_connection *
+gb_connection_create_static(struct gb_host_device *hd, u16 hd_cport_id,
+                                       gb_request_handler_t handler)
+{
+       return _gb_connection_create(hd, hd_cport_id, NULL, NULL, 0, handler,
+                                       GB_CONNECTION_FLAG_HIGH_PRIO);
+}
+
+struct gb_connection *
+gb_connection_create_control(struct gb_interface *intf)
+{
+       return _gb_connection_create(intf->hd, -1, intf, NULL, 0, NULL,
+                                       GB_CONNECTION_FLAG_CONTROL |
+                                       GB_CONNECTION_FLAG_HIGH_PRIO);
+}
+
+struct gb_connection *
+gb_connection_create(struct gb_bundle *bundle, u16 cport_id,
+                                       gb_request_handler_t handler)
+{
+       struct gb_interface *intf = bundle->intf;
+
+       return _gb_connection_create(intf->hd, -1, intf, bundle, cport_id,
+                                       handler, 0);
 }
+EXPORT_SYMBOL_GPL(gb_connection_create);
 
 struct gb_connection *
-gb_connection_create_static(struct gb_host_device *hd, u16 hd_cport_id)
+gb_connection_create_flags(struct gb_bundle *bundle, u16 cport_id,
+                                       gb_request_handler_t handler,
+                                       unsigned long flags)
 {
-       return gb_connection_create(hd, hd_cport_id, NULL, NULL, 0, 0);
+       struct gb_interface *intf = bundle->intf;
+
+       if (WARN_ON_ONCE(flags & GB_CONNECTION_FLAG_CORE_MASK))
+               flags &= ~GB_CONNECTION_FLAG_CORE_MASK;
+
+       return _gb_connection_create(intf->hd, -1, intf, bundle, cport_id,
+                                       handler, flags);
 }
+EXPORT_SYMBOL_GPL(gb_connection_create_flags);
 
 struct gb_connection *
-gb_connection_create_dynamic(struct gb_interface *intf,
-                                       struct gb_bundle *bundle,
-                                       u16 cport_id, u8 protocol_id)
+gb_connection_create_offloaded(struct gb_bundle *bundle, u16 cport_id,
+                                       unsigned long flags)
 {
-       return gb_connection_create(intf->hd, -1, intf, bundle, cport_id,
-                                                               protocol_id);
+       flags |= GB_CONNECTION_FLAG_OFFLOADED;
+
+       return gb_connection_create_flags(bundle, cport_id, NULL, flags);
 }
-EXPORT_SYMBOL_GPL(gb_connection_create_dynamic);
+EXPORT_SYMBOL_GPL(gb_connection_create_offloaded);
 
 static int gb_connection_hd_cport_enable(struct gb_connection *connection)
 {
@@ -240,10 +287,11 @@ static int gb_connection_hd_cport_enable(struct gb_connection *connection)
        if (!hd->driver->cport_enable)
                return 0;
 
-       ret = hd->driver->cport_enable(hd, connection->hd_cport_id);
+       ret = hd->driver->cport_enable(hd, connection->hd_cport_id,
+                                       connection->flags);
        if (ret) {
-               dev_err(&hd->dev,
-                       "failed to enable host cport: %d\n", ret);
+               dev_err(&hd->dev, "%s: failed to enable host cport: %d\n",
+                               connection->name, ret);
                return ret;
        }
 
@@ -253,11 +301,64 @@ static int gb_connection_hd_cport_enable(struct gb_connection *connection)
 static void gb_connection_hd_cport_disable(struct gb_connection *connection)
 {
        struct gb_host_device *hd = connection->hd;
+       int ret;
 
        if (!hd->driver->cport_disable)
                return;
 
-       hd->driver->cport_disable(hd, connection->hd_cport_id);
+       ret = hd->driver->cport_disable(hd, connection->hd_cport_id);
+       if (ret) {
+               dev_err(&hd->dev, "%s: failed to disable host cport: %d\n",
+                               connection->name, ret);
+       }
+}
+
+static int gb_connection_hd_cport_flush(struct gb_connection *connection)
+{
+       struct gb_host_device *hd = connection->hd;
+       int ret;
+
+       if (!hd->driver->cport_flush)
+               return 0;
+
+       ret = hd->driver->cport_flush(hd, connection->hd_cport_id);
+       if (ret) {
+               dev_err(&hd->dev, "%s: failed to flush host cport: %d\n",
+                               connection->name, ret);
+               return ret;
+       }
+
+       return 0;
+}
+
+static int
+gb_connection_hd_cport_features_enable(struct gb_connection *connection)
+{
+       struct gb_host_device *hd = connection->hd;
+       int ret;
+
+       if (!hd->driver->cport_features_enable)
+               return 0;
+
+       ret = hd->driver->cport_features_enable(hd, connection->hd_cport_id);
+       if (ret) {
+               dev_err(&hd->dev, "%s: failed to enable CPort features: %d\n",
+                       connection->name, ret);
+               return ret;
+       }
+
+       return 0;
+}
+
+static void
+gb_connection_hd_cport_features_disable(struct gb_connection *connection)
+{
+       struct gb_host_device *hd = connection->hd;
+
+       if (!hd->driver->cport_features_disable)
+               return;
+
+       hd->driver->cport_features_disable(hd, connection->hd_cport_id);
 }
 
 /*
@@ -269,18 +370,31 @@ gb_connection_svc_connection_create(struct gb_connection *connection)
 {
        struct gb_host_device *hd = connection->hd;
        struct gb_interface *intf;
+       u8 cport_flags;
        int ret;
 
        if (gb_connection_is_static(connection))
                return 0;
 
        intf = connection->intf;
+
+       /*
+        * Enable either E2EFC or CSD, unless no flow control is requested.
+        */
+       cport_flags = GB_SVC_CPORT_FLAG_CSV_N;
+       if (gb_connection_flow_control_disabled(connection)) {
+               cport_flags |= GB_SVC_CPORT_FLAG_CSD_N;
+       } else if (gb_connection_e2efc_enabled(connection)) {
+               cport_flags |= GB_SVC_CPORT_FLAG_CSD_N |
+                               GB_SVC_CPORT_FLAG_E2EFC;
+       }
+
        ret = gb_svc_connection_create(hd->svc,
                        hd->svc->ap_intf_id,
                        connection->hd_cport_id,
                        intf->interface_id,
                        connection->intf_cport_id,
-                       intf->boot_over_unipro);
+                       cport_flags);
        if (ret) {
                dev_err(&connection->hd->dev,
                        "%s: failed to create svc connection: %d\n",
@@ -304,6 +418,21 @@ gb_connection_svc_connection_destroy(struct gb_connection *connection)
                                  connection->intf_cport_id);
 }
 
+static void
+gb_connection_svc_connection_quiescing(struct gb_connection *connection)
+{
+       struct gb_host_device *hd = connection->hd;
+
+       if (gb_connection_is_static(connection))
+               return;
+
+       gb_svc_connection_quiescing(hd->svc,
+                                       hd->svc->ap_intf_id,
+                                       connection->hd_cport_id,
+                                       connection->intf->interface_id,
+                                       connection->intf_cport_id);
+}
+
 /* Inform Interface about active CPorts */
 static int gb_connection_control_connected(struct gb_connection *connection)
 {
@@ -314,11 +443,21 @@ static int gb_connection_control_connected(struct gb_connection *connection)
        if (gb_connection_is_static(connection))
                return 0;
 
-       control = connection->intf->control;
+       /*
+        * HACK: Suppress connected request for the offloaded camera
+        * connection as it is currently not supported by firmware. Note that
+        * the corresponding non-fatal disconnected event is still sent.
+        */
+       if (gb_connection_is_offloaded(connection) &&
+                       connection->flags & GB_CONNECTION_FLAG_CDSI1) {
+               return 0;
+       }
 
-       if (connection == control->connection)
+       if (gb_connection_is_control(connection))
                return 0;
 
+       control = connection->intf->control;
+
        ret = gb_control_connected_operation(control, cport_id);
        if (ret) {
                dev_err(&connection->bundle->dev,
@@ -329,7 +468,26 @@ static int gb_connection_control_connected(struct gb_connection *connection)
        return 0;
 }
 
-/* Inform Interface about inactive CPorts */
+static void
+gb_connection_control_disconnecting(struct gb_connection *connection)
+{
+       struct gb_control *control;
+       u16 cport_id = connection->intf_cport_id;
+       int ret;
+
+       if (gb_connection_is_static(connection))
+               return;
+
+       control = connection->intf->control;
+
+       ret = gb_control_disconnecting_operation(control, cport_id);
+       if (ret) {
+               dev_err(&connection->hd->dev,
+                               "%s: failed to send disconnecting: %d\n",
+                               connection->name, ret);
+       }
+}
+
 static void
 gb_connection_control_disconnected(struct gb_connection *connection)
 {
@@ -342,8 +500,20 @@ gb_connection_control_disconnected(struct gb_connection *connection)
 
        control = connection->intf->control;
 
-       if (connection == control->connection)
+       if (gb_connection_is_control(connection)) {
+               if (connection->mode_switch) {
+                       ret = gb_control_mode_switch_operation(control);
+                       if (ret) {
+                               /*
+                                * Allow mode switch to time out waiting for
+                                * mailbox event.
+                                */
+                               return;
+                       }
+               }
+
                return;
+       }
 
        ret = gb_control_disconnected_operation(control, cport_id);
        if (ret) {
@@ -352,13 +522,60 @@ gb_connection_control_disconnected(struct gb_connection *connection)
        }
 }
 
+static int gb_connection_ping_operation(struct gb_connection *connection)
+{
+       struct gb_operation *operation;
+       int ret;
+
+       operation = gb_operation_create_core(connection,
+                                               GB_REQUEST_TYPE_PING,
+                                               0, 0, 0,
+                                               GFP_KERNEL);
+       if (!operation)
+               return -ENOMEM;
+
+       ret = gb_operation_request_send_sync(operation);
+
+       gb_operation_put(operation);
+
+       return ret;
+}
+
+static int gb_connection_ping(struct gb_connection *connection)
+{
+       struct gb_host_device *hd = connection->hd;
+       int ret;
+
+       if (gb_connection_is_static(connection))
+               return 0;
+
+       if (gb_connection_is_offloaded(connection)) {
+               if (!hd->driver->cport_ping)
+                       return 0;
+
+               ret = hd->driver->cport_ping(hd, connection->intf_cport_id);
+       } else {
+               ret = gb_connection_ping_operation(connection);
+       }
+
+       if (ret) {
+               dev_err(&hd->dev, "%s: failed to send ping: %d\n",
+                               connection->name, ret);
+               return ret;
+       }
+
+       return 0;
+}
+
 /*
  * Cancel all active operations on a connection.
  *
- * Locking: Called with connection lock held and state set to DISABLED.
+ * Locking: Called with connection lock held and state set to DISABLED or
+ * DISCONNECTING.
  */
 static void gb_connection_cancel_operations(struct gb_connection *connection,
                                                int errno)
+       __must_hold(&connection->lock)
 {
        struct gb_operation *operation;
 
@@ -387,6 +604,7 @@ static void gb_connection_cancel_operations(struct gb_connection *connection,
 static void
 gb_connection_flush_incoming_operations(struct gb_connection *connection,
                                                int errno)
+       __must_hold(&connection->lock)
 {
        struct gb_operation *operation;
        bool incoming;
@@ -415,39 +633,46 @@ gb_connection_flush_incoming_operations(struct gb_connection *connection,
        }
 }
 
-int gb_connection_enable(struct gb_connection *connection,
-                               gb_request_handler_t handler)
+/*
+ * _gb_connection_enable() - enable a connection
+ * @connection:                connection to enable
+ * @rx:                        whether to enable incoming requests
+ *
+ * Connection-enable helper for DISABLED->ENABLED, DISABLED->ENABLED_TX, and
+ * ENABLED_TX->ENABLED state transitions.
+ *
+ * Locking: Caller holds connection->mutex.
+ */
+static int _gb_connection_enable(struct gb_connection *connection, bool rx)
 {
        int ret;
 
-       mutex_lock(&connection->mutex);
-
-       if (connection->state == GB_CONNECTION_STATE_ENABLED)
-               goto out_unlock;
-
+       /* Handle ENABLED_TX -> ENABLED transitions. */
        if (connection->state == GB_CONNECTION_STATE_ENABLED_TX) {
-               if (!handler)
-                       goto out_unlock;
+               if (!(connection->handler && rx))
+                       return 0;
 
                spin_lock_irq(&connection->lock);
-               connection->handler = handler;
                connection->state = GB_CONNECTION_STATE_ENABLED;
                spin_unlock_irq(&connection->lock);
 
-               goto out_unlock;
+               return 0;
        }
 
        ret = gb_connection_hd_cport_enable(connection);
        if (ret)
-               goto err_unlock;
+               return ret;
 
        ret = gb_connection_svc_connection_create(connection);
        if (ret)
                goto err_hd_cport_disable;
 
+       ret = gb_connection_hd_cport_features_enable(connection);
+       if (ret)
+               goto err_svc_connection_destroy;
+
        spin_lock_irq(&connection->lock);
-       connection->handler = handler;
-       if (handler)
+       if (connection->handler && rx)
                connection->state = GB_CONNECTION_STATE_ENABLED;
        else
                connection->state = GB_CONNECTION_STATE_ENABLED_TX;
@@ -455,30 +680,80 @@ int gb_connection_enable(struct gb_connection *connection,
 
        ret = gb_connection_control_connected(connection);
        if (ret)
-               goto err_svc_destroy;
-
-out_unlock:
-       mutex_unlock(&connection->mutex);
+               goto err_control_disconnecting;
 
        return 0;
 
-err_svc_destroy:
+err_control_disconnecting:
+       gb_connection_control_disconnecting(connection);
+
        spin_lock_irq(&connection->lock);
-       connection->state = GB_CONNECTION_STATE_DISABLED;
+       connection->state = GB_CONNECTION_STATE_DISCONNECTING;
        gb_connection_cancel_operations(connection, -ESHUTDOWN);
-       connection->handler = NULL;
        spin_unlock_irq(&connection->lock);
 
+       /* Transmit queue should already be empty. */
+       gb_connection_hd_cport_flush(connection);
+
+       gb_connection_ping(connection);
+       gb_connection_hd_cport_features_disable(connection);
+       gb_connection_svc_connection_quiescing(connection);
+       gb_connection_ping(connection);
+       gb_connection_control_disconnected(connection);
+       connection->state = GB_CONNECTION_STATE_DISABLED;
+err_svc_connection_destroy:
        gb_connection_svc_connection_destroy(connection);
 err_hd_cport_disable:
        gb_connection_hd_cport_disable(connection);
-err_unlock:
+
+       return ret;
+}
+
+int gb_connection_enable(struct gb_connection *connection)
+{
+       int ret = 0;
+
+       mutex_lock(&connection->mutex);
+
+       if (connection->state == GB_CONNECTION_STATE_ENABLED)
+               goto out_unlock;
+
+       ret = _gb_connection_enable(connection, true);
+       if (!ret)
+               trace_gb_connection_enable(connection);
+
+out_unlock:
        mutex_unlock(&connection->mutex);
 
        return ret;
 }
 EXPORT_SYMBOL_GPL(gb_connection_enable);
 
+int gb_connection_enable_tx(struct gb_connection *connection)
+{
+       int ret = 0;
+
+       mutex_lock(&connection->mutex);
+
+       if (connection->state == GB_CONNECTION_STATE_ENABLED) {
+               ret = -EINVAL;
+               goto out_unlock;
+       }
+
+       if (connection->state == GB_CONNECTION_STATE_ENABLED_TX)
+               goto out_unlock;
+
+       ret = _gb_connection_enable(connection, false);
+       if (!ret)
+               trace_gb_connection_enable(connection);
+
+out_unlock:
+       mutex_unlock(&connection->mutex);
+
+       return ret;
+}
+EXPORT_SYMBOL_GPL(gb_connection_enable_tx);
+
 void gb_connection_disable_rx(struct gb_connection *connection)
 {
        mutex_lock(&connection->mutex);
@@ -490,12 +765,26 @@ void gb_connection_disable_rx(struct gb_connection *connection)
        }
        connection->state = GB_CONNECTION_STATE_ENABLED_TX;
        gb_connection_flush_incoming_operations(connection, -ESHUTDOWN);
-       connection->handler = NULL;
        spin_unlock_irq(&connection->lock);
 
+       trace_gb_connection_disable(connection);
+
 out_unlock:
        mutex_unlock(&connection->mutex);
 }
+EXPORT_SYMBOL_GPL(gb_connection_disable_rx);
+
+void gb_connection_mode_switch_prepare(struct gb_connection *connection)
+{
+       connection->mode_switch = true;
+}
+
+void gb_connection_mode_switch_complete(struct gb_connection *connection)
+{
+       gb_connection_svc_connection_destroy(connection);
+       gb_connection_hd_cport_disable(connection);
+       connection->mode_switch = false;
+}
 
 void gb_connection_disable(struct gb_connection *connection)
 {
@@ -504,41 +793,83 @@ void gb_connection_disable(struct gb_connection *connection)
        if (connection->state == GB_CONNECTION_STATE_DISABLED)
                goto out_unlock;
 
+       trace_gb_connection_disable(connection);
+
+       gb_connection_control_disconnecting(connection);
+
+       spin_lock_irq(&connection->lock);
+       connection->state = GB_CONNECTION_STATE_DISCONNECTING;
+       gb_connection_cancel_operations(connection, -ESHUTDOWN);
+       spin_unlock_irq(&connection->lock);
+
+       gb_connection_hd_cport_flush(connection);
+
+       gb_connection_ping(connection);
+       gb_connection_hd_cport_features_disable(connection);
+       gb_connection_svc_connection_quiescing(connection);
+       gb_connection_ping(connection);
+
        gb_connection_control_disconnected(connection);
 
+       connection->state = GB_CONNECTION_STATE_DISABLED;
+
+       /* control-connection tear down is deferred when mode switching */
+       if (!connection->mode_switch) {
+               gb_connection_svc_connection_destroy(connection);
+               gb_connection_hd_cport_disable(connection);
+       }
+
+out_unlock:
+       mutex_unlock(&connection->mutex);
+}
+EXPORT_SYMBOL_GPL(gb_connection_disable);
+
+/* Disable a connection without communicating with the remote end. */
+void gb_connection_disable_forced(struct gb_connection *connection)
+{
+       mutex_lock(&connection->mutex);
+
+       if (connection->state == GB_CONNECTION_STATE_DISABLED)
+               goto out_unlock;
+
+       trace_gb_connection_disable(connection);
+
        spin_lock_irq(&connection->lock);
        connection->state = GB_CONNECTION_STATE_DISABLED;
        gb_connection_cancel_operations(connection, -ESHUTDOWN);
-       connection->handler = NULL;
        spin_unlock_irq(&connection->lock);
 
+       gb_connection_hd_cport_flush(connection);
+       gb_connection_hd_cport_features_disable(connection);
        gb_connection_svc_connection_destroy(connection);
        gb_connection_hd_cport_disable(connection);
 
 out_unlock:
        mutex_unlock(&connection->mutex);
 }
-EXPORT_SYMBOL_GPL(gb_connection_disable);
+EXPORT_SYMBOL_GPL(gb_connection_disable_forced);
 
 /* Caller must have disabled the connection before destroying it. */
 void gb_connection_destroy(struct gb_connection *connection)
 {
-       struct ida *id_map;
+       unsigned long flags;
 
        if (!connection)
                return;
 
+       if (WARN_ON(connection->state != GB_CONNECTION_STATE_DISABLED))
+               gb_connection_disable(connection);
+
        mutex_lock(&gb_connection_mutex);
 
-       spin_lock_irq(&gb_connections_lock);
+       spin_lock_irqsave(&gb_connections_lock, flags);
        list_del(&connection->bundle_links);
        list_del(&connection->hd_links);
-       spin_unlock_irq(&gb_connections_lock);
+       spin_unlock_irqrestore(&gb_connections_lock, flags);
 
        destroy_workqueue(connection->wq);
 
-       id_map = &connection->hd->cport_id_map;
-       ida_simple_remove(id_map, connection->hd_cport_id);
+       gb_hd_cport_release(connection->hd, connection->hd_cport_id);
        connection->hd_cport_id = CPORT_ID_BAD;
 
        mutex_unlock(&gb_connection_mutex);