]> git.proxmox.com Git - mirror_ubuntu-bionic-kernel.git/commitdiff
greybus: operation: fix connection tear down
authorJohan Hovold <johan@hovoldconsulting.com>
Tue, 14 Jul 2015 13:43:31 +0000 (15:43 +0200)
committerGreg Kroah-Hartman <gregkh@google.com>
Wed, 15 Jul 2015 19:39:13 +0000 (12:39 -0700)
Fix connection tear down, which was done in an unsafe way that could
result in use-after-free as the per-connection list of operations was
iterated without any locking or refcounting.

Specifically, the operations list was iterated without holding any locks or
operation refcounts even though operations were being both removed from
the list and deallocated during per-operation cancellation. Any
operation completing during tear down could also cause corruption.

Change the per-connection operation list to only contain active
operations and use the recently introduced active counter to maintain
the list.

Add new helper that is called on connection tear down to cancel all
outstanding operations in a safe way by using proper locks and making
sure to hold a reference to any operation being cancelled.

Note that by verifying the connection state before incrementing the
active count we can make sure that all active operations have been
cancelled and that no new ones have been started when the helper
returns.

Signed-off-by: Johan Hovold <johan@hovoldconsulting.com>
Signed-off-by: Greg Kroah-Hartman <gregkh@google.com>
drivers/staging/greybus/connection.c
drivers/staging/greybus/connection.h
drivers/staging/greybus/operation.c
drivers/staging/greybus/operation.h

index abc1f861ea28afeb25eaa182c07263afff9c068c..2d19082a7f35e8b09828c94c91d35d4a17d095f2 100644 (file)
@@ -221,23 +221,46 @@ struct gb_connection *gb_connection_create(struct gb_bundle *bundle,
        return connection;
 }
 
+/*
+ * Cancel all active operations on a connection.
+ *
+ * Should only be called during connection tear down.
+ */
+static void gb_connection_cancel_operations(struct gb_connection *connection,
+                                               int errno)
+{
+       struct gb_operation *operation;
+
+       spin_lock_irq(&connection->lock);
+
+       WARN_ON(!list_empty(&connection->operations));
+
+       while (!list_empty(&connection->operations)) {
+               operation = list_last_entry(&connection->operations,
+                                               struct gb_operation, links);
+               gb_operation_get(operation);
+               spin_unlock_irq(&connection->lock);
+
+               gb_operation_cancel(operation, errno);
+               gb_operation_put(operation);
+
+               spin_lock_irq(&connection->lock);
+       }
+       spin_unlock_irq(&connection->lock);
+}
+
 /*
  * Tear down a previously set up connection.
  */
 void gb_connection_destroy(struct gb_connection *connection)
 {
-       struct gb_operation *operation;
-       struct gb_operation *next;
        struct ida *id_map;
 
        if (WARN_ON(!connection))
                return;
 
-       if (WARN_ON(!list_empty(&connection->operations))) {
-               list_for_each_entry_safe(operation, next,
-                                        &connection->operations, links)
-                       gb_operation_cancel(operation, -ESHUTDOWN);
-       }
+       gb_connection_cancel_operations(connection, -ESHUTDOWN);
+
        spin_lock_irq(&gb_connections_lock);
        list_del(&connection->bundle_links);
        list_del(&connection->hd_links);
index c2b71fe7f39767c0699bc224b1b0817c94ac3c2e..fb7a1fb290ac001169febc55b10fbfb6d3a30304 100644 (file)
@@ -37,9 +37,9 @@ struct gb_connection {
 
        spinlock_t                      lock;
        enum gb_connection_state        state;
+       struct list_head                operations;
 
        atomic_t                        op_cycle;
-       struct list_head                operations;
 
        void                            *private;
 };
index 2729b486d68decb7e25db6956a5489fad0ddb629..5cd4665c645cf7c5f4857501a1f594bf9a43954a 100644 (file)
@@ -29,32 +29,65 @@ static struct workqueue_struct *gb_operation_workqueue;
 static DECLARE_WAIT_QUEUE_HEAD(gb_operation_cancellation_queue);
 
 /*
- * Protects access to connection operations lists, as well as
- * updates to operation->errno.
+ * Protects updates to operation->errno.
  */
 static DEFINE_SPINLOCK(gb_operations_lock);
 
 static int gb_operation_response_send(struct gb_operation *operation,
                                        int errno);
 
-/* Caller holds operation reference. */
-static inline void gb_operation_get_active(struct gb_operation *operation)
+/*
+ * Increment operation active count and add to connection list unless the
+ * connection is going away.
+ *
+ * Caller holds operation reference.
+ */
+static int gb_operation_get_active(struct gb_operation *operation)
 {
-       atomic_inc(&operation->active);
+       struct gb_connection *connection = operation->connection;
+       unsigned long flags;
+
+       spin_lock_irqsave(&connection->lock, flags);
+
+       if (connection->state != GB_CONNECTION_STATE_ENABLED) {
+               spin_unlock_irqrestore(&connection->lock, flags);
+               return -ENOTCONN;
+       }
+
+       if (operation->active++ == 0)
+               list_add_tail(&operation->links, &connection->operations);
+
+       spin_unlock_irqrestore(&connection->lock, flags);
+
+       return 0;
 }
 
 /* Caller holds operation reference. */
-static inline void gb_operation_put_active(struct gb_operation *operation)
+static void gb_operation_put_active(struct gb_operation *operation)
 {
-       if (atomic_dec_and_test(&operation->active)) {
+       struct gb_connection *connection = operation->connection;
+       unsigned long flags;
+
+       spin_lock_irqsave(&connection->lock, flags);
+       if (--operation->active == 0) {
+               list_del(&operation->links);
                if (atomic_read(&operation->waiters))
                        wake_up(&gb_operation_cancellation_queue);
        }
+       spin_unlock_irqrestore(&connection->lock, flags);
 }
 
-static inline bool gb_operation_is_active(struct gb_operation *operation)
+static bool gb_operation_is_active(struct gb_operation *operation)
 {
-       return atomic_read(&operation->active);
+       struct gb_connection *connection = operation->connection;
+       unsigned long flags;
+       bool ret;
+
+       spin_lock_irqsave(&connection->lock, flags);
+       ret = operation->active;
+       spin_unlock_irqrestore(&connection->lock, flags);
+
+       return ret;
 }
 
 /*
@@ -150,7 +183,7 @@ gb_operation_find_outgoing(struct gb_connection *connection, u16 operation_id)
        unsigned long flags;
        bool found = false;
 
-       spin_lock_irqsave(&gb_operations_lock, flags);
+       spin_lock_irqsave(&connection->lock, flags);
        list_for_each_entry(operation, &connection->operations, links)
                if (operation->id == operation_id &&
                                !gb_operation_is_incoming(operation)) {
@@ -158,7 +191,7 @@ gb_operation_find_outgoing(struct gb_connection *connection, u16 operation_id)
                        found = true;
                        break;
                }
-       spin_unlock_irqrestore(&gb_operations_lock, flags);
+       spin_unlock_irqrestore(&connection->lock, flags);
 
        return found ? operation : NULL;
 }
@@ -453,7 +486,6 @@ gb_operation_create_common(struct gb_connection *connection, u8 type,
 {
        struct greybus_host_device *hd = connection->hd;
        struct gb_operation *operation;
-       unsigned long flags;
 
        operation = kmem_cache_zalloc(gb_operation_cache, gfp_flags);
        if (!operation)
@@ -479,13 +511,8 @@ gb_operation_create_common(struct gb_connection *connection, u8 type,
        INIT_WORK(&operation->work, gb_operation_work);
        init_completion(&operation->completion);
        kref_init(&operation->kref);
-       atomic_set(&operation->active, 0);
        atomic_set(&operation->waiters, 0);
 
-       spin_lock_irqsave(&gb_operations_lock, flags);
-       list_add_tail(&operation->links, &connection->operations);
-       spin_unlock_irqrestore(&gb_operations_lock, flags);
-
        return operation;
 
 err_request:
@@ -570,10 +597,6 @@ static void _gb_operation_destroy(struct kref *kref)
 
        operation = container_of(kref, struct gb_operation, kref);
 
-       /* XXX Make sure it's not in flight */
-       list_del(&operation->links);
-       spin_unlock(&gb_operations_lock);
-
        if (operation->response)
                gb_operation_message_free(operation->response);
        gb_operation_message_free(operation->request);
@@ -590,8 +613,7 @@ void gb_operation_put(struct gb_operation *operation)
        if (WARN_ON(!operation))
                return;
 
-       kref_put_spinlock_irqsave(&operation->kref, _gb_operation_destroy,
-                       &gb_operations_lock);
+       kref_put(&operation->kref, _gb_operation_destroy);
 }
 EXPORT_SYMBOL_GPL(gb_operation_put);
 
@@ -621,15 +643,14 @@ int gb_operation_request_send(struct gb_operation *operation,
        if (!callback)
                return -EINVAL;
 
-       if (connection->state != GB_CONNECTION_STATE_ENABLED)
-               return -ENOTCONN;
-
        /*
         * First, get an extra reference on the operation.
         * It'll be dropped when the operation completes.
         */
        gb_operation_get(operation);
-       gb_operation_get_active(operation);
+       ret = gb_operation_get_active(operation);
+       if (ret)
+               goto err_put;
 
        /*
         * Record the callback function, which is executed in
@@ -651,10 +672,15 @@ int gb_operation_request_send(struct gb_operation *operation,
        gb_operation_result_set(operation, -EINPROGRESS);
 
        ret = gb_message_send(operation->request, gfp);
-       if (ret) {
-               gb_operation_put_active(operation);
-               gb_operation_put(operation);
-       }
+       if (ret)
+               goto err_put_active;
+
+       return 0;
+
+err_put_active:
+       gb_operation_put_active(operation);
+err_put:
+       gb_operation_put(operation);
 
        return ret;
 }
@@ -705,9 +731,6 @@ static int gb_operation_response_send(struct gb_operation *operation,
        struct gb_connection *connection = operation->connection;
        int ret;
 
-       if (connection->state != GB_CONNECTION_STATE_ENABLED)
-               return -ENOTCONN;
-
        if (!operation->response &&
                        !gb_operation_is_unidirectional(operation)) {
                if (!gb_operation_response_alloc(operation, 0))
@@ -726,16 +749,23 @@ static int gb_operation_response_send(struct gb_operation *operation,
 
        /* Reference will be dropped when message has been sent. */
        gb_operation_get(operation);
-       gb_operation_get_active(operation);
+       ret = gb_operation_get_active(operation);
+       if (ret)
+               goto err_put;
 
        /* Fill in the response header and send it */
        operation->response->header->result = gb_operation_errno_map(errno);
 
        ret = gb_message_send(operation->response, GFP_KERNEL);
-       if (ret) {
-               gb_operation_put_active(operation);
-               gb_operation_put(operation);
-       }
+       if (ret)
+               goto err_put_active;
+
+       return 0;
+
+err_put_active:
+       gb_operation_put_active(operation);
+err_put:
+       gb_operation_put(operation);
 
        return ret;
 }
@@ -785,6 +815,7 @@ static void gb_connection_recv_request(struct gb_connection *connection,
                                       void *data, size_t size)
 {
        struct gb_operation *operation;
+       int ret;
 
        operation = gb_operation_create_incoming(connection, operation_id,
                                                type, data, size);
@@ -793,7 +824,11 @@ static void gb_connection_recv_request(struct gb_connection *connection,
                return;         /* XXX Respond with pre-allocated ENOMEM */
        }
 
-       gb_operation_get_active(operation);
+       ret = gb_operation_get_active(operation);
+       if (ret) {
+               gb_operation_put(operation);
+               return;
+       }
 
        /*
         * The initial reference to the operation will be dropped when the
index c8aaf90a006ac9be2db80adfab79797efb42bd64..26ecd66710a31858ef85e0ee19205bda4fa8a5fb 100644 (file)
@@ -127,9 +127,9 @@ struct gb_operation {
        struct completion       completion;
 
        struct kref             kref;
-       atomic_t                active;
        atomic_t                waiters;
 
+       int                     active;
        struct list_head        links;          /* connection->operations */
 };