]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blobdiff - drivers/net/ethernet/qlogic/qed/qed_spq.c
qed*: Semantic changes
[mirror_ubuntu-artful-kernel.git] / drivers / net / ethernet / qlogic / qed / qed_spq.c
index d73456eab1d7e660a5a6ce404c3ceb00465fa62f..5e68a33b209728c9844a170ea17e8964ff489634 100644 (file)
@@ -41,8 +41,7 @@
 ***************************************************************************/
 static void qed_spq_blocking_cb(struct qed_hwfn *p_hwfn,
                                void *cookie,
-                               union event_ring_data *data,
-                               u8 fw_return_code)
+                               union event_ring_data *data, u8 fw_return_code)
 {
        struct qed_spq_comp_done *comp_done;
 
@@ -109,9 +108,8 @@ static int qed_spq_block(struct qed_hwfn *p_hwfn,
 /***************************************************************************
 * SPQ entries inner API
 ***************************************************************************/
-static int
-qed_spq_fill_entry(struct qed_hwfn *p_hwfn,
-                  struct qed_spq_entry *p_ent)
+static int qed_spq_fill_entry(struct qed_hwfn *p_hwfn,
+                             struct qed_spq_entry *p_ent)
 {
        p_ent->flags = 0;
 
@@ -189,8 +187,7 @@ static void qed_spq_hw_initialize(struct qed_hwfn *p_hwfn,
 }
 
 static int qed_spq_hw_post(struct qed_hwfn *p_hwfn,
-                          struct qed_spq *p_spq,
-                          struct qed_spq_entry *p_ent)
+                          struct qed_spq *p_spq, struct qed_spq_entry *p_ent)
 {
        struct qed_chain *p_chain = &p_hwfn->p_spq->chain;
        u16 echo = qed_chain_get_prod_idx(p_chain);
@@ -255,8 +252,7 @@ qed_async_event_completion(struct qed_hwfn *p_hwfn,
 /***************************************************************************
 * EQ API
 ***************************************************************************/
-void qed_eq_prod_update(struct qed_hwfn *p_hwfn,
-                       u16 prod)
+void qed_eq_prod_update(struct qed_hwfn *p_hwfn, u16 prod)
 {
        u32 addr = GTT_BAR0_MAP_REG_USDM_RAM +
                   USTORM_EQE_CONS_OFFSET(p_hwfn->rel_pf_id);
@@ -267,9 +263,7 @@ void qed_eq_prod_update(struct qed_hwfn *p_hwfn,
        mmiowb();
 }
 
-int qed_eq_completion(struct qed_hwfn *p_hwfn,
-                     void *cookie)
-
+int qed_eq_completion(struct qed_hwfn *p_hwfn, void *cookie)
 {
        struct qed_eq *p_eq = cookie;
        struct qed_chain *p_chain = &p_eq->chain;
@@ -323,8 +317,7 @@ int qed_eq_completion(struct qed_hwfn *p_hwfn,
        return rc;
 }
 
-struct qed_eq *qed_eq_alloc(struct qed_hwfn *p_hwfn,
-                           u16 num_elem)
+struct qed_eq *qed_eq_alloc(struct qed_hwfn *p_hwfn, u16 num_elem)
 {
        struct qed_eq *p_eq;
 
@@ -348,11 +341,8 @@ struct qed_eq *qed_eq_alloc(struct qed_hwfn *p_hwfn,
        }
 
        /* register EQ completion on the SP SB */
-       qed_int_register_cb(p_hwfn,
-                           qed_eq_completion,
-                           p_eq,
-                           &p_eq->eq_sb_index,
-                           &p_eq->p_fw_cons);
+       qed_int_register_cb(p_hwfn, qed_eq_completion,
+                           p_eq, &p_eq->eq_sb_index, &p_eq->p_fw_cons);
 
        return p_eq;
 
@@ -361,14 +351,12 @@ eq_allocate_fail:
        return NULL;
 }
 
-void qed_eq_setup(struct qed_hwfn *p_hwfn,
-                 struct qed_eq *p_eq)
+void qed_eq_setup(struct qed_hwfn *p_hwfn, struct qed_eq *p_eq)
 {
        qed_chain_reset(&p_eq->chain);
 }
 
-void qed_eq_free(struct qed_hwfn *p_hwfn,
-                struct qed_eq *p_eq)
+void qed_eq_free(struct qed_hwfn *p_hwfn, struct qed_eq *p_eq)
 {
        if (!p_eq)
                return;
@@ -379,10 +367,9 @@ void qed_eq_free(struct qed_hwfn *p_hwfn,
 /***************************************************************************
 * CQE API - manipulate EQ functionality
 ***************************************************************************/
-static int qed_cqe_completion(
-       struct qed_hwfn *p_hwfn,
-       struct eth_slow_path_rx_cqe *cqe,
-       enum protocol_type protocol)
+static int qed_cqe_completion(struct qed_hwfn *p_hwfn,
+                             struct eth_slow_path_rx_cqe *cqe,
+                             enum protocol_type protocol)
 {
        if (IS_VF(p_hwfn->cdev))
                return 0;
@@ -463,8 +450,7 @@ int qed_spq_alloc(struct qed_hwfn *p_hwfn)
        u32 capacity;
 
        /* SPQ struct */
-       p_spq =
-               kzalloc(sizeof(struct qed_spq), GFP_KERNEL);
+       p_spq = kzalloc(sizeof(struct qed_spq), GFP_KERNEL);
        if (!p_spq) {
                DP_NOTICE(p_hwfn, "Failed to allocate `struct qed_spq'\n");
                return -ENOMEM;
@@ -525,9 +511,7 @@ void qed_spq_free(struct qed_hwfn *p_hwfn)
        kfree(p_spq);
 }
 
-int
-qed_spq_get_entry(struct qed_hwfn *p_hwfn,
-                 struct qed_spq_entry **pp_ent)
+int qed_spq_get_entry(struct qed_hwfn *p_hwfn, struct qed_spq_entry **pp_ent)
 {
        struct qed_spq *p_spq = p_hwfn->p_spq;
        struct qed_spq_entry *p_ent = NULL;
@@ -538,14 +522,15 @@ qed_spq_get_entry(struct qed_hwfn *p_hwfn,
        if (list_empty(&p_spq->free_pool)) {
                p_ent = kzalloc(sizeof(*p_ent), GFP_ATOMIC);
                if (!p_ent) {
+                       DP_NOTICE(p_hwfn,
+                                 "Failed to allocate an SPQ entry for a pending ramrod\n");
                        rc = -ENOMEM;
                        goto out_unlock;
                }
                p_ent->queue = &p_spq->unlimited_pending;
        } else {
                p_ent = list_first_entry(&p_spq->free_pool,
-                                        struct qed_spq_entry,
-                                        list);
+                                        struct qed_spq_entry, list);
                list_del(&p_ent->list);
                p_ent->queue = &p_spq->pending;
        }
@@ -564,8 +549,7 @@ static void __qed_spq_return_entry(struct qed_hwfn *p_hwfn,
        list_add_tail(&p_ent->list, &p_hwfn->p_spq->free_pool);
 }
 
-void qed_spq_return_entry(struct qed_hwfn *p_hwfn,
-                         struct qed_spq_entry *p_ent)
+void qed_spq_return_entry(struct qed_hwfn *p_hwfn, struct qed_spq_entry *p_ent)
 {
        spin_lock_bh(&p_hwfn->p_spq->lock);
        __qed_spq_return_entry(p_hwfn, p_ent);
@@ -586,10 +570,9 @@ void qed_spq_return_entry(struct qed_hwfn *p_hwfn,
  *
  * @return int
  */
-static int
-qed_spq_add_entry(struct qed_hwfn *p_hwfn,
-                 struct qed_spq_entry *p_ent,
-                 enum spq_priority priority)
+static int qed_spq_add_entry(struct qed_hwfn *p_hwfn,
+                            struct qed_spq_entry *p_ent,
+                            enum spq_priority priority)
 {
        struct qed_spq *p_spq = p_hwfn->p_spq;
 
@@ -604,8 +587,7 @@ qed_spq_add_entry(struct qed_hwfn *p_hwfn,
                        struct qed_spq_entry *p_en2;
 
                        p_en2 = list_first_entry(&p_spq->free_pool,
-                                                struct qed_spq_entry,
-                                                list);
+                                                struct qed_spq_entry, list);
                        list_del(&p_en2->list);
 
                        /* Copy the ring element physical pointer to the new
@@ -655,8 +637,7 @@ u32 qed_spq_get_cid(struct qed_hwfn *p_hwfn)
 * Posting new Ramrods
 ***************************************************************************/
 static int qed_spq_post_list(struct qed_hwfn *p_hwfn,
-                            struct list_head *head,
-                            u32 keep_reserve)
+                            struct list_head *head, u32 keep_reserve)
 {
        struct qed_spq *p_spq = p_hwfn->p_spq;
        int rc;
@@ -690,8 +671,7 @@ static int qed_spq_pend_post(struct qed_hwfn *p_hwfn)
                        break;
 
                p_ent = list_first_entry(&p_spq->unlimited_pending,
-                                        struct qed_spq_entry,
-                                        list);
+                                        struct qed_spq_entry, list);
                if (!p_ent)
                        return -EINVAL;
 
@@ -705,8 +685,7 @@ static int qed_spq_pend_post(struct qed_hwfn *p_hwfn)
 }
 
 int qed_spq_post(struct qed_hwfn *p_hwfn,
-                struct qed_spq_entry *p_ent,
-                u8 *fw_return_code)
+                struct qed_spq_entry *p_ent, u8 *fw_return_code)
 {
        int rc = 0;
        struct qed_spq *p_spq = p_hwfn ? p_hwfn->p_spq : NULL;
@@ -803,8 +782,7 @@ int qed_spq_completion(struct qed_hwfn *p_hwfn,
                return -EINVAL;
 
        spin_lock_bh(&p_spq->lock);
-       list_for_each_entry_safe(p_ent, tmp, &p_spq->completion_pending,
-                                list) {
+       list_for_each_entry_safe(p_ent, tmp, &p_spq->completion_pending, list) {
                if (p_ent->elem.hdr.echo == echo) {
                        u16 pos = le16_to_cpu(echo) % SPQ_RING_SIZE;
 
@@ -901,14 +879,12 @@ consq_allocate_fail:
        return NULL;
 }
 
-void qed_consq_setup(struct qed_hwfn *p_hwfn,
-                    struct qed_consq *p_consq)
+void qed_consq_setup(struct qed_hwfn *p_hwfn, struct qed_consq *p_consq)
 {
        qed_chain_reset(&p_consq->chain);
 }
 
-void qed_consq_free(struct qed_hwfn *p_hwfn,
-                   struct qed_consq *p_consq)
+void qed_consq_free(struct qed_hwfn *p_hwfn, struct qed_consq *p_consq)
 {
        if (!p_consq)
                return;