Commit 8fe23fbd authored by Dmitry Kravkov's avatar Dmitry Kravkov Committed by David S. Miller

bnx2x: change type of spq_left to atomic

The field is now accessed from different contexts.
Signed-off-by: default avatarDmitry Kravkov <dmitry@broadcom.com>
Signed-off-by: default avatarEilon Greenstein <eilong@broadcom.com>
Signed-off-by: default avatarDavid S. Miller <davem@davemloft.net>
parent fb3bff17
......@@ -860,7 +860,7 @@ struct bnx2x {
struct eth_spe *spq_prod_bd;
struct eth_spe *spq_last_bd;
__le16 *dsb_sp_prod;
u16 spq_left; /* serialize spq */
atomic_t spq_left; /* serialize spq */
/* used to synchronize spq accesses */
spinlock_t spq_lock;
......
......@@ -1161,8 +1161,8 @@ void bnx2x_sp_event(struct bnx2x_fastpath *fp,
break;
}
bp->spq_left++;
smp_mb__before_atomic_inc();
atomic_inc(&bp->spq_left);
/* push the change in fp->state and towards the memory */
smp_wmb();
......@@ -2432,7 +2432,7 @@ int bnx2x_sp_post(struct bnx2x *bp, int command, int cid,
spin_lock_bh(&bp->spq_lock);
if (!bp->spq_left) {
if (!atomic_read(&bp->spq_left)) {
BNX2X_ERR("BUG! SPQ ring full!\n");
spin_unlock_bh(&bp->spq_lock);
bnx2x_panic();
......@@ -2472,7 +2472,7 @@ int bnx2x_sp_post(struct bnx2x *bp, int command, int cid,
* somewhere between the spin_lock and spin_unlock. Thus no
* more explict memory barrier is needed.
*/
bp->spq_left--;
atomic_dec(&bp->spq_left);
DP(BNX2X_MSG_SP/*NETIF_MSG_TIMER*/,
"SPQE[%x] (%x:%x) command %d hw_cid %x data (%x:%x) "
......@@ -2480,7 +2480,7 @@ int bnx2x_sp_post(struct bnx2x *bp, int command, int cid,
bp->spq_prod_idx, (u32)U64_HI(bp->spq_mapping),
(u32)(U64_LO(bp->spq_mapping) +
(void *)bp->spq_prod_bd - (void *)bp->spq), command,
HW_CID(bp, cid), data_hi, data_lo, type, bp->spq_left);
HW_CID(bp, cid), data_hi, data_lo, type, atomic_read(&bp->spq_left));
bnx2x_sp_prod_update(bp);
spin_unlock_bh(&bp->spq_lock);
......@@ -3290,7 +3290,7 @@ static void bnx2x_eq_int(struct bnx2x *bp)
sw_prod = bp->eq_prod;
DP(BNX2X_MSG_SP, "EQ: hw_cons %u sw_cons %u bp->spq_left %u\n",
hw_cons, sw_cons, bp->spq_left);
hw_cons, sw_cons, atomic_read(&bp->spq_left));
for (; sw_cons != hw_cons;
sw_prod = NEXT_EQ_IDX(sw_prod), sw_cons = NEXT_EQ_IDX(sw_cons)) {
......@@ -3360,7 +3360,8 @@ static void bnx2x_eq_int(struct bnx2x *bp)
spqe_cnt++;
} /* for */
bp->spq_left++;
smp_mb__before_atomic_inc();
atomic_add(spqe_cnt, &bp->spq_left);
bp->eq_cons = sw_cons;
bp->eq_prod = sw_prod;
......@@ -3737,8 +3738,8 @@ void bnx2x_update_coalesce(struct bnx2x *bp)
static void bnx2x_init_sp_ring(struct bnx2x *bp)
{
spin_lock_init(&bp->spq_lock);
atomic_set(&bp->spq_left, MAX_SPQ_PENDING);
bp->spq_left = MAX_SPQ_PENDING;
bp->spq_prod_idx = 0;
bp->dsb_sp_prod = BNX2X_SP_DSB_INDEX;
bp->spq_prod_bd = bp->spq;
......
......@@ -166,11 +166,8 @@ static void bnx2x_storm_stats_post(struct bnx2x *bp)
rc = bnx2x_sp_post(bp, RAMROD_CMD_ID_COMMON_STAT_QUERY, 0,
((u32 *)&ramrod_data)[1],
((u32 *)&ramrod_data)[0], 1);
if (rc == 0) {
/* stats ramrod has it's own slot on the spq */
bp->spq_left++;
if (rc == 0)
bp->stats_pending = 1;
}
spin_unlock_bh(&bp->stats_lock);
}
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment