About two thirds of the fields in 'struct qca_data' are related with in band sleep (IBS). Move these fields to a dedicated IBS struct which is part of 'struct qca_data'. Also congregate the IBS stats fields for debugfs in a separate sub-struct. Signed-off-by: Matthias Kaehlcke <mka@xxxxxxxxxxxx> --- drivers/bluetooth/hci_qca.c | 362 +++++++++++++++++++----------------- 1 file changed, 191 insertions(+), 171 deletions(-) diff --git a/drivers/bluetooth/hci_qca.c b/drivers/bluetooth/hci_qca.c index 3abf4e68de7a8c..c578c7c92680a6 100644 --- a/drivers/bluetooth/hci_qca.c +++ b/drivers/bluetooth/hci_qca.c @@ -125,51 +125,60 @@ struct qca_dump_size { u32 dump_size; } __packed; -struct qca_data { - struct hci_uart *hu; - struct sk_buff *rx_skb; - struct sk_buff_head txq; +struct qca_ibs_stats { + /* For debugging purpose */ + u64 sent_wacks; + u64 sent_slps; + u64 sent_wakes; + u64 recv_wacks; + u64 recv_slps; + u64 recv_wakes; + u64 vote_last_jif; + u32 vote_on_ms; + u32 vote_off_ms; + u64 tx_votes_on; + u64 rx_votes_on; + u64 tx_votes_off; + u64 rx_votes_off; + u64 votes_on; + u64 votes_off; +}; + +struct qca_ibs { struct sk_buff_head tx_wait_q; /* HCI_IBS wait queue */ - struct sk_buff_head rx_memdump_q; /* Memdump wait queue */ - spinlock_t hci_ibs_lock; /* HCI_IBS state lock */ - u8 tx_ibs_state; /* HCI_IBS transmit side power state*/ - u8 rx_ibs_state; /* HCI_IBS receive side power state */ + spinlock_t lock; /* HCI_IBS state lock */ + u8 tx_state; /* HCI_IBS transmit side power state*/ + u8 rx_state; /* HCI_IBS receive side power state */ bool tx_vote; /* Clock must be on for TX */ bool rx_vote; /* Clock must be on for RX */ struct timer_list tx_idle_timer; u32 tx_idle_delay; struct timer_list wake_retrans_timer; u32 wake_retrans; - struct workqueue_struct *workqueue; + struct work_struct ws_awake_rx; struct work_struct ws_awake_device; struct work_struct ws_rx_vote_off; struct work_struct ws_tx_vote_off; + + struct qca_ibs_stats stats; +}; + +struct qca_data { + struct hci_uart *hu; + struct sk_buff *rx_skb; + struct sk_buff_head txq; + struct sk_buff_head rx_memdump_q; /* Memdump wait queue */ + struct workqueue_struct *workqueue; struct work_struct ctrl_memdump_evt; struct delayed_work ctrl_memdump_timeout; struct qca_memdump_data *qca_memdump; + struct qca_ibs ibs; unsigned long flags; struct completion drop_ev_comp; wait_queue_head_t suspend_wait_q; enum qca_memdump_states memdump_state; struct mutex hci_memdump_lock; - - /* For debugging purpose */ - u64 ibs_sent_wacks; - u64 ibs_sent_slps; - u64 ibs_sent_wakes; - u64 ibs_recv_wacks; - u64 ibs_recv_slps; - u64 ibs_recv_wakes; - u64 vote_last_jif; - u32 vote_on_ms; - u32 vote_off_ms; - u64 tx_votes_on; - u64 rx_votes_on; - u64 tx_votes_off; - u64 rx_votes_off; - u64 votes_on; - u64 votes_off; }; enum qca_speed_type { @@ -267,41 +276,41 @@ static void serial_clock_vote(unsigned long vote, struct hci_uart *hu) struct qca_data *qca = hu->priv; unsigned int diff; - bool old_vote = (qca->tx_vote | qca->rx_vote); + bool old_vote = (qca->ibs.tx_vote | qca->ibs.rx_vote); bool new_vote; switch (vote) { case HCI_IBS_VOTE_STATS_UPDATE: - diff = jiffies_to_msecs(jiffies - qca->vote_last_jif); + diff = jiffies_to_msecs(jiffies - qca->ibs.stats.vote_last_jif); if (old_vote) - qca->vote_off_ms += diff; + qca->ibs.stats.vote_off_ms += diff; else - qca->vote_on_ms += diff; + qca->ibs.stats.vote_on_ms += diff; return; case HCI_IBS_TX_VOTE_CLOCK_ON: - qca->tx_vote = true; - qca->tx_votes_on++; + qca->ibs.tx_vote = true; + qca->ibs.stats.tx_votes_on++; new_vote = true; break; case HCI_IBS_RX_VOTE_CLOCK_ON: - qca->rx_vote = true; - qca->rx_votes_on++; + qca->ibs.rx_vote = true; + qca->ibs.stats.rx_votes_on++; new_vote = true; break; case HCI_IBS_TX_VOTE_CLOCK_OFF: - qca->tx_vote = false; - qca->tx_votes_off++; - new_vote = qca->rx_vote | qca->tx_vote; + qca->ibs.tx_vote = false; + qca->ibs.stats.tx_votes_off++; + new_vote = qca->ibs.rx_vote | qca->ibs.tx_vote; break; case HCI_IBS_RX_VOTE_CLOCK_OFF: - qca->rx_vote = false; - qca->rx_votes_off++; - new_vote = qca->rx_vote | qca->tx_vote; + qca->ibs.rx_vote = false; + qca->ibs.stats.rx_votes_off++; + new_vote = qca->ibs.rx_vote | qca->ibs.tx_vote; break; default: @@ -318,16 +327,16 @@ static void serial_clock_vote(unsigned long vote, struct hci_uart *hu) BT_DBG("Vote serial clock %s(%s)", new_vote ? "true" : "false", vote ? "true" : "false"); - diff = jiffies_to_msecs(jiffies - qca->vote_last_jif); + diff = jiffies_to_msecs(jiffies - qca->ibs.stats.vote_last_jif); if (new_vote) { - qca->votes_on++; - qca->vote_off_ms += diff; + qca->ibs.stats.votes_on++; + qca->ibs.stats.vote_off_ms += diff; } else { - qca->votes_off++; - qca->vote_on_ms += diff; + qca->ibs.stats.votes_off++; + qca->ibs.stats.vote_on_ms += diff; } - qca->vote_last_jif = jiffies; + qca->ibs.stats.vote_last_jif = jiffies; } } @@ -359,7 +368,7 @@ static int send_hci_ibs_cmd(u8 cmd, struct hci_uart *hu) static void qca_wq_awake_device(struct work_struct *work) { struct qca_data *qca = container_of(work, struct qca_data, - ws_awake_device); + ibs.ws_awake_device); struct hci_uart *hu = qca->hu; unsigned long retrans_delay; unsigned long flags; @@ -369,19 +378,19 @@ static void qca_wq_awake_device(struct work_struct *work) /* Vote for serial clock */ serial_clock_vote(HCI_IBS_TX_VOTE_CLOCK_ON, hu); - spin_lock_irqsave(&qca->hci_ibs_lock, flags); + spin_lock_irqsave(&qca->ibs.lock, flags); /* Send wake indication to device */ if (send_hci_ibs_cmd(HCI_IBS_WAKE_IND, hu) < 0) BT_ERR("Failed to send WAKE to device"); - qca->ibs_sent_wakes++; + qca->ibs.stats.sent_wakes++; /* Start retransmit timer */ - retrans_delay = msecs_to_jiffies(qca->wake_retrans); - mod_timer(&qca->wake_retrans_timer, jiffies + retrans_delay); + retrans_delay = msecs_to_jiffies(qca->ibs.wake_retrans); + mod_timer(&qca->ibs.wake_retrans_timer, jiffies + retrans_delay); - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); /* Actually send the packets */ hci_uart_tx_wakeup(hu); @@ -390,7 +399,7 @@ static void qca_wq_awake_device(struct work_struct *work) static void qca_wq_awake_rx(struct work_struct *work) { struct qca_data *qca = container_of(work, struct qca_data, - ws_awake_rx); + ibs.ws_awake_rx); struct hci_uart *hu = qca->hu; unsigned long flags; @@ -398,8 +407,8 @@ static void qca_wq_awake_rx(struct work_struct *work) serial_clock_vote(HCI_IBS_RX_VOTE_CLOCK_ON, hu); - spin_lock_irqsave(&qca->hci_ibs_lock, flags); - qca->rx_ibs_state = HCI_IBS_RX_AWAKE; + spin_lock_irqsave(&qca->ibs.lock, flags); + qca->ibs.rx_state = HCI_IBS_RX_AWAKE; /* Always acknowledge device wake up, * sending IBS message doesn't count as TX ON. @@ -407,9 +416,9 @@ static void qca_wq_awake_rx(struct work_struct *work) if (send_hci_ibs_cmd(HCI_IBS_WAKE_ACK, hu) < 0) BT_ERR("Failed to acknowledge device wake up"); - qca->ibs_sent_wacks++; + qca->ibs.stats.sent_wacks++; - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); /* Actually send the packets */ hci_uart_tx_wakeup(hu); @@ -418,7 +427,7 @@ static void qca_wq_awake_rx(struct work_struct *work) static void qca_wq_serial_rx_clock_vote_off(struct work_struct *work) { struct qca_data *qca = container_of(work, struct qca_data, - ws_rx_vote_off); + ibs.ws_rx_vote_off); struct hci_uart *hu = qca->hu; BT_DBG("hu %p rx clock vote off", hu); @@ -429,7 +438,7 @@ static void qca_wq_serial_rx_clock_vote_off(struct work_struct *work) static void qca_wq_serial_tx_clock_vote_off(struct work_struct *work) { struct qca_data *qca = container_of(work, struct qca_data, - ws_tx_vote_off); + ibs.ws_tx_vote_off); struct hci_uart *hu = qca->hu; BT_DBG("hu %p tx clock vote off", hu); @@ -445,25 +454,25 @@ static void qca_wq_serial_tx_clock_vote_off(struct work_struct *work) static void hci_ibs_tx_idle_timeout(struct timer_list *t) { - struct qca_data *qca = from_timer(qca, t, tx_idle_timer); + struct qca_data *qca = from_timer(qca, t, ibs.tx_idle_timer); struct hci_uart *hu = qca->hu; unsigned long flags; - BT_DBG("hu %p idle timeout in %d state", hu, qca->tx_ibs_state); + BT_DBG("hu %p idle timeout in %d state", hu, qca->ibs.tx_state); - spin_lock_irqsave_nested(&qca->hci_ibs_lock, + spin_lock_irqsave_nested(&qca->ibs.lock, flags, SINGLE_DEPTH_NESTING); - switch (qca->tx_ibs_state) { + switch (qca->ibs.tx_state) { case HCI_IBS_TX_AWAKE: /* TX_IDLE, go to SLEEP */ if (send_hci_ibs_cmd(HCI_IBS_SLEEP_IND, hu) < 0) { BT_ERR("Failed to send SLEEP to device"); break; } - qca->tx_ibs_state = HCI_IBS_TX_ASLEEP; - qca->ibs_sent_slps++; - queue_work(qca->workqueue, &qca->ws_tx_vote_off); + qca->ibs.tx_state = HCI_IBS_TX_ASLEEP; + qca->ibs.stats.sent_slps++; + queue_work(qca->workqueue, &qca->ibs.ws_tx_vote_off); break; case HCI_IBS_TX_ASLEEP: @@ -471,33 +480,33 @@ static void hci_ibs_tx_idle_timeout(struct timer_list *t) /* Fall through */ default: - BT_ERR("Spurious timeout tx state %d", qca->tx_ibs_state); + BT_ERR("Spurious timeout tx state %d", qca->ibs.tx_state); break; } - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); } static void hci_ibs_wake_retrans_timeout(struct timer_list *t) { - struct qca_data *qca = from_timer(qca, t, wake_retrans_timer); + struct qca_data *qca = from_timer(qca, t, ibs.wake_retrans_timer); struct hci_uart *hu = qca->hu; unsigned long flags, retrans_delay; bool retransmit = false; BT_DBG("hu %p wake retransmit timeout in %d state", - hu, qca->tx_ibs_state); + hu, qca->ibs.tx_state); - spin_lock_irqsave_nested(&qca->hci_ibs_lock, + spin_lock_irqsave_nested(&qca->ibs.lock, flags, SINGLE_DEPTH_NESTING); /* Don't retransmit the HCI_IBS_WAKE_IND when suspending. */ if (test_bit(QCA_SUSPENDING, &qca->flags)) { - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); return; } - switch (qca->tx_ibs_state) { + switch (qca->ibs.tx_state) { case HCI_IBS_TX_WAKING: /* No WAKE_ACK, retransmit WAKE */ retransmit = true; @@ -505,9 +514,10 @@ static void hci_ibs_wake_retrans_timeout(struct timer_list *t) BT_ERR("Failed to acknowledge device wake up"); break; } - qca->ibs_sent_wakes++; - retrans_delay = msecs_to_jiffies(qca->wake_retrans); - mod_timer(&qca->wake_retrans_timer, jiffies + retrans_delay); + qca->ibs.stats.sent_wakes++; + retrans_delay = msecs_to_jiffies(qca->ibs.wake_retrans); + mod_timer(&qca->ibs.wake_retrans_timer, + jiffies + retrans_delay); break; case HCI_IBS_TX_ASLEEP: @@ -515,11 +525,11 @@ static void hci_ibs_wake_retrans_timeout(struct timer_list *t) /* Fall through */ default: - BT_ERR("Spurious timeout tx state %d", qca->tx_ibs_state); + BT_ERR("Spurious timeout tx state %d", qca->ibs.tx_state); break; } - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); if (retransmit) hci_uart_tx_wakeup(hu); @@ -563,9 +573,9 @@ static int qca_open(struct hci_uart *hu) return -ENOMEM; skb_queue_head_init(&qca->txq); - skb_queue_head_init(&qca->tx_wait_q); + skb_queue_head_init(&qca->ibs.tx_wait_q); skb_queue_head_init(&qca->rx_memdump_q); - spin_lock_init(&qca->hci_ibs_lock); + spin_lock_init(&qca->ibs.lock); mutex_init(&qca->hci_memdump_lock); qca->workqueue = alloc_ordered_workqueue("qca_wq", 0); if (!qca->workqueue) { @@ -574,10 +584,10 @@ static int qca_open(struct hci_uart *hu) return -ENOMEM; } - INIT_WORK(&qca->ws_awake_rx, qca_wq_awake_rx); - INIT_WORK(&qca->ws_awake_device, qca_wq_awake_device); - INIT_WORK(&qca->ws_rx_vote_off, qca_wq_serial_rx_clock_vote_off); - INIT_WORK(&qca->ws_tx_vote_off, qca_wq_serial_tx_clock_vote_off); + INIT_WORK(&qca->ibs.ws_awake_rx, qca_wq_awake_rx); + INIT_WORK(&qca->ibs.ws_awake_device, qca_wq_awake_device); + INIT_WORK(&qca->ibs.ws_rx_vote_off, qca_wq_serial_rx_clock_vote_off); + INIT_WORK(&qca->ibs.ws_tx_vote_off, qca_wq_serial_tx_clock_vote_off); INIT_WORK(&qca->ctrl_memdump_evt, qca_controller_memdump); INIT_DELAYED_WORK(&qca->ctrl_memdump_timeout, qca_controller_memdump_timeout); @@ -587,10 +597,10 @@ static int qca_open(struct hci_uart *hu) init_completion(&qca->drop_ev_comp); /* Assume we start with both sides asleep -- extra wakes OK */ - qca->tx_ibs_state = HCI_IBS_TX_ASLEEP; - qca->rx_ibs_state = HCI_IBS_RX_ASLEEP; + qca->ibs.tx_state = HCI_IBS_TX_ASLEEP; + qca->ibs.rx_state = HCI_IBS_RX_ASLEEP; - qca->vote_last_jif = jiffies; + qca->ibs.stats.vote_last_jif = jiffies; hu->priv = qca; @@ -602,14 +612,15 @@ static int qca_open(struct hci_uart *hu) } } - timer_setup(&qca->wake_retrans_timer, hci_ibs_wake_retrans_timeout, 0); - qca->wake_retrans = IBS_WAKE_RETRANS_TIMEOUT_MS; + timer_setup(&qca->ibs.wake_retrans_timer, hci_ibs_wake_retrans_timeout, + 0); + qca->ibs.wake_retrans = IBS_WAKE_RETRANS_TIMEOUT_MS; - timer_setup(&qca->tx_idle_timer, hci_ibs_tx_idle_timeout, 0); - qca->tx_idle_delay = IBS_HOST_TX_IDLE_TIMEOUT_MS; + timer_setup(&qca->ibs.tx_idle_timer, hci_ibs_tx_idle_timeout, 0); + qca->ibs.tx_idle_delay = IBS_HOST_TX_IDLE_TIMEOUT_MS; BT_DBG("HCI_UART_QCA open, tx_idle_delay=%u, wake_retrans=%u", - qca->tx_idle_delay, qca->wake_retrans); + qca->ibs.tx_idle_delay, qca->ibs.wake_retrans); return 0; } @@ -628,36 +639,45 @@ static void qca_debugfs_init(struct hci_dev *hdev) /* read only */ mode = S_IRUGO; - debugfs_create_u8("tx_ibs_state", mode, ibs_dir, &qca->tx_ibs_state); - debugfs_create_u8("rx_ibs_state", mode, ibs_dir, &qca->rx_ibs_state); + debugfs_create_u8("tx_ibs_state", mode, ibs_dir, &qca->ibs.tx_state); + debugfs_create_u8("rx_ibs_state", mode, ibs_dir, &qca->ibs.rx_state); debugfs_create_u64("ibs_sent_sleeps", mode, ibs_dir, - &qca->ibs_sent_slps); + &qca->ibs.stats.sent_slps); debugfs_create_u64("ibs_sent_wakes", mode, ibs_dir, - &qca->ibs_sent_wakes); + &qca->ibs.stats.sent_wakes); debugfs_create_u64("ibs_sent_wake_acks", mode, ibs_dir, - &qca->ibs_sent_wacks); + &qca->ibs.stats.sent_wacks); debugfs_create_u64("ibs_recv_sleeps", mode, ibs_dir, - &qca->ibs_recv_slps); + &qca->ibs.stats.recv_slps); debugfs_create_u64("ibs_recv_wakes", mode, ibs_dir, - &qca->ibs_recv_wakes); + &qca->ibs.stats.recv_wakes); debugfs_create_u64("ibs_recv_wake_acks", mode, ibs_dir, - &qca->ibs_recv_wacks); - debugfs_create_bool("tx_vote", mode, ibs_dir, &qca->tx_vote); - debugfs_create_u64("tx_votes_on", mode, ibs_dir, &qca->tx_votes_on); - debugfs_create_u64("tx_votes_off", mode, ibs_dir, &qca->tx_votes_off); - debugfs_create_bool("rx_vote", mode, ibs_dir, &qca->rx_vote); - debugfs_create_u64("rx_votes_on", mode, ibs_dir, &qca->rx_votes_on); - debugfs_create_u64("rx_votes_off", mode, ibs_dir, &qca->rx_votes_off); - debugfs_create_u64("votes_on", mode, ibs_dir, &qca->votes_on); - debugfs_create_u64("votes_off", mode, ibs_dir, &qca->votes_off); - debugfs_create_u32("vote_on_ms", mode, ibs_dir, &qca->vote_on_ms); - debugfs_create_u32("vote_off_ms", mode, ibs_dir, &qca->vote_off_ms); + &qca->ibs.stats.recv_wacks); + debugfs_create_bool("tx_vote", mode, ibs_dir, &qca->ibs.tx_vote); + debugfs_create_u64("tx_votes_on", mode, ibs_dir, + &qca->ibs.stats.tx_votes_on); + debugfs_create_u64("tx_votes_off", mode, ibs_dir, + &qca->ibs.stats.tx_votes_off); + debugfs_create_bool("rx_vote", mode, ibs_dir, &qca->ibs.rx_vote); + debugfs_create_u64("rx_votes_on", mode, ibs_dir, + &qca->ibs.stats.rx_votes_on); + debugfs_create_u64("rx_votes_off", mode, ibs_dir, + &qca->ibs.stats.rx_votes_off); + debugfs_create_u64("votes_on", mode, ibs_dir, + &qca->ibs.stats.votes_on); + debugfs_create_u64("votes_off", mode, ibs_dir, + &qca->ibs.stats.votes_off); + debugfs_create_u32("vote_on_ms", mode, ibs_dir, + &qca->ibs.stats.vote_on_ms); + debugfs_create_u32("vote_off_ms", mode, ibs_dir, + &qca->ibs.stats.vote_off_ms); /* read/write */ mode = S_IRUGO | S_IWUSR; - debugfs_create_u32("wake_retrans", mode, ibs_dir, &qca->wake_retrans); + debugfs_create_u32("wake_retrans", mode, ibs_dir, + &qca->ibs.wake_retrans); debugfs_create_u32("tx_idle_delay", mode, ibs_dir, - &qca->tx_idle_delay); + &qca->ibs.tx_idle_delay); } /* Flush protocol data */ @@ -667,7 +687,7 @@ static int qca_flush(struct hci_uart *hu) BT_DBG("hu %p qca flush", hu); - skb_queue_purge(&qca->tx_wait_q); + skb_queue_purge(&qca->ibs.tx_wait_q); skb_queue_purge(&qca->txq); return 0; @@ -682,11 +702,11 @@ static int qca_close(struct hci_uart *hu) serial_clock_vote(HCI_IBS_VOTE_STATS_UPDATE, hu); - skb_queue_purge(&qca->tx_wait_q); + skb_queue_purge(&qca->ibs.tx_wait_q); skb_queue_purge(&qca->txq); skb_queue_purge(&qca->rx_memdump_q); - del_timer(&qca->tx_idle_timer); - del_timer(&qca->wake_retrans_timer); + del_timer(&qca->ibs.tx_idle_timer); + del_timer(&qca->ibs.wake_retrans_timer); destroy_workqueue(qca->workqueue); qca->hu = NULL; @@ -710,23 +730,23 @@ static void device_want_to_wakeup(struct hci_uart *hu) BT_DBG("hu %p want to wake up", hu); - spin_lock_irqsave(&qca->hci_ibs_lock, flags); + spin_lock_irqsave(&qca->ibs.lock, flags); - qca->ibs_recv_wakes++; + qca->ibs.stats.recv_wakes++; /* Don't wake the rx up when suspending. */ if (test_bit(QCA_SUSPENDING, &qca->flags)) { - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); return; } - switch (qca->rx_ibs_state) { + switch (qca->ibs.rx_state) { case HCI_IBS_RX_ASLEEP: /* Make sure clock is on - we may have turned clock off since * receiving the wake up indicator awake rx clock. */ - queue_work(qca->workqueue, &qca->ws_awake_rx); - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + queue_work(qca->workqueue, &qca->ibs.ws_awake_rx); + spin_unlock_irqrestore(&qca->ibs.lock, flags); return; case HCI_IBS_RX_AWAKE: @@ -737,17 +757,17 @@ static void device_want_to_wakeup(struct hci_uart *hu) BT_ERR("Failed to acknowledge device wake up"); break; } - qca->ibs_sent_wacks++; + qca->ibs.stats.sent_wacks++; break; default: /* Any other state is illegal */ BT_ERR("Received HCI_IBS_WAKE_IND in rx state %d", - qca->rx_ibs_state); + qca->ibs.rx_state); break; } - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); /* Actually send the packets */ hci_uart_tx_wakeup(hu); @@ -760,18 +780,18 @@ static void device_want_to_sleep(struct hci_uart *hu) unsigned long flags; struct qca_data *qca = hu->priv; - BT_DBG("hu %p want to sleep in %d state", hu, qca->rx_ibs_state); + BT_DBG("hu %p want to sleep in %d state", hu, qca->ibs.rx_state); - spin_lock_irqsave(&qca->hci_ibs_lock, flags); + spin_lock_irqsave(&qca->ibs.lock, flags); - qca->ibs_recv_slps++; + qca->ibs.stats.recv_slps++; - switch (qca->rx_ibs_state) { + switch (qca->ibs.rx_state) { case HCI_IBS_RX_AWAKE: /* Update state */ - qca->rx_ibs_state = HCI_IBS_RX_ASLEEP; + qca->ibs.rx_state = HCI_IBS_RX_ASLEEP; /* Vote off rx clock under workqueue */ - queue_work(qca->workqueue, &qca->ws_rx_vote_off); + queue_work(qca->workqueue, &qca->ibs.ws_rx_vote_off); break; case HCI_IBS_RX_ASLEEP: @@ -780,13 +800,13 @@ static void device_want_to_sleep(struct hci_uart *hu) default: /* Any other state is illegal */ BT_ERR("Received HCI_IBS_SLEEP_IND in rx state %d", - qca->rx_ibs_state); + qca->ibs.rx_state); break; } wake_up_interruptible(&qca->suspend_wait_q); - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); } /* Called upon wake-up-acknowledgement from the device @@ -799,33 +819,33 @@ static void device_woke_up(struct hci_uart *hu) BT_DBG("hu %p woke up", hu); - spin_lock_irqsave(&qca->hci_ibs_lock, flags); + spin_lock_irqsave(&qca->ibs.lock, flags); - qca->ibs_recv_wacks++; + qca->ibs.stats.recv_wacks++; /* Don't react to the wake-up-acknowledgment when suspending. */ if (test_bit(QCA_SUSPENDING, &qca->flags)) { - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); return; } - switch (qca->tx_ibs_state) { + switch (qca->ibs.tx_state) { case HCI_IBS_TX_AWAKE: /* Expect one if we send 2 WAKEs */ BT_DBG("Received HCI_IBS_WAKE_ACK in tx state %d", - qca->tx_ibs_state); + qca->ibs.tx_state); break; case HCI_IBS_TX_WAKING: /* Send pending packets */ - while ((skb = skb_dequeue(&qca->tx_wait_q))) + while ((skb = skb_dequeue(&qca->ibs.tx_wait_q))) skb_queue_tail(&qca->txq, skb); /* Switch timers and change state to HCI_IBS_TX_AWAKE */ - del_timer(&qca->wake_retrans_timer); - idle_delay = msecs_to_jiffies(qca->tx_idle_delay); - mod_timer(&qca->tx_idle_timer, jiffies + idle_delay); - qca->tx_ibs_state = HCI_IBS_TX_AWAKE; + del_timer(&qca->ibs.wake_retrans_timer); + idle_delay = msecs_to_jiffies(qca->ibs.tx_idle_delay); + mod_timer(&qca->ibs.tx_idle_timer, jiffies + idle_delay); + qca->ibs.tx_state = HCI_IBS_TX_AWAKE; break; case HCI_IBS_TX_ASLEEP: @@ -833,11 +853,11 @@ static void device_woke_up(struct hci_uart *hu) default: BT_ERR("Received HCI_IBS_WAKE_ACK in tx state %d", - qca->tx_ibs_state); + qca->ibs.tx_state); break; } - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); /* Actually send the packets */ hci_uart_tx_wakeup(hu); @@ -852,12 +872,12 @@ static int qca_enqueue(struct hci_uart *hu, struct sk_buff *skb) struct qca_data *qca = hu->priv; BT_DBG("hu %p qca enq skb %p tx_ibs_state %d", hu, skb, - qca->tx_ibs_state); + qca->ibs.tx_state); /* Prepend skb with frame type */ memcpy(skb_push(skb, 1), &hci_skb_pkt_type(skb), 1); - spin_lock_irqsave(&qca->hci_ibs_lock, flags); + spin_lock_irqsave(&qca->ibs.lock, flags); /* Don't go to sleep in middle of patch download or * Out-Of-Band(GPIOs control) sleep is selected. @@ -866,43 +886,43 @@ static int qca_enqueue(struct hci_uart *hu, struct sk_buff *skb) if (!test_bit(QCA_IBS_ENABLED, &qca->flags) || test_bit(QCA_SUSPENDING, &qca->flags)) { skb_queue_tail(&qca->txq, skb); - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); return 0; } /* Act according to current state */ - switch (qca->tx_ibs_state) { + switch (qca->ibs.tx_state) { case HCI_IBS_TX_AWAKE: BT_DBG("Device awake, sending normally"); skb_queue_tail(&qca->txq, skb); - idle_delay = msecs_to_jiffies(qca->tx_idle_delay); - mod_timer(&qca->tx_idle_timer, jiffies + idle_delay); + idle_delay = msecs_to_jiffies(qca->ibs.tx_idle_delay); + mod_timer(&qca->ibs.tx_idle_timer, jiffies + idle_delay); break; case HCI_IBS_TX_ASLEEP: BT_DBG("Device asleep, waking up and queueing packet"); /* Save packet for later */ - skb_queue_tail(&qca->tx_wait_q, skb); + skb_queue_tail(&qca->ibs.tx_wait_q, skb); - qca->tx_ibs_state = HCI_IBS_TX_WAKING; + qca->ibs.tx_state = HCI_IBS_TX_WAKING; /* Schedule a work queue to wake up device */ - queue_work(qca->workqueue, &qca->ws_awake_device); + queue_work(qca->workqueue, &qca->ibs.ws_awake_device); break; case HCI_IBS_TX_WAKING: BT_DBG("Device waking up, queueing packet"); /* Transient state; just keep packet for later */ - skb_queue_tail(&qca->tx_wait_q, skb); + skb_queue_tail(&qca->ibs.tx_wait_q, skb); break; default: BT_ERR("Illegal tx state: %d (losing packet)", - qca->tx_ibs_state); + qca->ibs.tx_state); kfree_skb(skb); break; } - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); return 0; } @@ -1732,10 +1752,10 @@ static void qca_power_shutdown(struct hci_uart *hu) * still open, stop queueing the IBS data and flush all the buffered * data in skb's. */ - spin_lock_irqsave(&qca->hci_ibs_lock, flags); + spin_lock_irqsave(&qca->ibs.lock, flags); clear_bit(QCA_IBS_ENABLED, &qca->flags); qca_flush(hu); - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); hu->hdev->hw_error = NULL; hu->hdev->cmd_timeout = NULL; @@ -1959,18 +1979,18 @@ static int __maybe_unused qca_suspend(struct device *dev) if (!test_bit(QCA_IBS_ENABLED, &qca->flags)) return 0; - cancel_work_sync(&qca->ws_awake_device); - cancel_work_sync(&qca->ws_awake_rx); + cancel_work_sync(&qca->ibs.ws_awake_device); + cancel_work_sync(&qca->ibs.ws_awake_rx); - spin_lock_irqsave_nested(&qca->hci_ibs_lock, + spin_lock_irqsave_nested(&qca->ibs.lock, flags, SINGLE_DEPTH_NESTING); - switch (qca->tx_ibs_state) { + switch (qca->ibs.tx_state) { case HCI_IBS_TX_WAKING: - del_timer(&qca->wake_retrans_timer); + del_timer(&qca->ibs.wake_retrans_timer); /* Fall through */ case HCI_IBS_TX_AWAKE: - del_timer(&qca->tx_idle_timer); + del_timer(&qca->ibs.tx_idle_timer); serdev_device_write_flush(hu->serdev); cmd = HCI_IBS_SLEEP_IND; @@ -1981,22 +2001,22 @@ static int __maybe_unused qca_suspend(struct device *dev) break; } - qca->tx_ibs_state = HCI_IBS_TX_ASLEEP; - qca->ibs_sent_slps++; + qca->ibs.tx_state = HCI_IBS_TX_ASLEEP; + qca->ibs.stats.sent_slps++; - qca_wq_serial_tx_clock_vote_off(&qca->ws_tx_vote_off); + qca_wq_serial_tx_clock_vote_off(&qca->ibs.ws_tx_vote_off); break; case HCI_IBS_TX_ASLEEP: break; default: - BT_ERR("Spurious tx state %d", qca->tx_ibs_state); + BT_ERR("Spurious tx state %d", qca->ibs.tx_state); ret = -EINVAL; break; } - spin_unlock_irqrestore(&qca->hci_ibs_lock, flags); + spin_unlock_irqrestore(&qca->ibs.lock, flags); if (ret < 0) goto error; @@ -2009,7 +2029,7 @@ static int __maybe_unused qca_suspend(struct device *dev) */ ret = wait_event_interruptible_timeout(qca->suspend_wait_q, - qca->rx_ibs_state == HCI_IBS_RX_ASLEEP, + qca->ibs.rx_state == HCI_IBS_RX_ASLEEP, msecs_to_jiffies(IBS_BTSOC_TX_IDLE_TIMEOUT_MS)); if (ret > 0) -- 2.25.1.481.gfbce0eb801-goog