| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * |
| * Authors: |
| * Alexander Aring <aar@pengutronix.de> |
| * |
| * Based on: net/mac80211/util.c |
| */ |
| |
| #include "ieee802154_i.h" |
| #include "driver-ops.h" |
| |
| /* privid for wpan_phys to determine whether they belong to us or not */ |
| const void *const mac802154_wpan_phy_privid = &mac802154_wpan_phy_privid; |
| |
| /** |
| * ieee802154_wake_queue - wake ieee802154 queue |
| * @hw: main hardware object |
| * |
| * Tranceivers usually have either one transmit framebuffer or one framebuffer |
| * for both transmitting and receiving. Hence, the core currently only handles |
| * one frame at a time for each phy, which means we had to stop the queue to |
| * avoid new skb to come during the transmission. The queue then needs to be |
| * woken up after the operation. |
| */ |
| static void ieee802154_wake_queue(struct ieee802154_hw *hw) |
| { |
| struct ieee802154_local *local = hw_to_local(hw); |
| struct ieee802154_sub_if_data *sdata; |
| |
| rcu_read_lock(); |
| clear_bit(WPAN_PHY_FLAG_STATE_QUEUE_STOPPED, &local->phy->flags); |
| list_for_each_entry_rcu(sdata, &local->interfaces, list) { |
| if (!sdata->dev) |
| continue; |
| |
| netif_wake_queue(sdata->dev); |
| } |
| rcu_read_unlock(); |
| } |
| |
| /** |
| * ieee802154_stop_queue - stop ieee802154 queue |
| * @hw: main hardware object |
| * |
| * Tranceivers usually have either one transmit framebuffer or one framebuffer |
| * for both transmitting and receiving. Hence, the core currently only handles |
| * one frame at a time for each phy, which means we need to tell upper layers to |
| * stop giving us new skbs while we are busy with the transmitted one. The queue |
| * must then be stopped before transmitting. |
| */ |
| static void ieee802154_stop_queue(struct ieee802154_hw *hw) |
| { |
| struct ieee802154_local *local = hw_to_local(hw); |
| struct ieee802154_sub_if_data *sdata; |
| |
| rcu_read_lock(); |
| list_for_each_entry_rcu(sdata, &local->interfaces, list) { |
| if (!sdata->dev) |
| continue; |
| |
| netif_stop_queue(sdata->dev); |
| } |
| rcu_read_unlock(); |
| } |
| |
| void ieee802154_hold_queue(struct ieee802154_local *local) |
| { |
| unsigned long flags; |
| |
| spin_lock_irqsave(&local->phy->queue_lock, flags); |
| if (!atomic_fetch_inc(&local->phy->hold_txs)) |
| ieee802154_stop_queue(&local->hw); |
| spin_unlock_irqrestore(&local->phy->queue_lock, flags); |
| } |
| |
| void ieee802154_release_queue(struct ieee802154_local *local) |
| { |
| unsigned long flags; |
| |
| spin_lock_irqsave(&local->phy->queue_lock, flags); |
| if (atomic_dec_and_test(&local->phy->hold_txs)) |
| ieee802154_wake_queue(&local->hw); |
| spin_unlock_irqrestore(&local->phy->queue_lock, flags); |
| } |
| |
| void ieee802154_disable_queue(struct ieee802154_local *local) |
| { |
| struct ieee802154_sub_if_data *sdata; |
| |
| rcu_read_lock(); |
| list_for_each_entry_rcu(sdata, &local->interfaces, list) { |
| if (!sdata->dev) |
| continue; |
| |
| netif_tx_disable(sdata->dev); |
| } |
| rcu_read_unlock(); |
| } |
| |
| enum hrtimer_restart ieee802154_xmit_ifs_timer(struct hrtimer *timer) |
| { |
| struct ieee802154_local *local = |
| container_of(timer, struct ieee802154_local, ifs_timer); |
| |
| ieee802154_release_queue(local); |
| |
| return HRTIMER_NORESTART; |
| } |
| |
| void ieee802154_xmit_complete(struct ieee802154_hw *hw, struct sk_buff *skb, |
| bool ifs_handling) |
| { |
| struct ieee802154_local *local = hw_to_local(hw); |
| |
| local->tx_result = IEEE802154_SUCCESS; |
| |
| if (ifs_handling) { |
| u8 max_sifs_size; |
| |
| /* If transceiver sets CRC on his own we need to use lifs |
| * threshold len above 16 otherwise 18, because it's not |
| * part of skb->len. |
| */ |
| if (hw->flags & IEEE802154_HW_TX_OMIT_CKSUM) |
| max_sifs_size = IEEE802154_MAX_SIFS_FRAME_SIZE - |
| IEEE802154_FCS_LEN; |
| else |
| max_sifs_size = IEEE802154_MAX_SIFS_FRAME_SIZE; |
| |
| if (skb->len > max_sifs_size) |
| hrtimer_start(&local->ifs_timer, |
| hw->phy->lifs_period * NSEC_PER_USEC, |
| HRTIMER_MODE_REL); |
| else |
| hrtimer_start(&local->ifs_timer, |
| hw->phy->sifs_period * NSEC_PER_USEC, |
| HRTIMER_MODE_REL); |
| } else { |
| ieee802154_release_queue(local); |
| } |
| |
| dev_consume_skb_any(skb); |
| if (atomic_dec_and_test(&hw->phy->ongoing_txs)) |
| wake_up(&hw->phy->sync_txq); |
| } |
| EXPORT_SYMBOL(ieee802154_xmit_complete); |
| |
| void ieee802154_xmit_error(struct ieee802154_hw *hw, struct sk_buff *skb, |
| int reason) |
| { |
| struct ieee802154_local *local = hw_to_local(hw); |
| |
| local->tx_result = reason; |
| ieee802154_release_queue(local); |
| dev_kfree_skb_any(skb); |
| if (atomic_dec_and_test(&hw->phy->ongoing_txs)) |
| wake_up(&hw->phy->sync_txq); |
| } |
| EXPORT_SYMBOL(ieee802154_xmit_error); |
| |
| void ieee802154_xmit_hw_error(struct ieee802154_hw *hw, struct sk_buff *skb) |
| { |
| ieee802154_xmit_error(hw, skb, IEEE802154_SYSTEM_ERROR); |
| } |
| EXPORT_SYMBOL(ieee802154_xmit_hw_error); |
| |
| void ieee802154_stop_device(struct ieee802154_local *local) |
| { |
| flush_workqueue(local->workqueue); |
| hrtimer_cancel(&local->ifs_timer); |
| drv_stop(local); |
| } |