Networking fixes for 5.18-rc7, including fixes from wireless,
and bluetooth. No outstanding fires. Current release - regressions: - eth: atlantic: always deep reset on pm op, fix null-deref Current release - new code bugs: - rds: use maybe_get_net() when acquiring refcount on TCP sockets [refinement of a previous fix] - eth: ocelot: mark traps with a bool instead of guessing type based on list membership Previous releases - regressions: - net: fix skipping features in for_each_netdev_feature() - phy: micrel: fix null-derefs on suspend/resume and probe - bcmgenet: check for Wake-on-LAN interrupt probe deferral Previous releases - always broken: - ipv4: drop dst in multicast routing path, prevent leaks - ping: fix address binding wrt vrf - net: fix wrong network header length when BPF protocol translation is used on skbs with a fraglist - bluetooth: fix the creation of hdev->name - rfkill: uapi: fix RFKILL_IOCTL_MAX_SIZE ioctl request definition - wifi: iwlwifi: iwl-dbg: use del_timer_sync() before freeing - wifi: ath11k: reduce the wait time of 11d scan and hw scan while adding an interface - mac80211: fix rx reordering with non explicit / psmp ack policy - mac80211: reset MBSSID parameters upon connection - nl80211: fix races in nl80211_set_tx_bitrate_mask() - tls: fix context leak on tls_device_down - sched: act_pedit: really ensure the skb is writable - batman-adv: don't skb_split skbuffs with frag_list - eth: ocelot: fix various issues with TC actions (null-deref; bad stats; ineffective drops; ineffective filter removal) Signed-off-by: Jakub Kicinski <kuba@kernel.org> -----BEGIN PGP SIGNATURE----- iQIzBAABCAAdFiEE6jPA+I1ugmIBA4hXMUZtbf5SIrsFAmJ9VBwACgkQMUZtbf5S IrumIg/+M4ZTUjz/34v2djiwivBHjfn3Aa2SepJ3516AqxauLUr4h4yfIcFXW8vA 81PyZkpqLBdgGklt+1GQfNCGrJKgC4+sM0eyRATMl1nf5zgrc0bZfp+ZN1bZeuoy iRB7vwCnAPdLg997inxQeG11RYS+DUDDz68MJReqnqB6o5+xX87vcDEpyKaISTVo SlUe/SYEk7S/Erk37l4wlz8bpr3EoIy1pYc4fDZOhBa8gKOVmmXidI5pwcjZXSo5 RQ9oH74/aaI7dgEQ4UvyAksCUrj17GRygBDyLE/p0eMc/2oV0lO6DQ3RGvh6ucPj wU4AAOI5SRrXiqSzbDFXLFvtN7wLuXU+s4a/xq0z/Z4w7CQb8HVH1S92lYIA+jkA Q8RCT3XY523Me8JWAyGJGl2PnbLqlGhFdqucmvHpxXOKy2F+QC69Z44kGWIaBliR C3zoUuMnAV41DIp7YXbxvxjMk8GtNUXU4Vdg7+v1DlZxxe+DDH4cxB5YmReq+bNC 60p79/vpNzJgYuZgtlk4xi1YNcOywVJb0L6zXus5p4SU/Ud63H7NHufndCnf/X6Q 9W2vWrTJb9r1+cWRBKvxyWdqyUT68UgBAbhY2WvPdmjbP5UkW7YLCMfBOsAg7vPi GZeroxYIg/NwyqWm/aUuPdiNgD73Qq4Zvan5JmavP/PwX8AP9eA= =jt7U -----END PGP SIGNATURE----- Merge tag 'net-5.18-rc7' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net Pull networking fixes from Jakub Kicinski: "Including fixes from wireless, and bluetooth. No outstanding fires. Current release - regressions: - eth: atlantic: always deep reset on pm op, fix null-deref Current release - new code bugs: - rds: use maybe_get_net() when acquiring refcount on TCP sockets [refinement of a previous fix] - eth: ocelot: mark traps with a bool instead of guessing type based on list membership Previous releases - regressions: - net: fix skipping features in for_each_netdev_feature() - phy: micrel: fix null-derefs on suspend/resume and probe - bcmgenet: check for Wake-on-LAN interrupt probe deferral Previous releases - always broken: - ipv4: drop dst in multicast routing path, prevent leaks - ping: fix address binding wrt vrf - net: fix wrong network header length when BPF protocol translation is used on skbs with a fraglist - bluetooth: fix the creation of hdev->name - rfkill: uapi: fix RFKILL_IOCTL_MAX_SIZE ioctl request definition - wifi: iwlwifi: iwl-dbg: use del_timer_sync() before freeing - wifi: ath11k: reduce the wait time of 11d scan and hw scan while adding an interface - mac80211: fix rx reordering with non explicit / psmp ack policy - mac80211: reset MBSSID parameters upon connection - nl80211: fix races in nl80211_set_tx_bitrate_mask() - tls: fix context leak on tls_device_down - sched: act_pedit: really ensure the skb is writable - batman-adv: don't skb_split skbuffs with frag_list - eth: ocelot: fix various issues with TC actions (null-deref; bad stats; ineffective drops; ineffective filter removal)" * tag 'net-5.18-rc7' of git://git.kernel.org/pub/scm/linux/kernel/git/netdev/net: (61 commits) tls: Fix context leak on tls_device_down net: sfc: ef10: fix memory leak in efx_ef10_mtd_probe() net/smc: non blocking recvmsg() return -EAGAIN when no data and signal_pending net: dsa: bcm_sf2: Fix Wake-on-LAN with mac_link_down() mlxsw: Avoid warning during ip6gre device removal net: bcmgenet: Check for Wake-on-LAN interrupt probe deferral net: ethernet: mediatek: ppe: fix wrong size passed to memset() Bluetooth: Fix the creation of hdev->name i40e: i40e_main: fix a missing check on list iterator net/sched: act_pedit: really ensure the skb is writable s390/lcs: fix variable dereferenced before check s390/ctcm: fix potential memory leak s390/ctcm: fix variable dereferenced before check net: atlantic: verify hw_head_ lies within TX buffer ring net: atlantic: add check for MAX_SKB_FRAGS net: atlantic: reduce scope of is_rsc_complete net: atlantic: fix "frag[0] not initialized" net: stmmac: fix missing pci_disable_device() on error in stmmac_pci_probe() net: phy: micrel: Fix incorrect variable type in micrel decnet: Use container_of() for struct dn_neigh casts ...
This commit is contained in:
commit
f3f19f939c
1
.mailmap
1
.mailmap
|
@ -205,6 +205,7 @@ Juha Yrjola <at solidboot.com>
|
|||
Juha Yrjola <juha.yrjola@nokia.com>
|
||||
Juha Yrjola <juha.yrjola@solidboot.com>
|
||||
Julien Thierry <julien.thierry.kdev@gmail.com> <julien.thierry@arm.com>
|
||||
Kalle Valo <kvalo@kernel.org> <kvalo@codeaurora.org>
|
||||
Kalyan Thota <quic_kalyant@quicinc.com> <kalyan_t@codeaurora.org>
|
||||
Kay Sievers <kay.sievers@vrfy.org>
|
||||
Kees Cook <keescook@chromium.org> <kees.cook@canonical.com>
|
||||
|
|
|
@ -3571,8 +3571,9 @@ M: Andy Gospodarek <andy@greyhouse.net>
|
|||
L: netdev@vger.kernel.org
|
||||
S: Supported
|
||||
W: http://sourceforge.net/projects/bonding/
|
||||
F: Documentation/networking/bonding.rst
|
||||
F: drivers/net/bonding/
|
||||
F: include/net/bonding.h
|
||||
F: include/net/bond*
|
||||
F: include/uapi/linux/if_bonding.h
|
||||
|
||||
BOSCH SENSORTEC BMA400 ACCELEROMETER IIO DRIVER
|
||||
|
@ -10131,7 +10132,7 @@ S: Supported
|
|||
F: drivers/net/wireless/intel/iwlegacy/
|
||||
|
||||
INTEL WIRELESS WIFI LINK (iwlwifi)
|
||||
M: Luca Coelho <luciano.coelho@intel.com>
|
||||
M: Gregory Greenman <gregory.greenman@intel.com>
|
||||
L: linux-wireless@vger.kernel.org
|
||||
S: Supported
|
||||
W: https://wireless.wiki.kernel.org/en/users/drivers/iwlwifi
|
||||
|
|
|
@ -809,6 +809,9 @@ static void bcm_sf2_sw_mac_link_down(struct dsa_switch *ds, int port,
|
|||
struct bcm_sf2_priv *priv = bcm_sf2_to_priv(ds);
|
||||
u32 reg, offset;
|
||||
|
||||
if (priv->wol_ports_mask & BIT(port))
|
||||
return;
|
||||
|
||||
if (port != core_readl(priv, CORE_IMP0_PRT_ID)) {
|
||||
if (priv->type == BCM4908_DEVICE_ID ||
|
||||
priv->type == BCM7445_DEVICE_ID)
|
||||
|
|
|
@ -403,6 +403,7 @@ static int felix_update_trapping_destinations(struct dsa_switch *ds,
|
|||
{
|
||||
struct ocelot *ocelot = ds->priv;
|
||||
struct felix *felix = ocelot_to_felix(ocelot);
|
||||
struct ocelot_vcap_block *block_vcap_is2;
|
||||
struct ocelot_vcap_filter *trap;
|
||||
enum ocelot_mask_mode mask_mode;
|
||||
unsigned long port_mask;
|
||||
|
@ -422,9 +423,13 @@ static int felix_update_trapping_destinations(struct dsa_switch *ds,
|
|||
/* We are sure that "cpu" was found, otherwise
|
||||
* dsa_tree_setup_default_cpu() would have failed earlier.
|
||||
*/
|
||||
block_vcap_is2 = &ocelot->block[VCAP_IS2];
|
||||
|
||||
/* Make sure all traps are set up for that destination */
|
||||
list_for_each_entry(trap, &ocelot->traps, trap_list) {
|
||||
list_for_each_entry(trap, &block_vcap_is2->rules, list) {
|
||||
if (!trap->is_trap)
|
||||
continue;
|
||||
|
||||
/* Figure out the current trapping destination */
|
||||
if (using_tag_8021q) {
|
||||
/* Redirect to the tag_8021q CPU port. If timestamps
|
||||
|
|
|
@ -449,7 +449,7 @@ static int aq_pm_freeze(struct device *dev)
|
|||
|
||||
static int aq_pm_suspend_poweroff(struct device *dev)
|
||||
{
|
||||
return aq_suspend_common(dev, false);
|
||||
return aq_suspend_common(dev, true);
|
||||
}
|
||||
|
||||
static int aq_pm_thaw(struct device *dev)
|
||||
|
@ -459,7 +459,7 @@ static int aq_pm_thaw(struct device *dev)
|
|||
|
||||
static int aq_pm_resume_restore(struct device *dev)
|
||||
{
|
||||
return atl_resume_common(dev, false);
|
||||
return atl_resume_common(dev, true);
|
||||
}
|
||||
|
||||
static const struct dev_pm_ops aq_pm_ops = {
|
||||
|
|
|
@ -346,7 +346,6 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
|
|||
int budget)
|
||||
{
|
||||
struct net_device *ndev = aq_nic_get_ndev(self->aq_nic);
|
||||
bool is_rsc_completed = true;
|
||||
int err = 0;
|
||||
|
||||
for (; (self->sw_head != self->hw_head) && budget;
|
||||
|
@ -364,12 +363,17 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
|
|||
continue;
|
||||
|
||||
if (!buff->is_eop) {
|
||||
unsigned int frag_cnt = 0U;
|
||||
buff_ = buff;
|
||||
do {
|
||||
bool is_rsc_completed = true;
|
||||
|
||||
if (buff_->next >= self->size) {
|
||||
err = -EIO;
|
||||
goto err_exit;
|
||||
}
|
||||
|
||||
frag_cnt++;
|
||||
next_ = buff_->next,
|
||||
buff_ = &self->buff_ring[next_];
|
||||
is_rsc_completed =
|
||||
|
@ -377,18 +381,17 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
|
|||
next_,
|
||||
self->hw_head);
|
||||
|
||||
if (unlikely(!is_rsc_completed))
|
||||
break;
|
||||
if (unlikely(!is_rsc_completed) ||
|
||||
frag_cnt > MAX_SKB_FRAGS) {
|
||||
err = 0;
|
||||
goto err_exit;
|
||||
}
|
||||
|
||||
buff->is_error |= buff_->is_error;
|
||||
buff->is_cso_err |= buff_->is_cso_err;
|
||||
|
||||
} while (!buff_->is_eop);
|
||||
|
||||
if (!is_rsc_completed) {
|
||||
err = 0;
|
||||
goto err_exit;
|
||||
}
|
||||
if (buff->is_error ||
|
||||
(buff->is_lro && buff->is_cso_err)) {
|
||||
buff_ = buff;
|
||||
|
@ -446,7 +449,7 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
|
|||
ALIGN(hdr_len, sizeof(long)));
|
||||
|
||||
if (buff->len - hdr_len > 0) {
|
||||
skb_add_rx_frag(skb, 0, buff->rxdata.page,
|
||||
skb_add_rx_frag(skb, i++, buff->rxdata.page,
|
||||
buff->rxdata.pg_off + hdr_len,
|
||||
buff->len - hdr_len,
|
||||
AQ_CFG_RX_FRAME_MAX);
|
||||
|
@ -455,7 +458,6 @@ int aq_ring_rx_clean(struct aq_ring_s *self,
|
|||
|
||||
if (!buff->is_eop) {
|
||||
buff_ = buff;
|
||||
i = 1U;
|
||||
do {
|
||||
next_ = buff_->next;
|
||||
buff_ = &self->buff_ring[next_];
|
||||
|
|
|
@ -889,6 +889,13 @@ int hw_atl_b0_hw_ring_tx_head_update(struct aq_hw_s *self,
|
|||
err = -ENXIO;
|
||||
goto err_exit;
|
||||
}
|
||||
|
||||
/* Validate that the new hw_head_ is reasonable. */
|
||||
if (hw_head_ >= ring->size) {
|
||||
err = -ENXIO;
|
||||
goto err_exit;
|
||||
}
|
||||
|
||||
ring->hw_head = hw_head_;
|
||||
err = aq_hw_err_from_flags(self);
|
||||
|
||||
|
|
|
@ -3999,6 +3999,10 @@ static int bcmgenet_probe(struct platform_device *pdev)
|
|||
goto err;
|
||||
}
|
||||
priv->wol_irq = platform_get_irq_optional(pdev, 2);
|
||||
if (priv->wol_irq == -EPROBE_DEFER) {
|
||||
err = priv->wol_irq;
|
||||
goto err;
|
||||
}
|
||||
|
||||
priv->base = devm_platform_ioremap_resource(pdev, 0);
|
||||
if (IS_ERR(priv->base)) {
|
||||
|
|
|
@ -2793,14 +2793,14 @@ int t4_get_raw_vpd_params(struct adapter *adapter, struct vpd_params *p)
|
|||
goto out;
|
||||
na = ret;
|
||||
|
||||
memcpy(p->id, vpd + id, min_t(int, id_len, ID_LEN));
|
||||
memcpy(p->id, vpd + id, min_t(unsigned int, id_len, ID_LEN));
|
||||
strim(p->id);
|
||||
memcpy(p->sn, vpd + sn, min_t(int, sn_len, SERNUM_LEN));
|
||||
memcpy(p->sn, vpd + sn, min_t(unsigned int, sn_len, SERNUM_LEN));
|
||||
strim(p->sn);
|
||||
memcpy(p->pn, vpd + pn, min_t(int, pn_len, PN_LEN));
|
||||
memcpy(p->pn, vpd + pn, min_t(unsigned int, pn_len, PN_LEN));
|
||||
strim(p->pn);
|
||||
memcpy(p->na, vpd + na, min_t(int, na_len, MACADDR_LEN));
|
||||
strim((char *)p->na);
|
||||
memcpy(p->na, vpd + na, min_t(unsigned int, na_len, MACADDR_LEN));
|
||||
strim(p->na);
|
||||
|
||||
out:
|
||||
vfree(vpd);
|
||||
|
|
|
@ -1399,8 +1399,10 @@ static int tulip_init_one(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||
|
||||
/* alloc_etherdev ensures aligned and zeroed private structures */
|
||||
dev = alloc_etherdev (sizeof (*tp));
|
||||
if (!dev)
|
||||
if (!dev) {
|
||||
pci_disable_device(pdev);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
SET_NETDEV_DEV(dev, &pdev->dev);
|
||||
if (pci_resource_len (pdev, 0) < tulip_tbl[chip_idx].io_size) {
|
||||
|
@ -1785,6 +1787,7 @@ err_out_free_res:
|
|||
|
||||
err_out_free_netdev:
|
||||
free_netdev (dev);
|
||||
pci_disable_device(pdev);
|
||||
return -ENODEV;
|
||||
}
|
||||
|
||||
|
|
|
@ -7549,42 +7549,43 @@ static void i40e_free_macvlan_channels(struct i40e_vsi *vsi)
|
|||
static int i40e_fwd_ring_up(struct i40e_vsi *vsi, struct net_device *vdev,
|
||||
struct i40e_fwd_adapter *fwd)
|
||||
{
|
||||
struct i40e_channel *ch = NULL, *ch_tmp, *iter;
|
||||
int ret = 0, num_tc = 1, i, aq_err;
|
||||
struct i40e_channel *ch, *ch_tmp;
|
||||
struct i40e_pf *pf = vsi->back;
|
||||
struct i40e_hw *hw = &pf->hw;
|
||||
|
||||
if (list_empty(&vsi->macvlan_list))
|
||||
return -EINVAL;
|
||||
|
||||
/* Go through the list and find an available channel */
|
||||
list_for_each_entry_safe(ch, ch_tmp, &vsi->macvlan_list, list) {
|
||||
if (!i40e_is_channel_macvlan(ch)) {
|
||||
ch->fwd = fwd;
|
||||
list_for_each_entry_safe(iter, ch_tmp, &vsi->macvlan_list, list) {
|
||||
if (!i40e_is_channel_macvlan(iter)) {
|
||||
iter->fwd = fwd;
|
||||
/* record configuration for macvlan interface in vdev */
|
||||
for (i = 0; i < num_tc; i++)
|
||||
netdev_bind_sb_channel_queue(vsi->netdev, vdev,
|
||||
i,
|
||||
ch->num_queue_pairs,
|
||||
ch->base_queue);
|
||||
for (i = 0; i < ch->num_queue_pairs; i++) {
|
||||
iter->num_queue_pairs,
|
||||
iter->base_queue);
|
||||
for (i = 0; i < iter->num_queue_pairs; i++) {
|
||||
struct i40e_ring *tx_ring, *rx_ring;
|
||||
u16 pf_q;
|
||||
|
||||
pf_q = ch->base_queue + i;
|
||||
pf_q = iter->base_queue + i;
|
||||
|
||||
/* Get to TX ring ptr */
|
||||
tx_ring = vsi->tx_rings[pf_q];
|
||||
tx_ring->ch = ch;
|
||||
tx_ring->ch = iter;
|
||||
|
||||
/* Get the RX ring ptr */
|
||||
rx_ring = vsi->rx_rings[pf_q];
|
||||
rx_ring->ch = ch;
|
||||
rx_ring->ch = iter;
|
||||
}
|
||||
ch = iter;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (!ch)
|
||||
return -EINVAL;
|
||||
|
||||
/* Guarantee all rings are updated before we update the
|
||||
* MAC address filter.
|
||||
*/
|
||||
|
|
|
@ -540,6 +540,7 @@ struct ice_pf {
|
|||
struct mutex avail_q_mutex; /* protects access to avail_[rx|tx]qs */
|
||||
struct mutex sw_mutex; /* lock for protecting VSI alloc flow */
|
||||
struct mutex tc_mutex; /* lock to protect TC changes */
|
||||
struct mutex adev_mutex; /* lock to protect aux device access */
|
||||
u32 msg_enable;
|
||||
struct ice_ptp ptp;
|
||||
struct tty_driver *ice_gnss_tty_driver;
|
||||
|
|
|
@ -37,14 +37,17 @@ void ice_send_event_to_aux(struct ice_pf *pf, struct iidc_event *event)
|
|||
if (WARN_ON_ONCE(!in_task()))
|
||||
return;
|
||||
|
||||
mutex_lock(&pf->adev_mutex);
|
||||
if (!pf->adev)
|
||||
return;
|
||||
goto finish;
|
||||
|
||||
device_lock(&pf->adev->dev);
|
||||
iadrv = ice_get_auxiliary_drv(pf);
|
||||
if (iadrv && iadrv->event_handler)
|
||||
iadrv->event_handler(pf, event);
|
||||
device_unlock(&pf->adev->dev);
|
||||
finish:
|
||||
mutex_unlock(&pf->adev_mutex);
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -290,7 +293,6 @@ int ice_plug_aux_dev(struct ice_pf *pf)
|
|||
return -ENOMEM;
|
||||
|
||||
adev = &iadev->adev;
|
||||
pf->adev = adev;
|
||||
iadev->pf = pf;
|
||||
|
||||
adev->id = pf->aux_idx;
|
||||
|
@ -300,18 +302,20 @@ int ice_plug_aux_dev(struct ice_pf *pf)
|
|||
|
||||
ret = auxiliary_device_init(adev);
|
||||
if (ret) {
|
||||
pf->adev = NULL;
|
||||
kfree(iadev);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ret = auxiliary_device_add(adev);
|
||||
if (ret) {
|
||||
pf->adev = NULL;
|
||||
auxiliary_device_uninit(adev);
|
||||
return ret;
|
||||
}
|
||||
|
||||
mutex_lock(&pf->adev_mutex);
|
||||
pf->adev = adev;
|
||||
mutex_unlock(&pf->adev_mutex);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -320,12 +324,17 @@ int ice_plug_aux_dev(struct ice_pf *pf)
|
|||
*/
|
||||
void ice_unplug_aux_dev(struct ice_pf *pf)
|
||||
{
|
||||
if (!pf->adev)
|
||||
return;
|
||||
struct auxiliary_device *adev;
|
||||
|
||||
auxiliary_device_delete(pf->adev);
|
||||
auxiliary_device_uninit(pf->adev);
|
||||
mutex_lock(&pf->adev_mutex);
|
||||
adev = pf->adev;
|
||||
pf->adev = NULL;
|
||||
mutex_unlock(&pf->adev_mutex);
|
||||
|
||||
if (adev) {
|
||||
auxiliary_device_delete(adev);
|
||||
auxiliary_device_uninit(adev);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -3769,6 +3769,7 @@ u16 ice_get_avail_rxq_count(struct ice_pf *pf)
|
|||
static void ice_deinit_pf(struct ice_pf *pf)
|
||||
{
|
||||
ice_service_task_stop(pf);
|
||||
mutex_destroy(&pf->adev_mutex);
|
||||
mutex_destroy(&pf->sw_mutex);
|
||||
mutex_destroy(&pf->tc_mutex);
|
||||
mutex_destroy(&pf->avail_q_mutex);
|
||||
|
@ -3847,6 +3848,7 @@ static int ice_init_pf(struct ice_pf *pf)
|
|||
|
||||
mutex_init(&pf->sw_mutex);
|
||||
mutex_init(&pf->tc_mutex);
|
||||
mutex_init(&pf->adev_mutex);
|
||||
|
||||
INIT_HLIST_HEAD(&pf->aq_wait_list);
|
||||
spin_lock_init(&pf->aq_wait_lock);
|
||||
|
|
|
@ -2287,6 +2287,7 @@ ice_ptp_init_tx_e810(struct ice_pf *pf, struct ice_ptp_tx *tx)
|
|||
|
||||
/**
|
||||
* ice_ptp_tx_tstamp_cleanup - Cleanup old timestamp requests that got dropped
|
||||
* @hw: pointer to the hw struct
|
||||
* @tx: PTP Tx tracker to clean up
|
||||
*
|
||||
* Loop through the Tx timestamp requests and see if any of them have been
|
||||
|
@ -2295,7 +2296,7 @@ ice_ptp_init_tx_e810(struct ice_pf *pf, struct ice_ptp_tx *tx)
|
|||
* timestamp will never be captured. This might happen if the packet gets
|
||||
* discarded before it reaches the PHY timestamping block.
|
||||
*/
|
||||
static void ice_ptp_tx_tstamp_cleanup(struct ice_ptp_tx *tx)
|
||||
static void ice_ptp_tx_tstamp_cleanup(struct ice_hw *hw, struct ice_ptp_tx *tx)
|
||||
{
|
||||
u8 idx;
|
||||
|
||||
|
@ -2304,11 +2305,16 @@ static void ice_ptp_tx_tstamp_cleanup(struct ice_ptp_tx *tx)
|
|||
|
||||
for_each_set_bit(idx, tx->in_use, tx->len) {
|
||||
struct sk_buff *skb;
|
||||
u64 raw_tstamp;
|
||||
|
||||
/* Check if this SKB has been waiting for too long */
|
||||
if (time_is_after_jiffies(tx->tstamps[idx].start + 2 * HZ))
|
||||
continue;
|
||||
|
||||
/* Read tstamp to be able to use this register again */
|
||||
ice_read_phy_tstamp(hw, tx->quad, idx + tx->quad_offset,
|
||||
&raw_tstamp);
|
||||
|
||||
spin_lock(&tx->lock);
|
||||
skb = tx->tstamps[idx].skb;
|
||||
tx->tstamps[idx].skb = NULL;
|
||||
|
@ -2330,7 +2336,7 @@ static void ice_ptp_periodic_work(struct kthread_work *work)
|
|||
|
||||
ice_ptp_update_cached_phctime(pf);
|
||||
|
||||
ice_ptp_tx_tstamp_cleanup(&pf->ptp.port.tx);
|
||||
ice_ptp_tx_tstamp_cleanup(&pf->hw, &pf->ptp.port.tx);
|
||||
|
||||
/* Run twice a second */
|
||||
kthread_queue_delayed_work(ptp->kworker, &ptp->work,
|
||||
|
|
|
@ -1307,13 +1307,52 @@ error_param:
|
|||
NULL, 0);
|
||||
}
|
||||
|
||||
/**
|
||||
* ice_vf_vsi_dis_single_txq - disable a single Tx queue
|
||||
* @vf: VF to disable queue for
|
||||
* @vsi: VSI for the VF
|
||||
* @q_id: VF relative (0-based) queue ID
|
||||
*
|
||||
* Attempt to disable the Tx queue passed in. If the Tx queue was successfully
|
||||
* disabled then clear q_id bit in the enabled queues bitmap and return
|
||||
* success. Otherwise return error.
|
||||
*/
|
||||
static int
|
||||
ice_vf_vsi_dis_single_txq(struct ice_vf *vf, struct ice_vsi *vsi, u16 q_id)
|
||||
{
|
||||
struct ice_txq_meta txq_meta = { 0 };
|
||||
struct ice_tx_ring *ring;
|
||||
int err;
|
||||
|
||||
if (!test_bit(q_id, vf->txq_ena))
|
||||
dev_dbg(ice_pf_to_dev(vsi->back), "Queue %u on VSI %u is not enabled, but stopping it anyway\n",
|
||||
q_id, vsi->vsi_num);
|
||||
|
||||
ring = vsi->tx_rings[q_id];
|
||||
if (!ring)
|
||||
return -EINVAL;
|
||||
|
||||
ice_fill_txq_meta(vsi, ring, &txq_meta);
|
||||
|
||||
err = ice_vsi_stop_tx_ring(vsi, ICE_NO_RESET, vf->vf_id, ring, &txq_meta);
|
||||
if (err) {
|
||||
dev_err(ice_pf_to_dev(vsi->back), "Failed to stop Tx ring %d on VSI %d\n",
|
||||
q_id, vsi->vsi_num);
|
||||
return err;
|
||||
}
|
||||
|
||||
/* Clear enabled queues flag */
|
||||
clear_bit(q_id, vf->txq_ena);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* ice_vc_dis_qs_msg
|
||||
* @vf: pointer to the VF info
|
||||
* @msg: pointer to the msg buffer
|
||||
*
|
||||
* called from the VF to disable all or specific
|
||||
* queue(s)
|
||||
* called from the VF to disable all or specific queue(s)
|
||||
*/
|
||||
static int ice_vc_dis_qs_msg(struct ice_vf *vf, u8 *msg)
|
||||
{
|
||||
|
@ -1350,30 +1389,15 @@ static int ice_vc_dis_qs_msg(struct ice_vf *vf, u8 *msg)
|
|||
q_map = vqs->tx_queues;
|
||||
|
||||
for_each_set_bit(vf_q_id, &q_map, ICE_MAX_RSS_QS_PER_VF) {
|
||||
struct ice_tx_ring *ring = vsi->tx_rings[vf_q_id];
|
||||
struct ice_txq_meta txq_meta = { 0 };
|
||||
|
||||
if (!ice_vc_isvalid_q_id(vf, vqs->vsi_id, vf_q_id)) {
|
||||
v_ret = VIRTCHNL_STATUS_ERR_PARAM;
|
||||
goto error_param;
|
||||
}
|
||||
|
||||
if (!test_bit(vf_q_id, vf->txq_ena))
|
||||
dev_dbg(ice_pf_to_dev(vsi->back), "Queue %u on VSI %u is not enabled, but stopping it anyway\n",
|
||||
vf_q_id, vsi->vsi_num);
|
||||
|
||||
ice_fill_txq_meta(vsi, ring, &txq_meta);
|
||||
|
||||
if (ice_vsi_stop_tx_ring(vsi, ICE_NO_RESET, vf->vf_id,
|
||||
ring, &txq_meta)) {
|
||||
dev_err(ice_pf_to_dev(vsi->back), "Failed to stop Tx ring %d on VSI %d\n",
|
||||
vf_q_id, vsi->vsi_num);
|
||||
if (ice_vf_vsi_dis_single_txq(vf, vsi, vf_q_id)) {
|
||||
v_ret = VIRTCHNL_STATUS_ERR_PARAM;
|
||||
goto error_param;
|
||||
}
|
||||
|
||||
/* Clear enabled queues flag */
|
||||
clear_bit(vf_q_id, vf->txq_ena);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -1622,6 +1646,14 @@ static int ice_vc_cfg_qs_msg(struct ice_vf *vf, u8 *msg)
|
|||
if (qpi->txq.ring_len > 0) {
|
||||
vsi->tx_rings[i]->dma = qpi->txq.dma_ring_addr;
|
||||
vsi->tx_rings[i]->count = qpi->txq.ring_len;
|
||||
|
||||
/* Disable any existing queue first */
|
||||
if (ice_vf_vsi_dis_single_txq(vf, vsi, q_idx)) {
|
||||
v_ret = VIRTCHNL_STATUS_ERR_PARAM;
|
||||
goto error_param;
|
||||
}
|
||||
|
||||
/* Configure a queue with the requested settings */
|
||||
if (ice_vsi_cfg_single_txq(vsi, vsi->tx_rings, q_idx)) {
|
||||
v_ret = VIRTCHNL_STATUS_ERR_PARAM;
|
||||
goto error_param;
|
||||
|
|
|
@ -395,7 +395,7 @@ static void mtk_ppe_init_foe_table(struct mtk_ppe *ppe)
|
|||
static const u8 skip[] = { 12, 25, 38, 51, 76, 89, 102 };
|
||||
int i, k;
|
||||
|
||||
memset(ppe->foe_table, 0, MTK_PPE_ENTRIES * sizeof(ppe->foe_table));
|
||||
memset(ppe->foe_table, 0, MTK_PPE_ENTRIES * sizeof(*ppe->foe_table));
|
||||
|
||||
if (!IS_ENABLED(CONFIG_SOC_MT7621))
|
||||
return;
|
||||
|
|
|
@ -568,10 +568,8 @@ static int
|
|||
mlxsw_sp2_ipip_rem_addr_set_gre6(struct mlxsw_sp *mlxsw_sp,
|
||||
struct mlxsw_sp_ipip_entry *ipip_entry)
|
||||
{
|
||||
struct __ip6_tnl_parm parms6;
|
||||
|
||||
parms6 = mlxsw_sp_ipip_netdev_parms6(ipip_entry->ol_dev);
|
||||
return mlxsw_sp_ipv6_addr_kvdl_index_get(mlxsw_sp, &parms6.raddr,
|
||||
return mlxsw_sp_ipv6_addr_kvdl_index_get(mlxsw_sp,
|
||||
&ipip_entry->parms.daddr.addr6,
|
||||
&ipip_entry->dip_kvdl_index);
|
||||
}
|
||||
|
||||
|
@ -579,10 +577,7 @@ static void
|
|||
mlxsw_sp2_ipip_rem_addr_unset_gre6(struct mlxsw_sp *mlxsw_sp,
|
||||
const struct mlxsw_sp_ipip_entry *ipip_entry)
|
||||
{
|
||||
struct __ip6_tnl_parm parms6;
|
||||
|
||||
parms6 = mlxsw_sp_ipip_netdev_parms6(ipip_entry->ol_dev);
|
||||
mlxsw_sp_ipv6_addr_put(mlxsw_sp, &parms6.raddr);
|
||||
mlxsw_sp_ipv6_addr_put(mlxsw_sp, &ipip_entry->parms.daddr.addr6);
|
||||
}
|
||||
|
||||
static const struct mlxsw_sp_ipip_ops mlxsw_sp2_ipip_gre6_ops = {
|
||||
|
|
|
@ -1622,7 +1622,7 @@ int ocelot_trap_add(struct ocelot *ocelot, int port,
|
|||
trap->action.mask_mode = OCELOT_MASK_MODE_PERMIT_DENY;
|
||||
trap->action.port_mask = 0;
|
||||
trap->take_ts = take_ts;
|
||||
list_add_tail(&trap->trap_list, &ocelot->traps);
|
||||
trap->is_trap = true;
|
||||
new = true;
|
||||
}
|
||||
|
||||
|
@ -1634,10 +1634,8 @@ int ocelot_trap_add(struct ocelot *ocelot, int port,
|
|||
err = ocelot_vcap_filter_replace(ocelot, trap);
|
||||
if (err) {
|
||||
trap->ingress_port_mask &= ~BIT(port);
|
||||
if (!trap->ingress_port_mask) {
|
||||
list_del(&trap->trap_list);
|
||||
if (!trap->ingress_port_mask)
|
||||
kfree(trap);
|
||||
}
|
||||
return err;
|
||||
}
|
||||
|
||||
|
@ -1657,11 +1655,8 @@ int ocelot_trap_del(struct ocelot *ocelot, int port, unsigned long cookie)
|
|||
return 0;
|
||||
|
||||
trap->ingress_port_mask &= ~BIT(port);
|
||||
if (!trap->ingress_port_mask) {
|
||||
list_del(&trap->trap_list);
|
||||
|
||||
if (!trap->ingress_port_mask)
|
||||
return ocelot_vcap_filter_del(ocelot, trap);
|
||||
}
|
||||
|
||||
return ocelot_vcap_filter_replace(ocelot, trap);
|
||||
}
|
||||
|
|
|
@ -280,9 +280,10 @@ static int ocelot_flower_parse_action(struct ocelot *ocelot, int port,
|
|||
filter->type = OCELOT_VCAP_FILTER_OFFLOAD;
|
||||
break;
|
||||
case FLOW_ACTION_TRAP:
|
||||
if (filter->block_id != VCAP_IS2) {
|
||||
if (filter->block_id != VCAP_IS2 ||
|
||||
filter->lookup != 0) {
|
||||
NL_SET_ERR_MSG_MOD(extack,
|
||||
"Trap action can only be offloaded to VCAP IS2");
|
||||
"Trap action can only be offloaded to VCAP IS2 lookup 0");
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
if (filter->goto_target != -1) {
|
||||
|
@ -295,7 +296,7 @@ static int ocelot_flower_parse_action(struct ocelot *ocelot, int port,
|
|||
filter->action.cpu_copy_ena = true;
|
||||
filter->action.cpu_qu_num = 0;
|
||||
filter->type = OCELOT_VCAP_FILTER_OFFLOAD;
|
||||
list_add_tail(&filter->trap_list, &ocelot->traps);
|
||||
filter->is_trap = true;
|
||||
break;
|
||||
case FLOW_ACTION_POLICE:
|
||||
if (filter->block_id == PSFP_BLOCK_ID) {
|
||||
|
@ -878,8 +879,6 @@ int ocelot_cls_flower_replace(struct ocelot *ocelot, int port,
|
|||
|
||||
ret = ocelot_flower_parse(ocelot, port, ingress, f, filter);
|
||||
if (ret) {
|
||||
if (!list_empty(&filter->trap_list))
|
||||
list_del(&filter->trap_list);
|
||||
kfree(filter);
|
||||
return ret;
|
||||
}
|
||||
|
|
|
@ -374,7 +374,6 @@ static void is2_entry_set(struct ocelot *ocelot, int ix,
|
|||
OCELOT_VCAP_BIT_0);
|
||||
vcap_key_set(vcap, &data, VCAP_IS2_HK_IGR_PORT_MASK, 0,
|
||||
~filter->ingress_port_mask);
|
||||
vcap_key_bit_set(vcap, &data, VCAP_IS2_HK_FIRST, OCELOT_VCAP_BIT_ANY);
|
||||
vcap_key_bit_set(vcap, &data, VCAP_IS2_HK_HOST_MATCH,
|
||||
OCELOT_VCAP_BIT_ANY);
|
||||
vcap_key_bit_set(vcap, &data, VCAP_IS2_HK_L2_MC, filter->dmac_mc);
|
||||
|
@ -1217,6 +1216,8 @@ int ocelot_vcap_filter_add(struct ocelot *ocelot,
|
|||
struct ocelot_vcap_filter *tmp;
|
||||
|
||||
tmp = ocelot_vcap_block_find_filter_by_index(block, i);
|
||||
/* Read back the filter's counters before moving it */
|
||||
vcap_entry_get(ocelot, i - 1, tmp);
|
||||
vcap_entry_set(ocelot, i, tmp);
|
||||
}
|
||||
|
||||
|
@ -1250,7 +1251,11 @@ int ocelot_vcap_filter_del(struct ocelot *ocelot,
|
|||
struct ocelot_vcap_filter del_filter;
|
||||
int i, index;
|
||||
|
||||
/* Need to inherit the block_id so that vcap_entry_set()
|
||||
* does not get confused and knows where to install it.
|
||||
*/
|
||||
memset(&del_filter, 0, sizeof(del_filter));
|
||||
del_filter.block_id = filter->block_id;
|
||||
|
||||
/* Gets index of the filter */
|
||||
index = ocelot_vcap_block_get_filter_index(block, filter);
|
||||
|
@ -1265,6 +1270,8 @@ int ocelot_vcap_filter_del(struct ocelot *ocelot,
|
|||
struct ocelot_vcap_filter *tmp;
|
||||
|
||||
tmp = ocelot_vcap_block_find_filter_by_index(block, i);
|
||||
/* Read back the filter's counters before moving it */
|
||||
vcap_entry_get(ocelot, i + 1, tmp);
|
||||
vcap_entry_set(ocelot, i, tmp);
|
||||
}
|
||||
|
||||
|
|
|
@ -256,7 +256,7 @@ static int ionic_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|||
|
||||
err = ionic_map_bars(ionic);
|
||||
if (err)
|
||||
goto err_out_pci_disable_device;
|
||||
goto err_out_pci_release_regions;
|
||||
|
||||
/* Configure the device */
|
||||
err = ionic_setup(ionic);
|
||||
|
@ -360,6 +360,7 @@ err_out_teardown:
|
|||
|
||||
err_out_unmap_bars:
|
||||
ionic_unmap_bars(ionic);
|
||||
err_out_pci_release_regions:
|
||||
pci_release_regions(pdev);
|
||||
err_out_pci_disable_device:
|
||||
pci_disable_device(pdev);
|
||||
|
|
|
@ -3579,6 +3579,11 @@ static int efx_ef10_mtd_probe(struct efx_nic *efx)
|
|||
n_parts++;
|
||||
}
|
||||
|
||||
if (!n_parts) {
|
||||
kfree(parts);
|
||||
return 0;
|
||||
}
|
||||
|
||||
rc = efx_mtd_add(efx, &parts[0].common, n_parts, sizeof(*parts));
|
||||
fail:
|
||||
if (rc)
|
||||
|
|
|
@ -867,7 +867,9 @@ static void efx_set_xdp_channels(struct efx_nic *efx)
|
|||
|
||||
int efx_realloc_channels(struct efx_nic *efx, u32 rxq_entries, u32 txq_entries)
|
||||
{
|
||||
struct efx_channel *other_channel[EFX_MAX_CHANNELS], *channel;
|
||||
struct efx_channel *other_channel[EFX_MAX_CHANNELS], *channel,
|
||||
*ptp_channel = efx_ptp_channel(efx);
|
||||
struct efx_ptp_data *ptp_data = efx->ptp_data;
|
||||
unsigned int i, next_buffer_table = 0;
|
||||
u32 old_rxq_entries, old_txq_entries;
|
||||
int rc, rc2;
|
||||
|
@ -938,6 +940,7 @@ int efx_realloc_channels(struct efx_nic *efx, u32 rxq_entries, u32 txq_entries)
|
|||
|
||||
efx_set_xdp_channels(efx);
|
||||
out:
|
||||
efx->ptp_data = NULL;
|
||||
/* Destroy unused channel structures */
|
||||
for (i = 0; i < efx->n_channels; i++) {
|
||||
channel = other_channel[i];
|
||||
|
@ -948,6 +951,7 @@ out:
|
|||
}
|
||||
}
|
||||
|
||||
efx->ptp_data = ptp_data;
|
||||
rc2 = efx_soft_enable_interrupts(efx);
|
||||
if (rc2) {
|
||||
rc = rc ? rc : rc2;
|
||||
|
@ -966,6 +970,7 @@ rollback:
|
|||
efx->txq_entries = old_txq_entries;
|
||||
for (i = 0; i < efx->n_channels; i++)
|
||||
swap(efx->channel[i], other_channel[i]);
|
||||
efx_ptp_update_channel(efx, ptp_channel);
|
||||
goto out;
|
||||
}
|
||||
|
||||
|
|
|
@ -45,6 +45,7 @@
|
|||
#include "farch_regs.h"
|
||||
#include "tx.h"
|
||||
#include "nic.h" /* indirectly includes ptp.h */
|
||||
#include "efx_channels.h"
|
||||
|
||||
/* Maximum number of events expected to make up a PTP event */
|
||||
#define MAX_EVENT_FRAGS 3
|
||||
|
@ -541,6 +542,12 @@ struct efx_channel *efx_ptp_channel(struct efx_nic *efx)
|
|||
return efx->ptp_data ? efx->ptp_data->channel : NULL;
|
||||
}
|
||||
|
||||
void efx_ptp_update_channel(struct efx_nic *efx, struct efx_channel *channel)
|
||||
{
|
||||
if (efx->ptp_data)
|
||||
efx->ptp_data->channel = channel;
|
||||
}
|
||||
|
||||
static u32 last_sync_timestamp_major(struct efx_nic *efx)
|
||||
{
|
||||
struct efx_channel *channel = efx_ptp_channel(efx);
|
||||
|
@ -1443,6 +1450,11 @@ int efx_ptp_probe(struct efx_nic *efx, struct efx_channel *channel)
|
|||
int rc = 0;
|
||||
unsigned int pos;
|
||||
|
||||
if (efx->ptp_data) {
|
||||
efx->ptp_data->channel = channel;
|
||||
return 0;
|
||||
}
|
||||
|
||||
ptp = kzalloc(sizeof(struct efx_ptp_data), GFP_KERNEL);
|
||||
efx->ptp_data = ptp;
|
||||
if (!efx->ptp_data)
|
||||
|
@ -2176,7 +2188,7 @@ static const struct efx_channel_type efx_ptp_channel_type = {
|
|||
.pre_probe = efx_ptp_probe_channel,
|
||||
.post_remove = efx_ptp_remove_channel,
|
||||
.get_name = efx_ptp_get_channel_name,
|
||||
/* no copy operation; there is no need to reallocate this channel */
|
||||
.copy = efx_copy_channel,
|
||||
.receive_skb = efx_ptp_rx,
|
||||
.want_txqs = efx_ptp_want_txqs,
|
||||
.keep_eventq = false,
|
||||
|
|
|
@ -16,6 +16,7 @@ struct ethtool_ts_info;
|
|||
int efx_ptp_probe(struct efx_nic *efx, struct efx_channel *channel);
|
||||
void efx_ptp_defer_probe_with_channel(struct efx_nic *efx);
|
||||
struct efx_channel *efx_ptp_channel(struct efx_nic *efx);
|
||||
void efx_ptp_update_channel(struct efx_nic *efx, struct efx_channel *channel);
|
||||
void efx_ptp_remove(struct efx_nic *efx);
|
||||
int efx_ptp_set_ts_config(struct efx_nic *efx, struct ifreq *ifr);
|
||||
int efx_ptp_get_ts_config(struct efx_nic *efx, struct ifreq *ifr);
|
||||
|
|
|
@ -181,7 +181,7 @@ static int stmmac_pci_probe(struct pci_dev *pdev,
|
|||
return -ENOMEM;
|
||||
|
||||
/* Enable pci device */
|
||||
ret = pci_enable_device(pdev);
|
||||
ret = pcim_enable_device(pdev);
|
||||
if (ret) {
|
||||
dev_err(&pdev->dev, "%s: ERROR: failed to enable device\n",
|
||||
__func__);
|
||||
|
@ -241,8 +241,6 @@ static void stmmac_pci_remove(struct pci_dev *pdev)
|
|||
pcim_iounmap_regions(pdev, BIT(i));
|
||||
break;
|
||||
}
|
||||
|
||||
pci_disable_device(pdev);
|
||||
}
|
||||
|
||||
static int __maybe_unused stmmac_pci_suspend(struct device *dev)
|
||||
|
|
|
@ -1743,7 +1743,7 @@ static int ksz886x_cable_test_get_status(struct phy_device *phydev,
|
|||
|
||||
static int lanphy_read_page_reg(struct phy_device *phydev, int page, u32 addr)
|
||||
{
|
||||
u32 data;
|
||||
int data;
|
||||
|
||||
phy_lock_mdio_bus(phydev);
|
||||
__phy_write(phydev, LAN_EXT_PAGE_ACCESS_CONTROL, page);
|
||||
|
@ -2444,8 +2444,7 @@ static int lan8804_config_init(struct phy_device *phydev)
|
|||
|
||||
static irqreturn_t lan8814_handle_interrupt(struct phy_device *phydev)
|
||||
{
|
||||
u16 tsu_irq_status;
|
||||
int irq_status;
|
||||
int irq_status, tsu_irq_status;
|
||||
|
||||
irq_status = phy_read(phydev, LAN8814_INTS);
|
||||
if (irq_status > 0 && (irq_status & LAN8814_INT_LINK))
|
||||
|
@ -2657,6 +2656,7 @@ static struct phy_driver ksphy_driver[] = {
|
|||
.name = "Micrel KS8737",
|
||||
/* PHY_BASIC_FEATURES */
|
||||
.driver_data = &ks8737_type,
|
||||
.probe = kszphy_probe,
|
||||
.config_init = kszphy_config_init,
|
||||
.config_intr = kszphy_config_intr,
|
||||
.handle_interrupt = kszphy_handle_interrupt,
|
||||
|
@ -2782,8 +2782,8 @@ static struct phy_driver ksphy_driver[] = {
|
|||
.config_init = ksz8061_config_init,
|
||||
.config_intr = kszphy_config_intr,
|
||||
.handle_interrupt = kszphy_handle_interrupt,
|
||||
.suspend = kszphy_suspend,
|
||||
.resume = kszphy_resume,
|
||||
.suspend = genphy_suspend,
|
||||
.resume = genphy_resume,
|
||||
}, {
|
||||
.phy_id = PHY_ID_KSZ9021,
|
||||
.phy_id_mask = 0x000ffffe,
|
||||
|
|
|
@ -970,8 +970,13 @@ static irqreturn_t phy_interrupt(int irq, void *phy_dat)
|
|||
{
|
||||
struct phy_device *phydev = phy_dat;
|
||||
struct phy_driver *drv = phydev->drv;
|
||||
irqreturn_t ret;
|
||||
|
||||
return drv->handle_interrupt(phydev);
|
||||
mutex_lock(&phydev->lock);
|
||||
ret = drv->handle_interrupt(phydev);
|
||||
mutex_unlock(&phydev->lock);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -1288,6 +1288,7 @@ static void ath11k_core_restart(struct work_struct *work)
|
|||
|
||||
ieee80211_stop_queues(ar->hw);
|
||||
ath11k_mac_drain_tx(ar);
|
||||
complete(&ar->completed_11d_scan);
|
||||
complete(&ar->scan.started);
|
||||
complete(&ar->scan.completed);
|
||||
complete(&ar->peer_assoc_done);
|
||||
|
|
|
@ -38,6 +38,8 @@
|
|||
|
||||
extern unsigned int ath11k_frame_mode;
|
||||
|
||||
#define ATH11K_SCAN_TIMEOUT_HZ (20 * HZ)
|
||||
|
||||
#define ATH11K_MON_TIMER_INTERVAL 10
|
||||
|
||||
enum ath11k_supported_bw {
|
||||
|
@ -189,6 +191,12 @@ enum ath11k_scan_state {
|
|||
ATH11K_SCAN_ABORTING,
|
||||
};
|
||||
|
||||
enum ath11k_11d_state {
|
||||
ATH11K_11D_IDLE,
|
||||
ATH11K_11D_PREPARING,
|
||||
ATH11K_11D_RUNNING,
|
||||
};
|
||||
|
||||
enum ath11k_dev_flags {
|
||||
ATH11K_CAC_RUNNING,
|
||||
ATH11K_FLAG_CORE_REGISTERED,
|
||||
|
@ -607,9 +615,8 @@ struct ath11k {
|
|||
bool dfs_block_radar_events;
|
||||
struct ath11k_thermal thermal;
|
||||
u32 vdev_id_11d_scan;
|
||||
struct completion finish_11d_scan;
|
||||
struct completion finish_11d_ch_list;
|
||||
bool pending_11d;
|
||||
struct completion completed_11d_scan;
|
||||
enum ath11k_11d_state state_11d;
|
||||
bool regdom_set_by_user;
|
||||
int hw_rate_code;
|
||||
u8 twt_enabled;
|
||||
|
|
|
@ -3601,26 +3601,6 @@ static int ath11k_mac_op_hw_scan(struct ieee80211_hw *hw,
|
|||
if (ret)
|
||||
goto exit;
|
||||
|
||||
/* Currently the pending_11d=true only happened 1 time while
|
||||
* wlan interface up in ath11k_mac_11d_scan_start(), it is called by
|
||||
* ath11k_mac_op_add_interface(), after wlan interface up,
|
||||
* pending_11d=false always.
|
||||
* If remove below wait, it always happened scan fail and lead connect
|
||||
* fail while wlan interface up, because it has a 11d scan which is running
|
||||
* in firmware, and lead this scan failed.
|
||||
*/
|
||||
if (ar->pending_11d) {
|
||||
long time_left;
|
||||
unsigned long timeout = 5 * HZ;
|
||||
|
||||
if (ar->supports_6ghz)
|
||||
timeout += 5 * HZ;
|
||||
|
||||
time_left = wait_for_completion_timeout(&ar->finish_11d_ch_list, timeout);
|
||||
ath11k_dbg(ar->ab, ATH11K_DBG_MAC,
|
||||
"mac wait 11d channel list time left %ld\n", time_left);
|
||||
}
|
||||
|
||||
memset(&arg, 0, sizeof(arg));
|
||||
ath11k_wmi_start_scan_init(ar, &arg);
|
||||
arg.vdev_id = arvif->vdev_id;
|
||||
|
@ -3686,6 +3666,10 @@ exit:
|
|||
kfree(arg.extraie.ptr);
|
||||
|
||||
mutex_unlock(&ar->conf_mutex);
|
||||
|
||||
if (ar->state_11d == ATH11K_11D_PREPARING)
|
||||
ath11k_mac_11d_scan_start(ar, arvif->vdev_id);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@ -5814,7 +5798,7 @@ static int ath11k_mac_op_start(struct ieee80211_hw *hw)
|
|||
|
||||
/* TODO: Do we need to enable ANI? */
|
||||
|
||||
ath11k_reg_update_chan_list(ar);
|
||||
ath11k_reg_update_chan_list(ar, false);
|
||||
|
||||
ar->num_started_vdevs = 0;
|
||||
ar->num_created_vdevs = 0;
|
||||
|
@ -5881,6 +5865,11 @@ static void ath11k_mac_op_stop(struct ieee80211_hw *hw)
|
|||
cancel_work_sync(&ar->ab->update_11d_work);
|
||||
cancel_work_sync(&ar->ab->rfkill_work);
|
||||
|
||||
if (ar->state_11d == ATH11K_11D_PREPARING) {
|
||||
ar->state_11d = ATH11K_11D_IDLE;
|
||||
complete(&ar->completed_11d_scan);
|
||||
}
|
||||
|
||||
spin_lock_bh(&ar->data_lock);
|
||||
list_for_each_entry_safe(ppdu_stats, tmp, &ar->ppdu_stats_info, list) {
|
||||
list_del(&ppdu_stats->list);
|
||||
|
@ -6051,7 +6040,7 @@ static bool ath11k_mac_vif_ap_active_any(struct ath11k_base *ab)
|
|||
return false;
|
||||
}
|
||||
|
||||
void ath11k_mac_11d_scan_start(struct ath11k *ar, u32 vdev_id, bool wait)
|
||||
void ath11k_mac_11d_scan_start(struct ath11k *ar, u32 vdev_id)
|
||||
{
|
||||
struct wmi_11d_scan_start_params param;
|
||||
int ret;
|
||||
|
@ -6079,28 +6068,22 @@ void ath11k_mac_11d_scan_start(struct ath11k *ar, u32 vdev_id, bool wait)
|
|||
|
||||
ath11k_dbg(ar->ab, ATH11K_DBG_MAC, "mac start 11d scan\n");
|
||||
|
||||
if (wait)
|
||||
reinit_completion(&ar->finish_11d_scan);
|
||||
|
||||
ret = ath11k_wmi_send_11d_scan_start_cmd(ar, ¶m);
|
||||
if (ret) {
|
||||
ath11k_warn(ar->ab, "failed to start 11d scan vdev %d ret: %d\n",
|
||||
vdev_id, ret);
|
||||
} else {
|
||||
ar->vdev_id_11d_scan = vdev_id;
|
||||
if (wait) {
|
||||
ar->pending_11d = true;
|
||||
ret = wait_for_completion_timeout(&ar->finish_11d_scan,
|
||||
5 * HZ);
|
||||
ath11k_dbg(ar->ab, ATH11K_DBG_MAC,
|
||||
"mac 11d scan left time %d\n", ret);
|
||||
|
||||
if (!ret)
|
||||
ar->pending_11d = false;
|
||||
}
|
||||
if (ar->state_11d == ATH11K_11D_PREPARING)
|
||||
ar->state_11d = ATH11K_11D_RUNNING;
|
||||
}
|
||||
|
||||
fin:
|
||||
if (ar->state_11d == ATH11K_11D_PREPARING) {
|
||||
ar->state_11d = ATH11K_11D_IDLE;
|
||||
complete(&ar->completed_11d_scan);
|
||||
}
|
||||
|
||||
mutex_unlock(&ar->ab->vdev_id_11d_lock);
|
||||
}
|
||||
|
||||
|
@ -6123,12 +6106,15 @@ void ath11k_mac_11d_scan_stop(struct ath11k *ar)
|
|||
vdev_id = ar->vdev_id_11d_scan;
|
||||
|
||||
ret = ath11k_wmi_send_11d_scan_stop_cmd(ar, vdev_id);
|
||||
if (ret)
|
||||
if (ret) {
|
||||
ath11k_warn(ar->ab,
|
||||
"failed to stopt 11d scan vdev %d ret: %d\n",
|
||||
vdev_id, ret);
|
||||
else
|
||||
} else {
|
||||
ar->vdev_id_11d_scan = ATH11K_11D_INVALID_VDEV_ID;
|
||||
ar->state_11d = ATH11K_11D_IDLE;
|
||||
complete(&ar->completed_11d_scan);
|
||||
}
|
||||
}
|
||||
mutex_unlock(&ar->ab->vdev_id_11d_lock);
|
||||
}
|
||||
|
@ -6324,8 +6310,10 @@ static int ath11k_mac_op_add_interface(struct ieee80211_hw *hw,
|
|||
goto err_peer_del;
|
||||
}
|
||||
|
||||
ath11k_mac_11d_scan_start(ar, arvif->vdev_id, true);
|
||||
|
||||
if (test_bit(WMI_TLV_SERVICE_11D_OFFLOAD, ab->wmi_ab.svc_map)) {
|
||||
reinit_completion(&ar->completed_11d_scan);
|
||||
ar->state_11d = ATH11K_11D_PREPARING;
|
||||
}
|
||||
break;
|
||||
case WMI_VDEV_TYPE_MONITOR:
|
||||
set_bit(ATH11K_FLAG_MONITOR_VDEV_CREATED, &ar->monitor_flags);
|
||||
|
@ -7190,7 +7178,7 @@ ath11k_mac_op_unassign_vif_chanctx(struct ieee80211_hw *hw,
|
|||
}
|
||||
|
||||
if (arvif->vdev_type == WMI_VDEV_TYPE_STA)
|
||||
ath11k_mac_11d_scan_start(ar, arvif->vdev_id, false);
|
||||
ath11k_mac_11d_scan_start(ar, arvif->vdev_id);
|
||||
|
||||
mutex_unlock(&ar->conf_mutex);
|
||||
}
|
||||
|
@ -8671,8 +8659,7 @@ int ath11k_mac_allocate(struct ath11k_base *ab)
|
|||
ar->monitor_vdev_id = -1;
|
||||
clear_bit(ATH11K_FLAG_MONITOR_VDEV_CREATED, &ar->monitor_flags);
|
||||
ar->vdev_id_11d_scan = ATH11K_11D_INVALID_VDEV_ID;
|
||||
init_completion(&ar->finish_11d_scan);
|
||||
init_completion(&ar->finish_11d_ch_list);
|
||||
init_completion(&ar->completed_11d_scan);
|
||||
}
|
||||
|
||||
return 0;
|
||||
|
|
|
@ -130,7 +130,7 @@ extern const struct htt_rx_ring_tlv_filter ath11k_mac_mon_status_filter_default;
|
|||
#define ATH11K_SCAN_11D_INTERVAL 600000
|
||||
#define ATH11K_11D_INVALID_VDEV_ID 0xFFFF
|
||||
|
||||
void ath11k_mac_11d_scan_start(struct ath11k *ar, u32 vdev_id, bool wait);
|
||||
void ath11k_mac_11d_scan_start(struct ath11k *ar, u32 vdev_id);
|
||||
void ath11k_mac_11d_scan_stop(struct ath11k *ar);
|
||||
void ath11k_mac_11d_scan_stop_all(struct ath11k_base *ab);
|
||||
|
||||
|
|
|
@ -102,7 +102,7 @@ ath11k_reg_notifier(struct wiphy *wiphy, struct regulatory_request *request)
|
|||
ar->regdom_set_by_user = true;
|
||||
}
|
||||
|
||||
int ath11k_reg_update_chan_list(struct ath11k *ar)
|
||||
int ath11k_reg_update_chan_list(struct ath11k *ar, bool wait)
|
||||
{
|
||||
struct ieee80211_supported_band **bands;
|
||||
struct scan_chan_list_params *params;
|
||||
|
@ -111,7 +111,32 @@ int ath11k_reg_update_chan_list(struct ath11k *ar)
|
|||
struct channel_param *ch;
|
||||
enum nl80211_band band;
|
||||
int num_channels = 0;
|
||||
int i, ret;
|
||||
int i, ret, left;
|
||||
|
||||
if (wait && ar->state_11d != ATH11K_11D_IDLE) {
|
||||
left = wait_for_completion_timeout(&ar->completed_11d_scan,
|
||||
ATH11K_SCAN_TIMEOUT_HZ);
|
||||
if (!left) {
|
||||
ath11k_dbg(ar->ab, ATH11K_DBG_REG,
|
||||
"failed to receive 11d scan complete: timed out\n");
|
||||
ar->state_11d = ATH11K_11D_IDLE;
|
||||
}
|
||||
ath11k_dbg(ar->ab, ATH11K_DBG_REG,
|
||||
"reg 11d scan wait left time %d\n", left);
|
||||
}
|
||||
|
||||
if (wait &&
|
||||
(ar->scan.state == ATH11K_SCAN_STARTING ||
|
||||
ar->scan.state == ATH11K_SCAN_RUNNING)) {
|
||||
left = wait_for_completion_timeout(&ar->scan.completed,
|
||||
ATH11K_SCAN_TIMEOUT_HZ);
|
||||
if (!left)
|
||||
ath11k_dbg(ar->ab, ATH11K_DBG_REG,
|
||||
"failed to receive hw scan complete: timed out\n");
|
||||
|
||||
ath11k_dbg(ar->ab, ATH11K_DBG_REG,
|
||||
"reg hw scan wait left time %d\n", left);
|
||||
}
|
||||
|
||||
bands = hw->wiphy->bands;
|
||||
for (band = 0; band < NUM_NL80211_BANDS; band++) {
|
||||
|
@ -193,11 +218,6 @@ int ath11k_reg_update_chan_list(struct ath11k *ar)
|
|||
ret = ath11k_wmi_send_scan_chan_list_cmd(ar, params);
|
||||
kfree(params);
|
||||
|
||||
if (ar->pending_11d) {
|
||||
complete(&ar->finish_11d_ch_list);
|
||||
ar->pending_11d = false;
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@ -263,15 +283,8 @@ int ath11k_regd_update(struct ath11k *ar)
|
|||
goto err;
|
||||
}
|
||||
|
||||
if (ar->pending_11d)
|
||||
complete(&ar->finish_11d_scan);
|
||||
|
||||
rtnl_lock();
|
||||
wiphy_lock(ar->hw->wiphy);
|
||||
|
||||
if (ar->pending_11d)
|
||||
reinit_completion(&ar->finish_11d_ch_list);
|
||||
|
||||
ret = regulatory_set_wiphy_regd_sync(ar->hw->wiphy, regd_copy);
|
||||
wiphy_unlock(ar->hw->wiphy);
|
||||
rtnl_unlock();
|
||||
|
@ -282,7 +295,7 @@ int ath11k_regd_update(struct ath11k *ar)
|
|||
goto err;
|
||||
|
||||
if (ar->state == ATH11K_STATE_ON) {
|
||||
ret = ath11k_reg_update_chan_list(ar);
|
||||
ret = ath11k_reg_update_chan_list(ar, true);
|
||||
if (ret)
|
||||
goto err;
|
||||
}
|
||||
|
|
|
@ -32,5 +32,5 @@ struct ieee80211_regdomain *
|
|||
ath11k_reg_build_regd(struct ath11k_base *ab,
|
||||
struct cur_regulatory_info *reg_info, bool intersect);
|
||||
int ath11k_regd_update(struct ath11k *ar);
|
||||
int ath11k_reg_update_chan_list(struct ath11k *ar);
|
||||
int ath11k_reg_update_chan_list(struct ath11k *ar, bool wait);
|
||||
#endif
|
||||
|
|
|
@ -2015,7 +2015,10 @@ void ath11k_wmi_start_scan_init(struct ath11k *ar,
|
|||
{
|
||||
/* setup commonly used values */
|
||||
arg->scan_req_id = 1;
|
||||
arg->scan_priority = WMI_SCAN_PRIORITY_LOW;
|
||||
if (ar->state_11d == ATH11K_11D_PREPARING)
|
||||
arg->scan_priority = WMI_SCAN_PRIORITY_MEDIUM;
|
||||
else
|
||||
arg->scan_priority = WMI_SCAN_PRIORITY_LOW;
|
||||
arg->dwell_time_active = 50;
|
||||
arg->dwell_time_active_2g = 0;
|
||||
arg->dwell_time_passive = 150;
|
||||
|
@ -6350,8 +6353,10 @@ static void ath11k_wmi_op_ep_tx_credits(struct ath11k_base *ab)
|
|||
static int ath11k_reg_11d_new_cc_event(struct ath11k_base *ab, struct sk_buff *skb)
|
||||
{
|
||||
const struct wmi_11d_new_cc_ev *ev;
|
||||
struct ath11k *ar;
|
||||
struct ath11k_pdev *pdev;
|
||||
const void **tb;
|
||||
int ret;
|
||||
int ret, i;
|
||||
|
||||
tb = ath11k_wmi_tlv_parse_alloc(ab, skb->data, skb->len, GFP_ATOMIC);
|
||||
if (IS_ERR(tb)) {
|
||||
|
@ -6377,6 +6382,13 @@ static int ath11k_reg_11d_new_cc_event(struct ath11k_base *ab, struct sk_buff *s
|
|||
|
||||
kfree(tb);
|
||||
|
||||
for (i = 0; i < ab->num_radios; i++) {
|
||||
pdev = &ab->pdevs[i];
|
||||
ar = pdev->ar;
|
||||
ar->state_11d = ATH11K_11D_IDLE;
|
||||
complete(&ar->completed_11d_scan);
|
||||
}
|
||||
|
||||
queue_work(ab->workqueue, &ab->update_11d_work);
|
||||
|
||||
return 0;
|
||||
|
|
|
@ -371,7 +371,7 @@ void iwl_dbg_tlv_del_timers(struct iwl_trans *trans)
|
|||
struct iwl_dbg_tlv_timer_node *node, *tmp;
|
||||
|
||||
list_for_each_entry_safe(node, tmp, timer_list, list) {
|
||||
del_timer(&node->timer);
|
||||
del_timer_sync(&node->timer);
|
||||
list_del(&node->list);
|
||||
kfree(node);
|
||||
}
|
||||
|
|
|
@ -2202,11 +2202,14 @@ mac80211_hwsim_sta_rc_update(struct ieee80211_hw *hw,
|
|||
if (!data->use_chanctx) {
|
||||
confbw = data->bw;
|
||||
} else {
|
||||
struct ieee80211_chanctx_conf *chanctx_conf =
|
||||
rcu_dereference(vif->chanctx_conf);
|
||||
struct ieee80211_chanctx_conf *chanctx_conf;
|
||||
|
||||
rcu_read_lock();
|
||||
chanctx_conf = rcu_dereference(vif->chanctx_conf);
|
||||
|
||||
if (!WARN_ON(!chanctx_conf))
|
||||
confbw = chanctx_conf->def.width;
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
WARN(bw > hwsim_get_chanwidth(confbw),
|
||||
|
@ -2475,11 +2478,13 @@ static void hw_scan_work(struct work_struct *work)
|
|||
if (req->ie_len)
|
||||
skb_put_data(probe, req->ie, req->ie_len);
|
||||
|
||||
rcu_read_lock();
|
||||
if (!ieee80211_tx_prepare_skb(hwsim->hw,
|
||||
hwsim->hw_scan_vif,
|
||||
probe,
|
||||
hwsim->tmp_chan->band,
|
||||
NULL)) {
|
||||
rcu_read_unlock();
|
||||
kfree_skb(probe);
|
||||
continue;
|
||||
}
|
||||
|
@ -2487,6 +2492,7 @@ static void hw_scan_work(struct work_struct *work)
|
|||
local_bh_disable();
|
||||
mac80211_hwsim_tx_frame(hwsim->hw, probe,
|
||||
hwsim->tmp_chan);
|
||||
rcu_read_unlock();
|
||||
local_bh_enable();
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1557,7 +1557,7 @@ ptp_ocp_signal_set(struct ptp_ocp *bp, int gen, struct ptp_ocp_signal *s)
|
|||
start_ns = ktime_set(ts.tv_sec, ts.tv_nsec) + NSEC_PER_MSEC;
|
||||
if (!s->start) {
|
||||
/* roundup() does not work on 32-bit systems */
|
||||
s->start = DIV_ROUND_UP_ULL(start_ns, s->period);
|
||||
s->start = DIV64_U64_ROUND_UP(start_ns, s->period);
|
||||
s->start = ktime_add(s->start, s->phase);
|
||||
}
|
||||
|
||||
|
|
|
@ -626,8 +626,6 @@ static void mpc_rcvd_sweep_resp(struct mpcg_info *mpcginfo)
|
|||
ctcm_clear_busy_do(dev);
|
||||
}
|
||||
|
||||
kfree(mpcginfo);
|
||||
|
||||
return;
|
||||
|
||||
}
|
||||
|
@ -1192,10 +1190,10 @@ static void ctcmpc_unpack_skb(struct channel *ch, struct sk_buff *pskb)
|
|||
CTCM_FUNTAIL, dev->name);
|
||||
priv->stats.rx_dropped++;
|
||||
/* mpcginfo only used for non-data transfers */
|
||||
kfree(mpcginfo);
|
||||
if (do_debug_data)
|
||||
ctcmpc_dump_skb(pskb, -8);
|
||||
}
|
||||
kfree(mpcginfo);
|
||||
}
|
||||
done:
|
||||
|
||||
|
@ -1977,7 +1975,6 @@ static void mpc_action_rcvd_xid0(fsm_instance *fsm, int event, void *arg)
|
|||
}
|
||||
break;
|
||||
}
|
||||
kfree(mpcginfo);
|
||||
|
||||
CTCM_PR_DEBUG("ctcmpc:%s() %s xid2:%i xid7:%i xidt_p2:%i \n",
|
||||
__func__, ch->id, grp->outstanding_xid2,
|
||||
|
@ -2038,7 +2035,6 @@ static void mpc_action_rcvd_xid7(fsm_instance *fsm, int event, void *arg)
|
|||
mpc_validate_xid(mpcginfo);
|
||||
break;
|
||||
}
|
||||
kfree(mpcginfo);
|
||||
return;
|
||||
}
|
||||
|
||||
|
|
|
@ -39,11 +39,12 @@ static ssize_t ctcm_buffer_write(struct device *dev,
|
|||
struct ctcm_priv *priv = dev_get_drvdata(dev);
|
||||
int rc;
|
||||
|
||||
ndev = priv->channel[CTCM_READ]->netdev;
|
||||
if (!(priv && priv->channel[CTCM_READ] && ndev)) {
|
||||
if (!(priv && priv->channel[CTCM_READ] &&
|
||||
priv->channel[CTCM_READ]->netdev)) {
|
||||
CTCM_DBF_TEXT(SETUP, CTC_DBF_ERROR, "bfnondev");
|
||||
return -ENODEV;
|
||||
}
|
||||
ndev = priv->channel[CTCM_READ]->netdev;
|
||||
|
||||
rc = kstrtouint(buf, 0, &bs1);
|
||||
if (rc)
|
||||
|
|
|
@ -1736,10 +1736,11 @@ lcs_get_control(struct lcs_card *card, struct lcs_cmd *cmd)
|
|||
lcs_schedule_recovery(card);
|
||||
break;
|
||||
case LCS_CMD_STOPLAN:
|
||||
pr_warn("Stoplan for %s initiated by LGW\n",
|
||||
card->dev->name);
|
||||
if (card->dev)
|
||||
if (card->dev) {
|
||||
pr_warn("Stoplan for %s initiated by LGW\n",
|
||||
card->dev->name);
|
||||
netif_carrier_off(card->dev);
|
||||
}
|
||||
break;
|
||||
default:
|
||||
LCS_DBF_TEXT(5, trace, "noLGWcmd");
|
||||
|
|
|
@ -169,7 +169,7 @@ enum {
|
|||
#define NETIF_F_HW_HSR_FWD __NETIF_F(HW_HSR_FWD)
|
||||
#define NETIF_F_HW_HSR_DUP __NETIF_F(HW_HSR_DUP)
|
||||
|
||||
/* Finds the next feature with the highest number of the range of start till 0.
|
||||
/* Finds the next feature with the highest number of the range of start-1 till 0.
|
||||
*/
|
||||
static inline int find_next_netdev_feature(u64 feature, unsigned long start)
|
||||
{
|
||||
|
@ -188,7 +188,7 @@ static inline int find_next_netdev_feature(u64 feature, unsigned long start)
|
|||
for ((bit) = find_next_netdev_feature((mask_addr), \
|
||||
NETDEV_FEATURE_COUNT); \
|
||||
(bit) >= 0; \
|
||||
(bit) = find_next_netdev_feature((mask_addr), (bit) - 1))
|
||||
(bit) = find_next_netdev_feature((mask_addr), (bit)))
|
||||
|
||||
/* Features valid for ethtool to change */
|
||||
/* = all defined minus driver/device-class-related */
|
||||
|
|
|
@ -36,6 +36,9 @@
|
|||
/* HCI priority */
|
||||
#define HCI_PRIO_MAX 7
|
||||
|
||||
/* HCI maximum id value */
|
||||
#define HCI_MAX_ID 10000
|
||||
|
||||
/* HCI Core structures */
|
||||
struct inquiry_data {
|
||||
bdaddr_t bdaddr;
|
||||
|
|
|
@ -14,6 +14,7 @@ struct tcf_pedit {
|
|||
struct tc_action common;
|
||||
unsigned char tcfp_nkeys;
|
||||
unsigned char tcfp_flags;
|
||||
u32 tcfp_off_max_hint;
|
||||
struct tc_pedit_key *tcfp_keys;
|
||||
struct tcf_pedit_key_ex *tcfp_keys_ex;
|
||||
};
|
||||
|
|
|
@ -681,7 +681,6 @@ struct ocelot_vcap_id {
|
|||
|
||||
struct ocelot_vcap_filter {
|
||||
struct list_head list;
|
||||
struct list_head trap_list;
|
||||
|
||||
enum ocelot_vcap_filter_type type;
|
||||
int block_id;
|
||||
|
@ -695,6 +694,7 @@ struct ocelot_vcap_filter {
|
|||
struct ocelot_vcap_stats stats;
|
||||
/* For VCAP IS1 and IS2 */
|
||||
bool take_ts;
|
||||
bool is_trap;
|
||||
unsigned long ingress_port_mask;
|
||||
/* For VCAP ES0 */
|
||||
struct ocelot_vcap_port ingress_port;
|
||||
|
|
|
@ -184,7 +184,7 @@ struct rfkill_event_ext {
|
|||
#define RFKILL_IOC_NOINPUT 1
|
||||
#define RFKILL_IOCTL_NOINPUT _IO(RFKILL_IOC_MAGIC, RFKILL_IOC_NOINPUT)
|
||||
#define RFKILL_IOC_MAX_SIZE 2
|
||||
#define RFKILL_IOCTL_MAX_SIZE _IOW(RFKILL_IOC_MAGIC, RFKILL_IOC_EXT_SIZE, __u32)
|
||||
#define RFKILL_IOCTL_MAX_SIZE _IOW(RFKILL_IOC_MAGIC, RFKILL_IOC_MAX_SIZE, __u32)
|
||||
|
||||
/* and that's all userspace gets */
|
||||
|
||||
|
|
|
@ -12,41 +12,41 @@
|
|||
* Each profile size must be of NET_DIM_PARAMS_NUM_PROFILES
|
||||
*/
|
||||
#define NET_DIM_PARAMS_NUM_PROFILES 5
|
||||
#define NET_DIM_DEFAULT_RX_CQ_MODERATION_PKTS_FROM_EQE 256
|
||||
#define NET_DIM_DEFAULT_TX_CQ_MODERATION_PKTS_FROM_EQE 128
|
||||
#define NET_DIM_DEFAULT_RX_CQ_PKTS_FROM_EQE 256
|
||||
#define NET_DIM_DEFAULT_TX_CQ_PKTS_FROM_EQE 128
|
||||
#define NET_DIM_DEF_PROFILE_CQE 1
|
||||
#define NET_DIM_DEF_PROFILE_EQE 1
|
||||
|
||||
#define NET_DIM_RX_EQE_PROFILES { \
|
||||
{1, NET_DIM_DEFAULT_RX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{8, NET_DIM_DEFAULT_RX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{64, NET_DIM_DEFAULT_RX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{128, NET_DIM_DEFAULT_RX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{256, NET_DIM_DEFAULT_RX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{.usec = 1, .pkts = NET_DIM_DEFAULT_RX_CQ_PKTS_FROM_EQE,}, \
|
||||
{.usec = 8, .pkts = NET_DIM_DEFAULT_RX_CQ_PKTS_FROM_EQE,}, \
|
||||
{.usec = 64, .pkts = NET_DIM_DEFAULT_RX_CQ_PKTS_FROM_EQE,}, \
|
||||
{.usec = 128, .pkts = NET_DIM_DEFAULT_RX_CQ_PKTS_FROM_EQE,}, \
|
||||
{.usec = 256, .pkts = NET_DIM_DEFAULT_RX_CQ_PKTS_FROM_EQE,} \
|
||||
}
|
||||
|
||||
#define NET_DIM_RX_CQE_PROFILES { \
|
||||
{2, 256}, \
|
||||
{8, 128}, \
|
||||
{16, 64}, \
|
||||
{32, 64}, \
|
||||
{64, 64} \
|
||||
{.usec = 2, .pkts = 256,}, \
|
||||
{.usec = 8, .pkts = 128,}, \
|
||||
{.usec = 16, .pkts = 64,}, \
|
||||
{.usec = 32, .pkts = 64,}, \
|
||||
{.usec = 64, .pkts = 64,} \
|
||||
}
|
||||
|
||||
#define NET_DIM_TX_EQE_PROFILES { \
|
||||
{1, NET_DIM_DEFAULT_TX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{8, NET_DIM_DEFAULT_TX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{32, NET_DIM_DEFAULT_TX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{64, NET_DIM_DEFAULT_TX_CQ_MODERATION_PKTS_FROM_EQE}, \
|
||||
{128, NET_DIM_DEFAULT_TX_CQ_MODERATION_PKTS_FROM_EQE} \
|
||||
{.usec = 1, .pkts = NET_DIM_DEFAULT_TX_CQ_PKTS_FROM_EQE,}, \
|
||||
{.usec = 8, .pkts = NET_DIM_DEFAULT_TX_CQ_PKTS_FROM_EQE,}, \
|
||||
{.usec = 32, .pkts = NET_DIM_DEFAULT_TX_CQ_PKTS_FROM_EQE,}, \
|
||||
{.usec = 64, .pkts = NET_DIM_DEFAULT_TX_CQ_PKTS_FROM_EQE,}, \
|
||||
{.usec = 128, .pkts = NET_DIM_DEFAULT_TX_CQ_PKTS_FROM_EQE,} \
|
||||
}
|
||||
|
||||
#define NET_DIM_TX_CQE_PROFILES { \
|
||||
{5, 128}, \
|
||||
{8, 64}, \
|
||||
{16, 32}, \
|
||||
{32, 32}, \
|
||||
{64, 32} \
|
||||
{.usec = 5, .pkts = 128,}, \
|
||||
{.usec = 8, .pkts = 64,}, \
|
||||
{.usec = 16, .pkts = 32,}, \
|
||||
{.usec = 32, .pkts = 32,}, \
|
||||
{.usec = 64, .pkts = 32,} \
|
||||
}
|
||||
|
||||
static const struct dim_cq_moder
|
||||
|
|
|
@ -475,6 +475,17 @@ int batadv_frag_send_packet(struct sk_buff *skb,
|
|||
goto free_skb;
|
||||
}
|
||||
|
||||
/* GRO might have added fragments to the fragment list instead of
|
||||
* frags[]. But this is not handled by skb_split and must be
|
||||
* linearized to avoid incorrect length information after all
|
||||
* batman-adv fragments were created and submitted to the
|
||||
* hard-interface
|
||||
*/
|
||||
if (skb_has_frag_list(skb) && __skb_linearize(skb)) {
|
||||
ret = -ENOMEM;
|
||||
goto free_skb;
|
||||
}
|
||||
|
||||
/* Create one header to be copied to all fragments */
|
||||
frag_header.packet_type = BATADV_UNICAST_FRAG;
|
||||
frag_header.version = BATADV_COMPAT_VERSION;
|
||||
|
|
|
@ -2555,10 +2555,10 @@ int hci_register_dev(struct hci_dev *hdev)
|
|||
*/
|
||||
switch (hdev->dev_type) {
|
||||
case HCI_PRIMARY:
|
||||
id = ida_simple_get(&hci_index_ida, 0, 0, GFP_KERNEL);
|
||||
id = ida_simple_get(&hci_index_ida, 0, HCI_MAX_ID, GFP_KERNEL);
|
||||
break;
|
||||
case HCI_AMP:
|
||||
id = ida_simple_get(&hci_index_ida, 1, 0, GFP_KERNEL);
|
||||
id = ida_simple_get(&hci_index_ida, 1, HCI_MAX_ID, GFP_KERNEL);
|
||||
break;
|
||||
default:
|
||||
return -EINVAL;
|
||||
|
@ -2567,7 +2567,7 @@ int hci_register_dev(struct hci_dev *hdev)
|
|||
if (id < 0)
|
||||
return id;
|
||||
|
||||
sprintf(hdev->name, "hci%d", id);
|
||||
snprintf(hdev->name, sizeof(hdev->name), "hci%d", id);
|
||||
hdev->id = id;
|
||||
|
||||
BT_DBG("%p name %s bus %d", hdev, hdev->name, hdev->bus);
|
||||
|
|
|
@ -3897,7 +3897,7 @@ struct sk_buff *skb_segment_list(struct sk_buff *skb,
|
|||
unsigned int delta_len = 0;
|
||||
struct sk_buff *tail = NULL;
|
||||
struct sk_buff *nskb, *tmp;
|
||||
int err;
|
||||
int len_diff, err;
|
||||
|
||||
skb_push(skb, -skb_network_offset(skb) + offset);
|
||||
|
||||
|
@ -3937,9 +3937,11 @@ struct sk_buff *skb_segment_list(struct sk_buff *skb,
|
|||
skb_push(nskb, -skb_network_offset(nskb) + offset);
|
||||
|
||||
skb_release_head_state(nskb);
|
||||
len_diff = skb_network_header_len(nskb) - skb_network_header_len(skb);
|
||||
__copy_skb_header(nskb, skb);
|
||||
|
||||
skb_headers_offset_update(nskb, skb_headroom(nskb) - skb_headroom(skb));
|
||||
nskb->transport_header += len_diff;
|
||||
skb_copy_from_linear_data_offset(skb, -tnl_hlen,
|
||||
nskb->data - tnl_hlen,
|
||||
offset + tnl_hlen);
|
||||
|
|
|
@ -854,7 +854,7 @@ static void dn_send_endnode_hello(struct net_device *dev, struct dn_ifaddr *ifa)
|
|||
memcpy(msg->neighbor, dn_hiord, ETH_ALEN);
|
||||
|
||||
if (dn_db->router) {
|
||||
struct dn_neigh *dn = (struct dn_neigh *)dn_db->router;
|
||||
struct dn_neigh *dn = container_of(dn_db->router, struct dn_neigh, n);
|
||||
dn_dn2eth(msg->neighbor, dn->addr);
|
||||
}
|
||||
|
||||
|
@ -902,7 +902,7 @@ static void dn_send_router_hello(struct net_device *dev, struct dn_ifaddr *ifa)
|
|||
{
|
||||
int n;
|
||||
struct dn_dev *dn_db = rcu_dereference_raw(dev->dn_ptr);
|
||||
struct dn_neigh *dn = (struct dn_neigh *)dn_db->router;
|
||||
struct dn_neigh *dn = container_of(dn_db->router, struct dn_neigh, n);
|
||||
struct sk_buff *skb;
|
||||
size_t size;
|
||||
unsigned char *ptr;
|
||||
|
|
|
@ -426,7 +426,8 @@ int dn_neigh_router_hello(struct net *net, struct sock *sk, struct sk_buff *skb)
|
|||
if (!dn_db->router) {
|
||||
dn_db->router = neigh_clone(neigh);
|
||||
} else {
|
||||
if (msg->priority > ((struct dn_neigh *)dn_db->router)->priority)
|
||||
if (msg->priority > container_of(dn_db->router,
|
||||
struct dn_neigh, n)->priority)
|
||||
neigh_release(xchg(&dn_db->router, neigh_clone(neigh)));
|
||||
}
|
||||
}
|
||||
|
|
|
@ -1120,7 +1120,7 @@ source_ok:
|
|||
/* Ok then, we assume its directly connected and move on */
|
||||
select_source:
|
||||
if (neigh)
|
||||
gateway = ((struct dn_neigh *)neigh)->addr;
|
||||
gateway = container_of(neigh, struct dn_neigh, n)->addr;
|
||||
if (gateway == 0)
|
||||
gateway = fld.daddr;
|
||||
if (fld.saddr == 0) {
|
||||
|
@ -1429,7 +1429,7 @@ static int dn_route_input_slow(struct sk_buff *skb)
|
|||
/* Use the default router if there is one */
|
||||
neigh = neigh_clone(dn_db->router);
|
||||
if (neigh) {
|
||||
gateway = ((struct dn_neigh *)neigh)->addr;
|
||||
gateway = container_of(neigh, struct dn_neigh, n)->addr;
|
||||
goto make_route;
|
||||
}
|
||||
|
||||
|
|
|
@ -451,6 +451,7 @@ out_rollback_unoffload:
|
|||
switchdev_bridge_port_unoffload(brport_dev, dp,
|
||||
&dsa_slave_switchdev_notifier,
|
||||
&dsa_slave_switchdev_blocking_notifier);
|
||||
dsa_flush_workqueue();
|
||||
out_rollback_unbridge:
|
||||
dsa_broadcast(DSA_NOTIFIER_BRIDGE_LEAVE, &info);
|
||||
out_rollback:
|
||||
|
|
|
@ -305,6 +305,7 @@ static int ping_check_bind_addr(struct sock *sk, struct inet_sock *isk,
|
|||
struct net *net = sock_net(sk);
|
||||
if (sk->sk_family == AF_INET) {
|
||||
struct sockaddr_in *addr = (struct sockaddr_in *) uaddr;
|
||||
u32 tb_id = RT_TABLE_LOCAL;
|
||||
int chk_addr_ret;
|
||||
|
||||
if (addr_len < sizeof(*addr))
|
||||
|
@ -318,7 +319,8 @@ static int ping_check_bind_addr(struct sock *sk, struct inet_sock *isk,
|
|||
pr_debug("ping_check_bind_addr(sk=%p,addr=%pI4,port=%d)\n",
|
||||
sk, &addr->sin_addr.s_addr, ntohs(addr->sin_port));
|
||||
|
||||
chk_addr_ret = inet_addr_type(net, addr->sin_addr.s_addr);
|
||||
tb_id = l3mdev_fib_table_by_index(net, sk->sk_bound_dev_if) ? : tb_id;
|
||||
chk_addr_ret = inet_addr_type_table(net, addr->sin_addr.s_addr, tb_id);
|
||||
|
||||
if (!inet_addr_valid_or_nonlocal(net, inet_sk(sk),
|
||||
addr->sin_addr.s_addr,
|
||||
|
@ -355,6 +357,14 @@ static int ping_check_bind_addr(struct sock *sk, struct inet_sock *isk,
|
|||
return -ENODEV;
|
||||
}
|
||||
}
|
||||
|
||||
if (!dev && sk->sk_bound_dev_if) {
|
||||
dev = dev_get_by_index_rcu(net, sk->sk_bound_dev_if);
|
||||
if (!dev) {
|
||||
rcu_read_unlock();
|
||||
return -ENODEV;
|
||||
}
|
||||
}
|
||||
has_addr = pingv6_ops.ipv6_chk_addr(net, &addr->sin6_addr, dev,
|
||||
scoped);
|
||||
rcu_read_unlock();
|
||||
|
|
|
@ -1753,6 +1753,7 @@ static int ip_route_input_mc(struct sk_buff *skb, __be32 daddr, __be32 saddr,
|
|||
#endif
|
||||
RT_CACHE_STAT_INC(in_slow_mc);
|
||||
|
||||
skb_dst_drop(skb);
|
||||
skb_dst_set(skb, &rth->dst);
|
||||
return 0;
|
||||
}
|
||||
|
|
|
@ -3657,6 +3657,12 @@ static bool ieee80211_assoc_success(struct ieee80211_sub_if_data *sdata,
|
|||
cbss->transmitted_bss->bssid);
|
||||
bss_conf->bssid_indicator = cbss->max_bssid_indicator;
|
||||
bss_conf->bssid_index = cbss->bssid_index;
|
||||
} else {
|
||||
bss_conf->nontransmitted = false;
|
||||
memset(bss_conf->transmitter_bssid, 0,
|
||||
sizeof(bss_conf->transmitter_bssid));
|
||||
bss_conf->bssid_indicator = 0;
|
||||
bss_conf->bssid_index = 0;
|
||||
}
|
||||
|
||||
/*
|
||||
|
|
|
@ -1405,8 +1405,7 @@ static void ieee80211_rx_reorder_ampdu(struct ieee80211_rx_data *rx,
|
|||
goto dont_reorder;
|
||||
|
||||
/* not part of a BA session */
|
||||
if (ack_policy != IEEE80211_QOS_CTL_ACK_POLICY_BLOCKACK &&
|
||||
ack_policy != IEEE80211_QOS_CTL_ACK_POLICY_NORMAL)
|
||||
if (ack_policy == IEEE80211_QOS_CTL_ACK_POLICY_NOACK)
|
||||
goto dont_reorder;
|
||||
|
||||
/* new, potentially un-ordered, ampdu frame - process it */
|
||||
|
|
|
@ -1975,7 +1975,6 @@ static int netlink_recvmsg(struct socket *sock, struct msghdr *msg, size_t len,
|
|||
copied = len;
|
||||
}
|
||||
|
||||
skb_reset_transport_header(data_skb);
|
||||
err = skb_copy_datagram_msg(data_skb, 0, msg, copied);
|
||||
|
||||
if (msg->msg_name) {
|
||||
|
|
|
@ -487,11 +487,11 @@ struct rds_tcp_net {
|
|||
/* All module specific customizations to the RDS-TCP socket should be done in
|
||||
* rds_tcp_tune() and applied after socket creation.
|
||||
*/
|
||||
void rds_tcp_tune(struct socket *sock)
|
||||
bool rds_tcp_tune(struct socket *sock)
|
||||
{
|
||||
struct sock *sk = sock->sk;
|
||||
struct net *net = sock_net(sk);
|
||||
struct rds_tcp_net *rtn = net_generic(net, rds_tcp_netid);
|
||||
struct rds_tcp_net *rtn;
|
||||
|
||||
tcp_sock_set_nodelay(sock->sk);
|
||||
lock_sock(sk);
|
||||
|
@ -499,10 +499,15 @@ void rds_tcp_tune(struct socket *sock)
|
|||
* a process which created this net namespace terminated.
|
||||
*/
|
||||
if (!sk->sk_net_refcnt) {
|
||||
if (!maybe_get_net(net)) {
|
||||
release_sock(sk);
|
||||
return false;
|
||||
}
|
||||
sk->sk_net_refcnt = 1;
|
||||
get_net_track(net, &sk->ns_tracker, GFP_KERNEL);
|
||||
netns_tracker_alloc(net, &sk->ns_tracker, GFP_KERNEL);
|
||||
sock_inuse_add(net, 1);
|
||||
}
|
||||
rtn = net_generic(net, rds_tcp_netid);
|
||||
if (rtn->sndbuf_size > 0) {
|
||||
sk->sk_sndbuf = rtn->sndbuf_size;
|
||||
sk->sk_userlocks |= SOCK_SNDBUF_LOCK;
|
||||
|
@ -512,6 +517,7 @@ void rds_tcp_tune(struct socket *sock)
|
|||
sk->sk_userlocks |= SOCK_RCVBUF_LOCK;
|
||||
}
|
||||
release_sock(sk);
|
||||
return true;
|
||||
}
|
||||
|
||||
static void rds_tcp_accept_worker(struct work_struct *work)
|
||||
|
|
|
@ -49,7 +49,7 @@ struct rds_tcp_statistics {
|
|||
};
|
||||
|
||||
/* tcp.c */
|
||||
void rds_tcp_tune(struct socket *sock);
|
||||
bool rds_tcp_tune(struct socket *sock);
|
||||
void rds_tcp_set_callbacks(struct socket *sock, struct rds_conn_path *cp);
|
||||
void rds_tcp_reset_callbacks(struct socket *sock, struct rds_conn_path *cp);
|
||||
void rds_tcp_restore_callbacks(struct socket *sock,
|
||||
|
|
|
@ -124,7 +124,10 @@ int rds_tcp_conn_path_connect(struct rds_conn_path *cp)
|
|||
if (ret < 0)
|
||||
goto out;
|
||||
|
||||
rds_tcp_tune(sock);
|
||||
if (!rds_tcp_tune(sock)) {
|
||||
ret = -EINVAL;
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (isv6) {
|
||||
sin6.sin6_family = AF_INET6;
|
||||
|
|
|
@ -133,7 +133,10 @@ int rds_tcp_accept_one(struct socket *sock)
|
|||
__module_get(new_sock->ops->owner);
|
||||
|
||||
rds_tcp_keepalive(new_sock);
|
||||
rds_tcp_tune(new_sock);
|
||||
if (!rds_tcp_tune(new_sock)) {
|
||||
ret = -EINVAL;
|
||||
goto out;
|
||||
}
|
||||
|
||||
inet = inet_sk(new_sock->sk);
|
||||
|
||||
|
|
|
@ -149,7 +149,7 @@ static int tcf_pedit_init(struct net *net, struct nlattr *nla,
|
|||
struct nlattr *pattr;
|
||||
struct tcf_pedit *p;
|
||||
int ret = 0, err;
|
||||
int ksize;
|
||||
int i, ksize;
|
||||
u32 index;
|
||||
|
||||
if (!nla) {
|
||||
|
@ -228,6 +228,18 @@ static int tcf_pedit_init(struct net *net, struct nlattr *nla,
|
|||
p->tcfp_nkeys = parm->nkeys;
|
||||
}
|
||||
memcpy(p->tcfp_keys, parm->keys, ksize);
|
||||
p->tcfp_off_max_hint = 0;
|
||||
for (i = 0; i < p->tcfp_nkeys; ++i) {
|
||||
u32 cur = p->tcfp_keys[i].off;
|
||||
|
||||
/* The AT option can read a single byte, we can bound the actual
|
||||
* value with uchar max.
|
||||
*/
|
||||
cur += (0xff & p->tcfp_keys[i].offmask) >> p->tcfp_keys[i].shift;
|
||||
|
||||
/* Each key touches 4 bytes starting from the computed offset */
|
||||
p->tcfp_off_max_hint = max(p->tcfp_off_max_hint, cur + 4);
|
||||
}
|
||||
|
||||
p->tcfp_flags = parm->flags;
|
||||
goto_ch = tcf_action_set_ctrlact(*a, parm->action, goto_ch);
|
||||
|
@ -308,13 +320,18 @@ static int tcf_pedit_act(struct sk_buff *skb, const struct tc_action *a,
|
|||
struct tcf_result *res)
|
||||
{
|
||||
struct tcf_pedit *p = to_pedit(a);
|
||||
u32 max_offset;
|
||||
int i;
|
||||
|
||||
if (skb_unclone(skb, GFP_ATOMIC))
|
||||
return p->tcf_action;
|
||||
|
||||
spin_lock(&p->tcf_lock);
|
||||
|
||||
max_offset = (skb_transport_header_was_set(skb) ?
|
||||
skb_transport_offset(skb) :
|
||||
skb_network_offset(skb)) +
|
||||
p->tcfp_off_max_hint;
|
||||
if (skb_ensure_writable(skb, min(skb->len, max_offset)))
|
||||
goto unlock;
|
||||
|
||||
tcf_lastuse_update(&p->tcf_tm);
|
||||
|
||||
if (p->tcfp_nkeys > 0) {
|
||||
|
@ -403,6 +420,7 @@ bad:
|
|||
p->tcf_qstats.overlimits++;
|
||||
done:
|
||||
bstats_update(&p->tcf_bstats, skb);
|
||||
unlock:
|
||||
spin_unlock(&p->tcf_lock);
|
||||
return p->tcf_action;
|
||||
}
|
||||
|
|
|
@ -355,12 +355,12 @@ int smc_rx_recvmsg(struct smc_sock *smc, struct msghdr *msg,
|
|||
}
|
||||
break;
|
||||
}
|
||||
if (!timeo)
|
||||
return -EAGAIN;
|
||||
if (signal_pending(current)) {
|
||||
read_done = sock_intr_errno(timeo);
|
||||
break;
|
||||
}
|
||||
if (!timeo)
|
||||
return -EAGAIN;
|
||||
}
|
||||
|
||||
if (!smc_rx_data_available(conn)) {
|
||||
|
|
|
@ -1347,7 +1347,10 @@ static int tls_device_down(struct net_device *netdev)
|
|||
|
||||
/* Device contexts for RX and TX will be freed in on sk_destruct
|
||||
* by tls_device_free_ctx. rx_conf and tx_conf stay in TLS_HW.
|
||||
* Now release the ref taken above.
|
||||
*/
|
||||
if (refcount_dec_and_test(&ctx->refcount))
|
||||
tls_device_free_ctx(ctx);
|
||||
}
|
||||
|
||||
up_write(&device_offload_lock);
|
||||
|
|
|
@ -3173,6 +3173,15 @@ int nl80211_parse_chandef(struct cfg80211_registered_device *rdev,
|
|||
} else if (attrs[NL80211_ATTR_CHANNEL_WIDTH]) {
|
||||
chandef->width =
|
||||
nla_get_u32(attrs[NL80211_ATTR_CHANNEL_WIDTH]);
|
||||
if (chandef->chan->band == NL80211_BAND_S1GHZ) {
|
||||
/* User input error for channel width doesn't match channel */
|
||||
if (chandef->width != ieee80211_s1g_channel_width(chandef->chan)) {
|
||||
NL_SET_ERR_MSG_ATTR(extack,
|
||||
attrs[NL80211_ATTR_CHANNEL_WIDTH],
|
||||
"bad channel width");
|
||||
return -EINVAL;
|
||||
}
|
||||
}
|
||||
if (attrs[NL80211_ATTR_CENTER_FREQ1]) {
|
||||
chandef->center_freq1 =
|
||||
nla_get_u32(attrs[NL80211_ATTR_CENTER_FREQ1]);
|
||||
|
@ -11657,18 +11666,23 @@ static int nl80211_set_tx_bitrate_mask(struct sk_buff *skb,
|
|||
struct cfg80211_bitrate_mask mask;
|
||||
struct cfg80211_registered_device *rdev = info->user_ptr[0];
|
||||
struct net_device *dev = info->user_ptr[1];
|
||||
struct wireless_dev *wdev = dev->ieee80211_ptr;
|
||||
int err;
|
||||
|
||||
if (!rdev->ops->set_bitrate_mask)
|
||||
return -EOPNOTSUPP;
|
||||
|
||||
wdev_lock(wdev);
|
||||
err = nl80211_parse_tx_bitrate_mask(info, info->attrs,
|
||||
NL80211_ATTR_TX_RATES, &mask,
|
||||
dev, true);
|
||||
if (err)
|
||||
return err;
|
||||
goto out;
|
||||
|
||||
return rdev_set_bitrate_mask(rdev, dev, NULL, &mask);
|
||||
err = rdev_set_bitrate_mask(rdev, dev, NULL, &mask);
|
||||
out:
|
||||
wdev_unlock(wdev);
|
||||
return err;
|
||||
}
|
||||
|
||||
static int nl80211_register_mgmt(struct sk_buff *skb, struct genl_info *info)
|
||||
|
|
|
@ -1829,7 +1829,7 @@ int cfg80211_get_ies_channel_number(const u8 *ie, size_t ielen,
|
|||
if (tmp && tmp->datalen >= sizeof(struct ieee80211_s1g_oper_ie)) {
|
||||
struct ieee80211_s1g_oper_ie *s1gop = (void *)tmp->data;
|
||||
|
||||
return s1gop->primary_ch;
|
||||
return s1gop->oper_ch;
|
||||
}
|
||||
} else {
|
||||
tmp = cfg80211_find_elem(WLAN_EID_DS_PARAMS, ie, ielen);
|
||||
|
|
|
@ -25,6 +25,7 @@ TEST_PROGS += bareudp.sh
|
|||
TEST_PROGS += amt.sh
|
||||
TEST_PROGS += unicast_extensions.sh
|
||||
TEST_PROGS += udpgro_fwd.sh
|
||||
TEST_PROGS += udpgro_frglist.sh
|
||||
TEST_PROGS += veth.sh
|
||||
TEST_PROGS += ioam6.sh
|
||||
TEST_PROGS += gro.sh
|
||||
|
@ -61,6 +62,8 @@ TEST_FILES := settings
|
|||
KSFT_KHDR_INSTALL := 1
|
||||
include ../lib.mk
|
||||
|
||||
include bpf/Makefile
|
||||
|
||||
$(OUTPUT)/reuseport_bpf_numa: LDLIBS += -lnuma
|
||||
$(OUTPUT)/tcp_mmap: LDLIBS += -lpthread
|
||||
$(OUTPUT)/tcp_inq: LDLIBS += -lpthread
|
||||
|
|
|
@ -0,0 +1,14 @@
|
|||
# SPDX-License-Identifier: GPL-2.0
|
||||
|
||||
CLANG ?= clang
|
||||
CCINCLUDE += -I../../bpf
|
||||
CCINCLUDE += -I../../../../../usr/include/
|
||||
|
||||
TEST_CUSTOM_PROGS = $(OUTPUT)/bpf/nat6to4.o
|
||||
all: $(TEST_CUSTOM_PROGS)
|
||||
|
||||
$(OUTPUT)/%.o: %.c
|
||||
$(CLANG) -O2 -target bpf -c $< $(CCINCLUDE) -o $@
|
||||
|
||||
clean:
|
||||
rm -f $(TEST_CUSTOM_PROGS)
|
|
@ -0,0 +1,285 @@
|
|||
// SPDX-License-Identifier: GPL-2.0-only
|
||||
/*
|
||||
* This code is taken from the Android Open Source Project and the author
|
||||
* (Maciej Żenczykowski) has gave permission to relicense it under the
|
||||
* GPLv2. Therefore this program is free software;
|
||||
* You can redistribute it and/or modify it under the terms of the GNU
|
||||
* General Public License version 2 as published by the Free Software
|
||||
* Foundation
|
||||
|
||||
* The original headers, including the original license headers, are
|
||||
* included below for completeness.
|
||||
*
|
||||
* Copyright (C) 2019 The Android Open Source Project
|
||||
*
|
||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||
* you may not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
#include <linux/bpf.h>
|
||||
#include <linux/if.h>
|
||||
#include <linux/if_ether.h>
|
||||
#include <linux/if_packet.h>
|
||||
#include <linux/in.h>
|
||||
#include <linux/in6.h>
|
||||
#include <linux/ip.h>
|
||||
#include <linux/ipv6.h>
|
||||
#include <linux/pkt_cls.h>
|
||||
#include <linux/swab.h>
|
||||
#include <stdbool.h>
|
||||
#include <stdint.h>
|
||||
|
||||
|
||||
#include <linux/udp.h>
|
||||
|
||||
#include <bpf/bpf_helpers.h>
|
||||
#include <bpf/bpf_endian.h>
|
||||
|
||||
#define IP_DF 0x4000 // Flag: "Don't Fragment"
|
||||
|
||||
SEC("schedcls/ingress6/nat_6")
|
||||
int sched_cls_ingress6_nat_6_prog(struct __sk_buff *skb)
|
||||
{
|
||||
const int l2_header_size = sizeof(struct ethhdr);
|
||||
void *data = (void *)(long)skb->data;
|
||||
const void *data_end = (void *)(long)skb->data_end;
|
||||
const struct ethhdr * const eth = data; // used iff is_ethernet
|
||||
const struct ipv6hdr * const ip6 = (void *)(eth + 1);
|
||||
|
||||
// Require ethernet dst mac address to be our unicast address.
|
||||
if (skb->pkt_type != PACKET_HOST)
|
||||
return TC_ACT_OK;
|
||||
|
||||
// Must be meta-ethernet IPv6 frame
|
||||
if (skb->protocol != bpf_htons(ETH_P_IPV6))
|
||||
return TC_ACT_OK;
|
||||
|
||||
// Must have (ethernet and) ipv6 header
|
||||
if (data + l2_header_size + sizeof(*ip6) > data_end)
|
||||
return TC_ACT_OK;
|
||||
|
||||
// Ethertype - if present - must be IPv6
|
||||
if (eth->h_proto != bpf_htons(ETH_P_IPV6))
|
||||
return TC_ACT_OK;
|
||||
|
||||
// IP version must be 6
|
||||
if (ip6->version != 6)
|
||||
return TC_ACT_OK;
|
||||
// Maximum IPv6 payload length that can be translated to IPv4
|
||||
if (bpf_ntohs(ip6->payload_len) > 0xFFFF - sizeof(struct iphdr))
|
||||
return TC_ACT_OK;
|
||||
switch (ip6->nexthdr) {
|
||||
case IPPROTO_TCP: // For TCP & UDP the checksum neutrality of the chosen IPv6
|
||||
case IPPROTO_UDP: // address means there is no need to update their checksums.
|
||||
case IPPROTO_GRE: // We do not need to bother looking at GRE/ESP headers,
|
||||
case IPPROTO_ESP: // since there is never a checksum to update.
|
||||
break;
|
||||
default: // do not know how to handle anything else
|
||||
return TC_ACT_OK;
|
||||
}
|
||||
|
||||
struct ethhdr eth2; // used iff is_ethernet
|
||||
|
||||
eth2 = *eth; // Copy over the ethernet header (src/dst mac)
|
||||
eth2.h_proto = bpf_htons(ETH_P_IP); // But replace the ethertype
|
||||
|
||||
struct iphdr ip = {
|
||||
.version = 4, // u4
|
||||
.ihl = sizeof(struct iphdr) / sizeof(__u32), // u4
|
||||
.tos = (ip6->priority << 4) + (ip6->flow_lbl[0] >> 4), // u8
|
||||
.tot_len = bpf_htons(bpf_ntohs(ip6->payload_len) + sizeof(struct iphdr)), // u16
|
||||
.id = 0, // u16
|
||||
.frag_off = bpf_htons(IP_DF), // u16
|
||||
.ttl = ip6->hop_limit, // u8
|
||||
.protocol = ip6->nexthdr, // u8
|
||||
.check = 0, // u16
|
||||
.saddr = 0x0201a8c0, // u32
|
||||
.daddr = 0x0101a8c0, // u32
|
||||
};
|
||||
|
||||
// Calculate the IPv4 one's complement checksum of the IPv4 header.
|
||||
__wsum sum4 = 0;
|
||||
|
||||
for (int i = 0; i < sizeof(ip) / sizeof(__u16); ++i)
|
||||
sum4 += ((__u16 *)&ip)[i];
|
||||
|
||||
// Note that sum4 is guaranteed to be non-zero by virtue of ip.version == 4
|
||||
sum4 = (sum4 & 0xFFFF) + (sum4 >> 16); // collapse u32 into range 1 .. 0x1FFFE
|
||||
sum4 = (sum4 & 0xFFFF) + (sum4 >> 16); // collapse any potential carry into u16
|
||||
ip.check = (__u16)~sum4; // sum4 cannot be zero, so this is never 0xFFFF
|
||||
|
||||
// Calculate the *negative* IPv6 16-bit one's complement checksum of the IPv6 header.
|
||||
__wsum sum6 = 0;
|
||||
// We'll end up with a non-zero sum due to ip6->version == 6 (which has '0' bits)
|
||||
for (int i = 0; i < sizeof(*ip6) / sizeof(__u16); ++i)
|
||||
sum6 += ~((__u16 *)ip6)[i]; // note the bitwise negation
|
||||
|
||||
// Note that there is no L4 checksum update: we are relying on the checksum neutrality
|
||||
// of the ipv6 address chosen by netd's ClatdController.
|
||||
|
||||
// Packet mutations begin - point of no return, but if this first modification fails
|
||||
// the packet is probably still pristine, so let clatd handle it.
|
||||
if (bpf_skb_change_proto(skb, bpf_htons(ETH_P_IP), 0))
|
||||
return TC_ACT_OK;
|
||||
bpf_csum_update(skb, sum6);
|
||||
|
||||
data = (void *)(long)skb->data;
|
||||
data_end = (void *)(long)skb->data_end;
|
||||
if (data + l2_header_size + sizeof(struct iphdr) > data_end)
|
||||
return TC_ACT_SHOT;
|
||||
|
||||
struct ethhdr *new_eth = data;
|
||||
|
||||
// Copy over the updated ethernet header
|
||||
*new_eth = eth2;
|
||||
|
||||
// Copy over the new ipv4 header.
|
||||
*(struct iphdr *)(new_eth + 1) = ip;
|
||||
return bpf_redirect(skb->ifindex, BPF_F_INGRESS);
|
||||
}
|
||||
|
||||
SEC("schedcls/egress4/snat4")
|
||||
int sched_cls_egress4_snat4_prog(struct __sk_buff *skb)
|
||||
{
|
||||
const int l2_header_size = sizeof(struct ethhdr);
|
||||
void *data = (void *)(long)skb->data;
|
||||
const void *data_end = (void *)(long)skb->data_end;
|
||||
const struct ethhdr *const eth = data; // used iff is_ethernet
|
||||
const struct iphdr *const ip4 = (void *)(eth + 1);
|
||||
|
||||
// Must be meta-ethernet IPv4 frame
|
||||
if (skb->protocol != bpf_htons(ETH_P_IP))
|
||||
return TC_ACT_OK;
|
||||
|
||||
// Must have ipv4 header
|
||||
if (data + l2_header_size + sizeof(struct ipv6hdr) > data_end)
|
||||
return TC_ACT_OK;
|
||||
|
||||
// Ethertype - if present - must be IPv4
|
||||
if (eth->h_proto != bpf_htons(ETH_P_IP))
|
||||
return TC_ACT_OK;
|
||||
|
||||
// IP version must be 4
|
||||
if (ip4->version != 4)
|
||||
return TC_ACT_OK;
|
||||
|
||||
// We cannot handle IP options, just standard 20 byte == 5 dword minimal IPv4 header
|
||||
if (ip4->ihl != 5)
|
||||
return TC_ACT_OK;
|
||||
|
||||
// Maximum IPv6 payload length that can be translated to IPv4
|
||||
if (bpf_htons(ip4->tot_len) > 0xFFFF - sizeof(struct ipv6hdr))
|
||||
return TC_ACT_OK;
|
||||
|
||||
// Calculate the IPv4 one's complement checksum of the IPv4 header.
|
||||
__wsum sum4 = 0;
|
||||
|
||||
for (int i = 0; i < sizeof(*ip4) / sizeof(__u16); ++i)
|
||||
sum4 += ((__u16 *)ip4)[i];
|
||||
|
||||
// Note that sum4 is guaranteed to be non-zero by virtue of ip4->version == 4
|
||||
sum4 = (sum4 & 0xFFFF) + (sum4 >> 16); // collapse u32 into range 1 .. 0x1FFFE
|
||||
sum4 = (sum4 & 0xFFFF) + (sum4 >> 16); // collapse any potential carry into u16
|
||||
// for a correct checksum we should get *a* zero, but sum4 must be positive, ie 0xFFFF
|
||||
if (sum4 != 0xFFFF)
|
||||
return TC_ACT_OK;
|
||||
|
||||
// Minimum IPv4 total length is the size of the header
|
||||
if (bpf_ntohs(ip4->tot_len) < sizeof(*ip4))
|
||||
return TC_ACT_OK;
|
||||
|
||||
// We are incapable of dealing with IPv4 fragments
|
||||
if (ip4->frag_off & ~bpf_htons(IP_DF))
|
||||
return TC_ACT_OK;
|
||||
|
||||
switch (ip4->protocol) {
|
||||
case IPPROTO_TCP: // For TCP & UDP the checksum neutrality of the chosen IPv6
|
||||
case IPPROTO_GRE: // address means there is no need to update their checksums.
|
||||
case IPPROTO_ESP: // We do not need to bother looking at GRE/ESP headers,
|
||||
break; // since there is never a checksum to update.
|
||||
|
||||
case IPPROTO_UDP: // See above comment, but must also have UDP header...
|
||||
if (data + sizeof(*ip4) + sizeof(struct udphdr) > data_end)
|
||||
return TC_ACT_OK;
|
||||
const struct udphdr *uh = (const struct udphdr *)(ip4 + 1);
|
||||
// If IPv4/UDP checksum is 0 then fallback to clatd so it can calculate the
|
||||
// checksum. Otherwise the network or more likely the NAT64 gateway might
|
||||
// drop the packet because in most cases IPv6/UDP packets with a zero checksum
|
||||
// are invalid. See RFC 6935. TODO: calculate checksum via bpf_csum_diff()
|
||||
if (!uh->check)
|
||||
return TC_ACT_OK;
|
||||
break;
|
||||
|
||||
default: // do not know how to handle anything else
|
||||
return TC_ACT_OK;
|
||||
}
|
||||
struct ethhdr eth2; // used iff is_ethernet
|
||||
|
||||
eth2 = *eth; // Copy over the ethernet header (src/dst mac)
|
||||
eth2.h_proto = bpf_htons(ETH_P_IPV6); // But replace the ethertype
|
||||
|
||||
struct ipv6hdr ip6 = {
|
||||
.version = 6, // __u8:4
|
||||
.priority = ip4->tos >> 4, // __u8:4
|
||||
.flow_lbl = {(ip4->tos & 0xF) << 4, 0, 0}, // __u8[3]
|
||||
.payload_len = bpf_htons(bpf_ntohs(ip4->tot_len) - 20), // __be16
|
||||
.nexthdr = ip4->protocol, // __u8
|
||||
.hop_limit = ip4->ttl, // __u8
|
||||
};
|
||||
ip6.saddr.in6_u.u6_addr32[0] = bpf_htonl(0x20010db8);
|
||||
ip6.saddr.in6_u.u6_addr32[1] = 0;
|
||||
ip6.saddr.in6_u.u6_addr32[2] = 0;
|
||||
ip6.saddr.in6_u.u6_addr32[3] = bpf_htonl(1);
|
||||
ip6.daddr.in6_u.u6_addr32[0] = bpf_htonl(0x20010db8);
|
||||
ip6.daddr.in6_u.u6_addr32[1] = 0;
|
||||
ip6.daddr.in6_u.u6_addr32[2] = 0;
|
||||
ip6.daddr.in6_u.u6_addr32[3] = bpf_htonl(2);
|
||||
|
||||
// Calculate the IPv6 16-bit one's complement checksum of the IPv6 header.
|
||||
__wsum sum6 = 0;
|
||||
// We'll end up with a non-zero sum due to ip6.version == 6
|
||||
for (int i = 0; i < sizeof(ip6) / sizeof(__u16); ++i)
|
||||
sum6 += ((__u16 *)&ip6)[i];
|
||||
|
||||
// Packet mutations begin - point of no return, but if this first modification fails
|
||||
// the packet is probably still pristine, so let clatd handle it.
|
||||
if (bpf_skb_change_proto(skb, bpf_htons(ETH_P_IPV6), 0))
|
||||
return TC_ACT_OK;
|
||||
|
||||
// This takes care of updating the skb->csum field for a CHECKSUM_COMPLETE packet.
|
||||
// In such a case, skb->csum is a 16-bit one's complement sum of the entire payload,
|
||||
// thus we need to subtract out the ipv4 header's sum, and add in the ipv6 header's sum.
|
||||
// However, we've already verified the ipv4 checksum is correct and thus 0.
|
||||
// Thus we only need to add the ipv6 header's sum.
|
||||
//
|
||||
// bpf_csum_update() always succeeds if the skb is CHECKSUM_COMPLETE and returns an error
|
||||
// (-ENOTSUPP) if it isn't. So we just ignore the return code (see above for more details).
|
||||
bpf_csum_update(skb, sum6);
|
||||
|
||||
// bpf_skb_change_proto() invalidates all pointers - reload them.
|
||||
data = (void *)(long)skb->data;
|
||||
data_end = (void *)(long)skb->data_end;
|
||||
|
||||
// I cannot think of any valid way for this error condition to trigger, however I do
|
||||
// believe the explicit check is required to keep the in kernel ebpf verifier happy.
|
||||
if (data + l2_header_size + sizeof(ip6) > data_end)
|
||||
return TC_ACT_SHOT;
|
||||
|
||||
struct ethhdr *new_eth = data;
|
||||
|
||||
// Copy over the updated ethernet header
|
||||
*new_eth = eth2;
|
||||
// Copy over the new ipv4 header.
|
||||
*(struct ipv6hdr *)(new_eth + 1) = ip6;
|
||||
return TC_ACT_OK;
|
||||
}
|
||||
|
||||
char _license[] SEC("license") = ("GPL");
|
|
@ -810,10 +810,16 @@ ipv4_ping()
|
|||
setup
|
||||
set_sysctl net.ipv4.raw_l3mdev_accept=1 2>/dev/null
|
||||
ipv4_ping_novrf
|
||||
setup
|
||||
set_sysctl net.ipv4.ping_group_range='0 2147483647' 2>/dev/null
|
||||
ipv4_ping_novrf
|
||||
|
||||
log_subsection "With VRF"
|
||||
setup "yes"
|
||||
ipv4_ping_vrf
|
||||
setup "yes"
|
||||
set_sysctl net.ipv4.ping_group_range='0 2147483647' 2>/dev/null
|
||||
ipv4_ping_vrf
|
||||
}
|
||||
|
||||
################################################################################
|
||||
|
@ -2348,10 +2354,16 @@ ipv6_ping()
|
|||
log_subsection "No VRF"
|
||||
setup
|
||||
ipv6_ping_novrf
|
||||
setup
|
||||
set_sysctl net.ipv4.ping_group_range='0 2147483647' 2>/dev/null
|
||||
ipv6_ping_novrf
|
||||
|
||||
log_subsection "With VRF"
|
||||
setup "yes"
|
||||
ipv6_ping_vrf
|
||||
setup "yes"
|
||||
set_sysctl net.ipv4.ping_group_range='0 2147483647' 2>/dev/null
|
||||
ipv6_ping_vrf
|
||||
}
|
||||
|
||||
################################################################################
|
||||
|
|
|
@ -0,0 +1,101 @@
|
|||
#!/bin/bash
|
||||
# SPDX-License-Identifier: GPL-2.0
|
||||
#
|
||||
# Run a series of udpgro benchmarks
|
||||
|
||||
readonly PEER_NS="ns-peer-$(mktemp -u XXXXXX)"
|
||||
|
||||
cleanup() {
|
||||
local -r jobs="$(jobs -p)"
|
||||
local -r ns="$(ip netns list|grep $PEER_NS)"
|
||||
|
||||
[ -n "${jobs}" ] && kill -INT ${jobs} 2>/dev/null
|
||||
[ -n "$ns" ] && ip netns del $ns 2>/dev/null
|
||||
}
|
||||
trap cleanup EXIT
|
||||
|
||||
run_one() {
|
||||
# use 'rx' as separator between sender args and receiver args
|
||||
local -r all="$@"
|
||||
local -r tx_args=${all%rx*}
|
||||
local rx_args=${all#*rx}
|
||||
|
||||
|
||||
|
||||
ip netns add "${PEER_NS}"
|
||||
ip -netns "${PEER_NS}" link set lo up
|
||||
ip link add type veth
|
||||
ip link set dev veth0 up
|
||||
ip addr add dev veth0 192.168.1.2/24
|
||||
ip addr add dev veth0 2001:db8::2/64 nodad
|
||||
|
||||
ip link set dev veth1 netns "${PEER_NS}"
|
||||
ip -netns "${PEER_NS}" addr add dev veth1 192.168.1.1/24
|
||||
ip -netns "${PEER_NS}" addr add dev veth1 2001:db8::1/64 nodad
|
||||
ip -netns "${PEER_NS}" link set dev veth1 up
|
||||
ip netns exec "${PEER_NS}" ethtool -K veth1 rx-gro-list on
|
||||
|
||||
|
||||
ip -n "${PEER_NS}" link set veth1 xdp object ../bpf/xdp_dummy.o section xdp_dummy
|
||||
tc -n "${PEER_NS}" qdisc add dev veth1 clsact
|
||||
tc -n "${PEER_NS}" filter add dev veth1 ingress prio 4 protocol ipv6 bpf object-file ../bpf/nat6to4.o section schedcls/ingress6/nat_6 direct-action
|
||||
tc -n "${PEER_NS}" filter add dev veth1 egress prio 4 protocol ip bpf object-file ../bpf/nat6to4.o section schedcls/egress4/snat4 direct-action
|
||||
echo ${rx_args}
|
||||
ip netns exec "${PEER_NS}" ./udpgso_bench_rx ${rx_args} -r &
|
||||
|
||||
# Hack: let bg programs complete the startup
|
||||
sleep 0.1
|
||||
./udpgso_bench_tx ${tx_args}
|
||||
}
|
||||
|
||||
run_in_netns() {
|
||||
local -r args=$@
|
||||
echo ${args}
|
||||
./in_netns.sh $0 __subprocess ${args}
|
||||
}
|
||||
|
||||
run_udp() {
|
||||
local -r args=$@
|
||||
|
||||
echo "udp gso - over veth touching data"
|
||||
run_in_netns ${args} -u -S 0 rx -4 -v
|
||||
|
||||
echo "udp gso and gro - over veth touching data"
|
||||
run_in_netns ${args} -S 0 rx -4 -G
|
||||
}
|
||||
|
||||
run_tcp() {
|
||||
local -r args=$@
|
||||
|
||||
echo "tcp - over veth touching data"
|
||||
run_in_netns ${args} -t rx -4 -t
|
||||
}
|
||||
|
||||
run_all() {
|
||||
local -r core_args="-l 4"
|
||||
local -r ipv4_args="${core_args} -4 -D 192.168.1.1"
|
||||
local -r ipv6_args="${core_args} -6 -D 2001:db8::1"
|
||||
|
||||
echo "ipv6"
|
||||
run_tcp "${ipv6_args}"
|
||||
run_udp "${ipv6_args}"
|
||||
}
|
||||
|
||||
if [ ! -f ../bpf/xdp_dummy.o ]; then
|
||||
echo "Missing xdp_dummy helper. Build bpf selftest first"
|
||||
exit -1
|
||||
fi
|
||||
|
||||
if [ ! -f bpf/nat6to4.o ]; then
|
||||
echo "Missing nat6to4 helper. Build bpfnat6to4.o selftest first"
|
||||
exit -1
|
||||
fi
|
||||
|
||||
if [[ $# -eq 0 ]]; then
|
||||
run_all
|
||||
elif [[ $1 == "__subprocess" ]]; then
|
||||
shift
|
||||
run_one $@
|
||||
else
|
||||
run_in_netns $@
|
||||
fi
|
Loading…
Reference in New Issue