net: atlantic: changes for multi-TC support
This patch contains the following changes: * add cfg->is_ptp (used for PTP enable/disable switch, which is described in more details below); * add cfg->tc_mode (A1 supports 2 HW modes only); * setup queue to TC mapping based on TC mode on A2; * remove hw_tx_tc_mode_get / hw_rx_tc_mode_get hw_ops. In the first generation of our hardware (A1), a whole traffic class is consumed for PTP handling in FW (FW uses it to send the ptp data and to send back timestamps). The 'is_ptp' flag introduced in this patch will be used in to automatically disable PTP when a conflicting configuration is detected, e.g. when multiple TCs are enabled. Signed-off-by: Dmitry Bezrukov <dbezrukov@marvell.com> Co-developed-by: Mark Starovoytov <mstarovoitov@marvell.com> Signed-off-by: Mark Starovoytov <mstarovoitov@marvell.com> Signed-off-by: Igor Russkikh <irusskikh@marvell.com> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
593dd0fc20
commit
0aa7bc3ee4
|
@ -18,6 +18,12 @@
|
|||
#define AQ_HW_MAC_COUNTER_HZ 312500000ll
|
||||
#define AQ_HW_PHY_COUNTER_HZ 160000000ll
|
||||
|
||||
enum aq_tc_mode {
|
||||
AQ_TC_MODE_INVALID = -1,
|
||||
AQ_TC_MODE_8TCS,
|
||||
AQ_TC_MODE_4TCS,
|
||||
};
|
||||
|
||||
#define AQ_RX_FIRST_LOC_FVLANID 0U
|
||||
#define AQ_RX_LAST_LOC_FVLANID 15U
|
||||
#define AQ_RX_FIRST_LOC_FETHERT 16U
|
||||
|
@ -281,10 +287,6 @@ struct aq_hw_ops {
|
|||
int (*hw_set_offload)(struct aq_hw_s *self,
|
||||
struct aq_nic_cfg_s *aq_nic_cfg);
|
||||
|
||||
int (*hw_tx_tc_mode_get)(struct aq_hw_s *self, u32 *tc_mode);
|
||||
|
||||
int (*hw_rx_tc_mode_get)(struct aq_hw_s *self, u32 *tc_mode);
|
||||
|
||||
int (*hw_ring_hwts_rx_fill)(struct aq_hw_s *self,
|
||||
struct aq_ring_s *aq_ring);
|
||||
|
||||
|
|
|
@ -89,6 +89,7 @@ void aq_nic_cfg_start(struct aq_nic_s *self)
|
|||
cfg->is_autoneg = AQ_CFG_IS_AUTONEG_DEF;
|
||||
|
||||
cfg->is_lro = AQ_CFG_IS_LRO_DEF;
|
||||
cfg->is_ptp = true;
|
||||
|
||||
/*descriptors */
|
||||
cfg->rxds = min(cfg->aq_hw_caps->rxds_max, AQ_CFG_RXDS_DEF);
|
||||
|
@ -122,6 +123,11 @@ void aq_nic_cfg_start(struct aq_nic_s *self)
|
|||
cfg->vecs = 1U;
|
||||
}
|
||||
|
||||
if (cfg->vecs <= 4)
|
||||
cfg->tc_mode = AQ_TC_MODE_8TCS;
|
||||
else
|
||||
cfg->tc_mode = AQ_TC_MODE_4TCS;
|
||||
|
||||
/* Check if we have enough vectors allocated for
|
||||
* link status IRQ. If no - we'll know link state from
|
||||
* slower service task.
|
||||
|
@ -409,17 +415,19 @@ int aq_nic_init(struct aq_nic_s *self)
|
|||
aq_vec_init(aq_vec, self->aq_hw_ops, self->aq_hw);
|
||||
}
|
||||
|
||||
err = aq_ptp_init(self, self->irqvecs - 1);
|
||||
if (err < 0)
|
||||
goto err_exit;
|
||||
if (aq_nic_get_cfg(self)->is_ptp) {
|
||||
err = aq_ptp_init(self, self->irqvecs - 1);
|
||||
if (err < 0)
|
||||
goto err_exit;
|
||||
|
||||
err = aq_ptp_ring_alloc(self);
|
||||
if (err < 0)
|
||||
goto err_exit;
|
||||
err = aq_ptp_ring_alloc(self);
|
||||
if (err < 0)
|
||||
goto err_exit;
|
||||
|
||||
err = aq_ptp_ring_init(self);
|
||||
if (err < 0)
|
||||
goto err_exit;
|
||||
err = aq_ptp_ring_init(self);
|
||||
if (err < 0)
|
||||
goto err_exit;
|
||||
}
|
||||
|
||||
netif_carrier_off(self->ndev);
|
||||
|
||||
|
|
|
@ -59,6 +59,8 @@ struct aq_nic_cfg_s {
|
|||
bool is_polling;
|
||||
bool is_rss;
|
||||
bool is_lro;
|
||||
bool is_ptp;
|
||||
enum aq_tc_mode tc_mode;
|
||||
u32 priv_flags;
|
||||
u8 tcs;
|
||||
struct aq_rss_parameters aq_rss;
|
||||
|
|
|
@ -945,26 +945,29 @@ void aq_ptp_ring_deinit(struct aq_nic_s *aq_nic)
|
|||
#define PTP_4TC_RING_IDX 16
|
||||
#define PTP_HWST_RING_IDX 31
|
||||
|
||||
/* Index must be 8 (8 TCs) or 16 (4 TCs).
|
||||
* It depends on Traffic Class mode.
|
||||
*/
|
||||
static unsigned int ptp_ring_idx(const enum aq_tc_mode tc_mode)
|
||||
{
|
||||
if (tc_mode == AQ_TC_MODE_8TCS)
|
||||
return PTP_8TC_RING_IDX;
|
||||
|
||||
return PTP_4TC_RING_IDX;
|
||||
}
|
||||
|
||||
int aq_ptp_ring_alloc(struct aq_nic_s *aq_nic)
|
||||
{
|
||||
struct aq_ptp_s *aq_ptp = aq_nic->aq_ptp;
|
||||
unsigned int tx_ring_idx, rx_ring_idx;
|
||||
struct aq_ring_s *hwts;
|
||||
u32 tx_tc_mode, rx_tc_mode;
|
||||
struct aq_ring_s *ring;
|
||||
int err;
|
||||
|
||||
if (!aq_ptp)
|
||||
return 0;
|
||||
|
||||
/* Index must to be 8 (8 TCs) or 16 (4 TCs).
|
||||
* It depends from Traffic Class mode.
|
||||
*/
|
||||
aq_nic->aq_hw_ops->hw_tx_tc_mode_get(aq_nic->aq_hw, &tx_tc_mode);
|
||||
if (tx_tc_mode == 0)
|
||||
tx_ring_idx = PTP_8TC_RING_IDX;
|
||||
else
|
||||
tx_ring_idx = PTP_4TC_RING_IDX;
|
||||
tx_ring_idx = ptp_ring_idx(aq_nic->aq_nic_cfg.tc_mode);
|
||||
|
||||
ring = aq_ring_tx_alloc(&aq_ptp->ptp_tx, aq_nic,
|
||||
tx_ring_idx, &aq_nic->aq_nic_cfg);
|
||||
|
@ -973,11 +976,7 @@ int aq_ptp_ring_alloc(struct aq_nic_s *aq_nic)
|
|||
goto err_exit;
|
||||
}
|
||||
|
||||
aq_nic->aq_hw_ops->hw_rx_tc_mode_get(aq_nic->aq_hw, &rx_tc_mode);
|
||||
if (rx_tc_mode == 0)
|
||||
rx_ring_idx = PTP_8TC_RING_IDX;
|
||||
else
|
||||
rx_ring_idx = PTP_4TC_RING_IDX;
|
||||
rx_ring_idx = ptp_ring_idx(aq_nic->aq_nic_cfg.tc_mode);
|
||||
|
||||
ring = aq_ring_rx_alloc(&aq_ptp->ptp_rx, aq_nic,
|
||||
rx_ring_idx, &aq_nic->aq_nic_cfg);
|
||||
|
|
|
@ -131,13 +131,16 @@ static int hw_atl_b0_tc_ptp_set(struct aq_hw_s *self)
|
|||
|
||||
static int hw_atl_b0_hw_qos_set(struct aq_hw_s *self)
|
||||
{
|
||||
struct aq_nic_cfg_s *cfg = self->aq_nic_cfg;
|
||||
u32 tx_buff_size = HW_ATL_B0_TXBUF_MAX;
|
||||
u32 rx_buff_size = HW_ATL_B0_RXBUF_MAX;
|
||||
unsigned int i_priority = 0U;
|
||||
u32 tc = 0U;
|
||||
|
||||
tx_buff_size -= HW_ATL_B0_PTP_TXBUF_SIZE;
|
||||
rx_buff_size -= HW_ATL_B0_PTP_RXBUF_SIZE;
|
||||
if (cfg->is_ptp) {
|
||||
tx_buff_size -= HW_ATL_B0_PTP_TXBUF_SIZE;
|
||||
rx_buff_size -= HW_ATL_B0_PTP_RXBUF_SIZE;
|
||||
}
|
||||
|
||||
/* TPS Descriptor rate init */
|
||||
hw_atl_tps_tx_pkt_shed_desc_rate_curr_time_res_set(self, 0x0U);
|
||||
|
@ -180,7 +183,8 @@ static int hw_atl_b0_hw_qos_set(struct aq_hw_s *self)
|
|||
|
||||
hw_atl_b0_set_fc(self, self->aq_nic_cfg->fc.req, tc);
|
||||
|
||||
hw_atl_b0_tc_ptp_set(self);
|
||||
if (cfg->is_ptp)
|
||||
hw_atl_b0_tc_ptp_set(self);
|
||||
|
||||
/* QoS 802.1p priority -> TC mapping */
|
||||
for (i_priority = 8U; i_priority--;)
|
||||
|
@ -1079,18 +1083,6 @@ int hw_atl_b0_hw_ring_rx_stop(struct aq_hw_s *self, struct aq_ring_s *ring)
|
|||
return aq_hw_err_from_flags(self);
|
||||
}
|
||||
|
||||
static int hw_atl_b0_tx_tc_mode_get(struct aq_hw_s *self, u32 *tc_mode)
|
||||
{
|
||||
*tc_mode = hw_atl_tpb_tps_tx_tc_mode_get(self);
|
||||
return aq_hw_err_from_flags(self);
|
||||
}
|
||||
|
||||
static int hw_atl_b0_rx_tc_mode_get(struct aq_hw_s *self, u32 *tc_mode)
|
||||
{
|
||||
*tc_mode = hw_atl_rpb_rpf_rx_traf_class_mode_get(self);
|
||||
return aq_hw_err_from_flags(self);
|
||||
}
|
||||
|
||||
#define get_ptp_ts_val_u64(self, indx) \
|
||||
((u64)(hw_atl_pcs_ptp_clock_get(self, indx) & 0xffff))
|
||||
|
||||
|
@ -1508,9 +1500,6 @@ const struct aq_hw_ops hw_atl_ops_b0 = {
|
|||
.hw_get_hw_stats = hw_atl_utils_get_hw_stats,
|
||||
.hw_get_fw_version = hw_atl_utils_get_fw_version,
|
||||
|
||||
.hw_tx_tc_mode_get = hw_atl_b0_tx_tc_mode_get,
|
||||
.hw_rx_tc_mode_get = hw_atl_b0_rx_tc_mode_get,
|
||||
|
||||
.hw_ring_hwts_rx_fill = hw_atl_b0_hw_ring_hwts_rx_fill,
|
||||
.hw_ring_hwts_rx_receive = hw_atl_b0_hw_ring_hwts_rx_receive,
|
||||
|
||||
|
|
|
@ -91,16 +91,36 @@ static int hw_atl2_hw_reset(struct aq_hw_s *self)
|
|||
|
||||
static int hw_atl2_hw_queue_to_tc_map_set(struct aq_hw_s *self)
|
||||
{
|
||||
if (!hw_atl_rpb_rpf_rx_traf_class_mode_get(self)) {
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(0), 0x11110000);
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(8), 0x33332222);
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(16), 0x55554444);
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(24), 0x77776666);
|
||||
} else {
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(0), 0x00000000);
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(8), 0x11111111);
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(16), 0x22222222);
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(24), 0x33333333);
|
||||
struct aq_nic_cfg_s *cfg = self->aq_nic_cfg;
|
||||
unsigned int tcs, q_per_tc;
|
||||
unsigned int tc, q;
|
||||
u32 value = 0;
|
||||
|
||||
switch (cfg->tc_mode) {
|
||||
case AQ_TC_MODE_8TCS:
|
||||
tcs = 8;
|
||||
q_per_tc = 4;
|
||||
break;
|
||||
case AQ_TC_MODE_4TCS:
|
||||
tcs = 4;
|
||||
q_per_tc = 8;
|
||||
break;
|
||||
default:
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
for (tc = 0; tc != tcs; tc++) {
|
||||
unsigned int tc_q_offset = tc * q_per_tc;
|
||||
|
||||
for (q = tc_q_offset; q != tc_q_offset + q_per_tc; q++)
|
||||
value |= tc << HW_ATL2_RX_Q_TC_MAP_SHIFT(q);
|
||||
|
||||
if (HW_ATL2_RX_Q_TC_MAP_ADR(q) !=
|
||||
HW_ATL2_RX_Q_TC_MAP_ADR(q - 1)) {
|
||||
aq_hw_write_reg(self, HW_ATL2_RX_Q_TC_MAP_ADR(q - 1),
|
||||
value);
|
||||
value = 0;
|
||||
}
|
||||
}
|
||||
|
||||
return aq_hw_err_from_flags(self);
|
||||
|
|
Loading…
Reference in New Issue