net/mlx5: DR, Replace the check for valid STE entry
Validity check is done by reading the next lu_type from the STE, this check can be replaced by checking the refcount. This will make the check independent on internal STE structure. Signed-off-by: Alex Vesker <valex@nvidia.com> Signed-off-by: Yevgeny Kliteynik <kliteyn@nvidia.com> Signed-off-by: Saeed Mahameed <saeedm@nvidia.com>
This commit is contained in:
Родитель
f2e834694b
Коммит
97ffd895fe
|
@ -242,7 +242,7 @@ dr_rule_rehash_copy_ste(struct mlx5dr_matcher *matcher,
|
|||
new_idx = mlx5dr_ste_calc_hash_index(hw_ste, new_htbl);
|
||||
new_ste = &new_htbl->ste_arr[new_idx];
|
||||
|
||||
if (mlx5dr_ste_not_used_ste(new_ste)) {
|
||||
if (mlx5dr_ste_is_not_used(new_ste)) {
|
||||
mlx5dr_htbl_get(new_htbl);
|
||||
list_add_tail(&new_ste->miss_list_node,
|
||||
mlx5dr_ste_get_miss_list(new_ste));
|
||||
|
@ -335,7 +335,7 @@ static int dr_rule_rehash_copy_htbl(struct mlx5dr_matcher *matcher,
|
|||
|
||||
for (i = 0; i < cur_entries; i++) {
|
||||
cur_ste = &cur_htbl->ste_arr[i];
|
||||
if (mlx5dr_ste_not_used_ste(cur_ste)) /* Empty, nothing to copy */
|
||||
if (mlx5dr_ste_is_not_used(cur_ste)) /* Empty, nothing to copy */
|
||||
continue;
|
||||
|
||||
err = dr_rule_rehash_copy_miss_list(matcher,
|
||||
|
@ -791,7 +791,7 @@ again:
|
|||
miss_list = &cur_htbl->chunk->miss_list[index];
|
||||
ste = &cur_htbl->ste_arr[index];
|
||||
|
||||
if (mlx5dr_ste_not_used_ste(ste)) {
|
||||
if (mlx5dr_ste_is_not_used(ste)) {
|
||||
if (dr_rule_handle_empty_entry(matcher, nic_matcher, cur_htbl,
|
||||
ste, ste_location,
|
||||
hw_ste, miss_list,
|
||||
|
|
|
@ -466,10 +466,10 @@ int mlx5dr_send_postsend_htbl(struct mlx5dr_domain *dmn,
|
|||
* need to add the bit_mask
|
||||
*/
|
||||
for (j = 0; j < num_stes_per_iter; j++) {
|
||||
u8 *hw_ste = htbl->ste_arr[ste_index + j].hw_ste;
|
||||
struct mlx5dr_ste *ste = &htbl->ste_arr[ste_index + j];
|
||||
u32 ste_off = j * DR_STE_SIZE;
|
||||
|
||||
if (mlx5dr_ste_is_not_valid_entry(hw_ste)) {
|
||||
if (mlx5dr_ste_is_not_used(ste)) {
|
||||
memcpy(data + ste_off,
|
||||
formatted_ste, DR_STE_SIZE);
|
||||
} else {
|
||||
|
|
|
@ -549,25 +549,6 @@ void mlx5dr_ste_always_miss_addr(struct mlx5dr_ste *ste, u64 miss_addr)
|
|||
dr_ste_set_always_miss((struct dr_hw_ste_format *)ste->hw_ste);
|
||||
}
|
||||
|
||||
/* The assumption here is that we don't update the ste->hw_ste if it is not
|
||||
* used ste, so it will be all zero, checking the next_lu_type.
|
||||
*/
|
||||
bool mlx5dr_ste_is_not_valid_entry(u8 *p_hw_ste)
|
||||
{
|
||||
struct dr_hw_ste_format *hw_ste = (struct dr_hw_ste_format *)p_hw_ste;
|
||||
|
||||
if (MLX5_GET(ste_general, hw_ste, next_lu_type) ==
|
||||
MLX5DR_STE_LU_TYPE_NOP)
|
||||
return true;
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
bool mlx5dr_ste_not_used_ste(struct mlx5dr_ste *ste)
|
||||
{
|
||||
return !ste->refcount;
|
||||
}
|
||||
|
||||
/* Init one ste as a pattern for ste data array */
|
||||
void mlx5dr_ste_set_formatted_ste(u16 gvmi,
|
||||
struct mlx5dr_domain_rx_tx *nic_dmn,
|
||||
|
|
|
@ -227,7 +227,6 @@ void mlx5dr_ste_set_hit_gvmi(u8 *hw_ste_p, u16 gvmi);
|
|||
void mlx5dr_ste_set_hit_addr(u8 *hw_ste, u64 icm_addr, u32 ht_size);
|
||||
void mlx5dr_ste_always_miss_addr(struct mlx5dr_ste *ste, u64 miss_addr);
|
||||
void mlx5dr_ste_set_bit_mask(u8 *hw_ste_p, u8 *bit_mask);
|
||||
bool mlx5dr_ste_not_used_ste(struct mlx5dr_ste *ste);
|
||||
bool mlx5dr_ste_is_last_in_rule(struct mlx5dr_matcher_rx_tx *nic_matcher,
|
||||
u8 ste_location);
|
||||
void mlx5dr_ste_rx_set_flow_tag(u8 *hw_ste_p, u32 flow_tag);
|
||||
|
@ -266,6 +265,11 @@ static inline void mlx5dr_ste_get(struct mlx5dr_ste *ste)
|
|||
ste->refcount++;
|
||||
}
|
||||
|
||||
static inline bool mlx5dr_ste_is_not_used(struct mlx5dr_ste *ste)
|
||||
{
|
||||
return !ste->refcount;
|
||||
}
|
||||
|
||||
void mlx5dr_ste_set_hit_addr_by_next_htbl(u8 *hw_ste,
|
||||
struct mlx5dr_ste_htbl *next_htbl);
|
||||
bool mlx5dr_ste_equal_tag(void *src, void *dst);
|
||||
|
@ -991,7 +995,6 @@ struct mlx5dr_icm_chunk *
|
|||
mlx5dr_icm_alloc_chunk(struct mlx5dr_icm_pool *pool,
|
||||
enum mlx5dr_icm_chunk_size chunk_size);
|
||||
void mlx5dr_icm_free_chunk(struct mlx5dr_icm_chunk *chunk);
|
||||
bool mlx5dr_ste_is_not_valid_entry(u8 *p_hw_ste);
|
||||
int mlx5dr_ste_htbl_init_and_postsend(struct mlx5dr_domain *dmn,
|
||||
struct mlx5dr_domain_rx_tx *nic_dmn,
|
||||
struct mlx5dr_ste_htbl *htbl,
|
||||
|
|
Загрузка…
Ссылка в новой задаче