2020-07-19 14:25:21 +07:00
|
|
|
/* SPDX-License-Identifier: GPL-2.0 OR BSD-3-Clause */
|
2016-09-06 18:34:49 +07:00
|
|
|
/*
|
2019-01-24 10:30:40 +07:00
|
|
|
* Copyright(c) 2016 - 2018 Intel Corporation.
|
2016-09-06 18:34:49 +07:00
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef IB_HDRS_H
|
|
|
|
#define IB_HDRS_H
|
|
|
|
|
|
|
|
#include <linux/types.h>
|
|
|
|
#include <asm/unaligned.h>
|
|
|
|
#include <rdma/ib_verbs.h>
|
|
|
|
|
|
|
|
#define IB_SEQ_NAK (3 << 29)
|
|
|
|
|
|
|
|
/* AETH NAK opcode values */
|
|
|
|
#define IB_RNR_NAK 0x20
|
|
|
|
#define IB_NAK_PSN_ERROR 0x60
|
|
|
|
#define IB_NAK_INVALID_REQUEST 0x61
|
|
|
|
#define IB_NAK_REMOTE_ACCESS_ERROR 0x62
|
|
|
|
#define IB_NAK_REMOTE_OPERATIONAL_ERROR 0x63
|
|
|
|
#define IB_NAK_INVALID_RD_REQUEST 0x64
|
|
|
|
|
|
|
|
#define IB_BTH_REQ_ACK BIT(31)
|
|
|
|
#define IB_BTH_SOLICITED BIT(23)
|
|
|
|
#define IB_BTH_MIG_REQ BIT(22)
|
|
|
|
|
|
|
|
#define IB_GRH_VERSION 6
|
|
|
|
#define IB_GRH_VERSION_MASK 0xF
|
|
|
|
#define IB_GRH_VERSION_SHIFT 28
|
|
|
|
#define IB_GRH_TCLASS_MASK 0xFF
|
|
|
|
#define IB_GRH_TCLASS_SHIFT 20
|
|
|
|
#define IB_GRH_FLOW_MASK 0xFFFFF
|
|
|
|
#define IB_GRH_FLOW_SHIFT 0
|
|
|
|
#define IB_GRH_NEXT_HDR 0x1B
|
2017-04-10 00:16:28 +07:00
|
|
|
#define IB_FECN_SHIFT 31
|
|
|
|
#define IB_FECN_MASK 1
|
|
|
|
#define IB_FECN_SMASK BIT(IB_FECN_SHIFT)
|
|
|
|
#define IB_BECN_SHIFT 30
|
|
|
|
#define IB_BECN_MASK 1
|
|
|
|
#define IB_BECN_SMASK BIT(IB_BECN_SHIFT)
|
2016-09-06 18:34:49 +07:00
|
|
|
|
2017-02-08 20:28:25 +07:00
|
|
|
#define IB_AETH_CREDIT_SHIFT 24
|
|
|
|
#define IB_AETH_CREDIT_MASK 0x1F
|
|
|
|
#define IB_AETH_CREDIT_INVAL 0x1F
|
|
|
|
#define IB_AETH_NAK_SHIFT 29
|
|
|
|
#define IB_MSN_MASK 0xFFFFFF
|
|
|
|
|
2016-09-06 18:34:49 +07:00
|
|
|
struct ib_reth {
|
|
|
|
__be64 vaddr; /* potentially unaligned */
|
|
|
|
__be32 rkey;
|
|
|
|
__be32 length;
|
|
|
|
} __packed;
|
|
|
|
|
|
|
|
struct ib_atomic_eth {
|
|
|
|
__be64 vaddr; /* potentially unaligned */
|
|
|
|
__be32 rkey;
|
|
|
|
__be64 swap_data; /* potentially unaligned */
|
|
|
|
__be64 compare_data; /* potentially unaligned */
|
|
|
|
} __packed;
|
|
|
|
|
2019-01-24 10:30:40 +07:00
|
|
|
#include <rdma/tid_rdma_defs.h>
|
|
|
|
|
2016-09-06 18:34:49 +07:00
|
|
|
union ib_ehdrs {
|
|
|
|
struct {
|
|
|
|
__be32 deth[2];
|
|
|
|
__be32 imm_data;
|
|
|
|
} ud;
|
|
|
|
struct {
|
|
|
|
struct ib_reth reth;
|
|
|
|
__be32 imm_data;
|
|
|
|
} rc;
|
|
|
|
struct {
|
|
|
|
__be32 aeth;
|
|
|
|
__be64 atomic_ack_eth; /* potentially unaligned */
|
|
|
|
} __packed at;
|
|
|
|
__be32 imm_data;
|
|
|
|
__be32 aeth;
|
|
|
|
__be32 ieth;
|
|
|
|
struct ib_atomic_eth atomic_eth;
|
2019-01-24 10:30:40 +07:00
|
|
|
/* TID RDMA headers */
|
|
|
|
union {
|
|
|
|
struct tid_rdma_read_req r_req;
|
|
|
|
struct tid_rdma_read_resp r_rsp;
|
2019-01-24 12:48:28 +07:00
|
|
|
struct tid_rdma_write_req w_req;
|
|
|
|
struct tid_rdma_write_resp w_rsp;
|
|
|
|
struct tid_rdma_write_data w_data;
|
|
|
|
struct tid_rdma_resync resync;
|
|
|
|
struct tid_rdma_ack ack;
|
2019-01-24 10:30:40 +07:00
|
|
|
} tid_rdma;
|
2016-09-06 18:34:49 +07:00
|
|
|
} __packed;
|
|
|
|
|
|
|
|
struct ib_other_headers {
|
|
|
|
__be32 bth[3];
|
|
|
|
union ib_ehdrs u;
|
|
|
|
} __packed;
|
|
|
|
|
|
|
|
struct ib_header {
|
|
|
|
__be16 lrh[4];
|
|
|
|
union {
|
|
|
|
struct {
|
|
|
|
struct ib_grh grh;
|
|
|
|
struct ib_other_headers oth;
|
|
|
|
} l;
|
|
|
|
struct ib_other_headers oth;
|
|
|
|
} u;
|
|
|
|
} __packed;
|
|
|
|
|
|
|
|
/* accessors for unaligned __be64 items */
|
|
|
|
|
|
|
|
static inline u64 ib_u64_get(__be64 *p)
|
|
|
|
{
|
|
|
|
return get_unaligned_be64(p);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void ib_u64_put(u64 val, __be64 *p)
|
|
|
|
{
|
|
|
|
put_unaligned_be64(val, p);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u64 get_ib_reth_vaddr(struct ib_reth *reth)
|
|
|
|
{
|
|
|
|
return ib_u64_get(&reth->vaddr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void put_ib_reth_vaddr(u64 val, struct ib_reth *reth)
|
|
|
|
{
|
|
|
|
ib_u64_put(val, &reth->vaddr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u64 get_ib_ateth_vaddr(struct ib_atomic_eth *ateth)
|
|
|
|
{
|
|
|
|
return ib_u64_get(&ateth->vaddr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void put_ib_ateth_vaddr(u64 val, struct ib_atomic_eth *ateth)
|
|
|
|
{
|
|
|
|
ib_u64_put(val, &ateth->vaddr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u64 get_ib_ateth_swap(struct ib_atomic_eth *ateth)
|
|
|
|
{
|
|
|
|
return ib_u64_get(&ateth->swap_data);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void put_ib_ateth_swap(u64 val, struct ib_atomic_eth *ateth)
|
|
|
|
{
|
|
|
|
ib_u64_put(val, &ateth->swap_data);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u64 get_ib_ateth_compare(struct ib_atomic_eth *ateth)
|
|
|
|
{
|
|
|
|
return ib_u64_get(&ateth->compare_data);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void put_ib_ateth_compare(u64 val, struct ib_atomic_eth *ateth)
|
|
|
|
{
|
|
|
|
ib_u64_put(val, &ateth->compare_data);
|
|
|
|
}
|
|
|
|
|
2017-04-10 00:16:22 +07:00
|
|
|
/*
|
|
|
|
* 9B/IB Packet Format
|
|
|
|
*/
|
|
|
|
#define IB_LNH_MASK 3
|
|
|
|
#define IB_SC_MASK 0xf
|
|
|
|
#define IB_SC_SHIFT 12
|
2017-05-12 23:19:55 +07:00
|
|
|
#define IB_SC5_MASK 0x10
|
2017-04-10 00:16:22 +07:00
|
|
|
#define IB_SL_MASK 0xf
|
|
|
|
#define IB_SL_SHIFT 4
|
2017-05-12 23:19:55 +07:00
|
|
|
#define IB_SL_SHIFT 4
|
|
|
|
#define IB_LVER_MASK 0xf
|
|
|
|
#define IB_LVER_SHIFT 8
|
2017-04-10 00:16:22 +07:00
|
|
|
|
|
|
|
static inline u8 ib_get_lnh(struct ib_header *hdr)
|
|
|
|
{
|
|
|
|
return (be16_to_cpu(hdr->lrh[0]) & IB_LNH_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u8 ib_get_sc(struct ib_header *hdr)
|
|
|
|
{
|
|
|
|
return ((be16_to_cpu(hdr->lrh[0]) >> IB_SC_SHIFT) & IB_SC_MASK);
|
|
|
|
}
|
|
|
|
|
2017-05-12 23:19:55 +07:00
|
|
|
static inline bool ib_is_sc5(u16 sc5)
|
|
|
|
{
|
|
|
|
return !!(sc5 & IB_SC5_MASK);
|
|
|
|
}
|
|
|
|
|
2017-04-10 00:16:22 +07:00
|
|
|
static inline u8 ib_get_sl(struct ib_header *hdr)
|
|
|
|
{
|
|
|
|
return ((be16_to_cpu(hdr->lrh[0]) >> IB_SL_SHIFT) & IB_SL_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u16 ib_get_dlid(struct ib_header *hdr)
|
|
|
|
{
|
|
|
|
return (be16_to_cpu(hdr->lrh[1]));
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u16 ib_get_slid(struct ib_header *hdr)
|
|
|
|
{
|
|
|
|
return (be16_to_cpu(hdr->lrh[3]));
|
|
|
|
}
|
|
|
|
|
2017-05-12 23:19:55 +07:00
|
|
|
static inline u8 ib_get_lver(struct ib_header *hdr)
|
|
|
|
{
|
|
|
|
return (u8)((be16_to_cpu(hdr->lrh[0]) >> IB_LVER_SHIFT) &
|
|
|
|
IB_LVER_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u16 ib_get_len(struct ib_header *hdr)
|
|
|
|
{
|
|
|
|
return (u16)(be16_to_cpu(hdr->lrh[2]));
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 ib_get_qkey(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return be32_to_cpu(ohdr->u.ud.deth[0]);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 ib_get_sqpn(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return ((be32_to_cpu(ohdr->u.ud.deth[1])) & IB_QPN_MASK);
|
|
|
|
}
|
|
|
|
|
2017-04-10 00:16:22 +07:00
|
|
|
/*
|
|
|
|
* BTH
|
|
|
|
*/
|
|
|
|
#define IB_BTH_OPCODE_MASK 0xff
|
|
|
|
#define IB_BTH_OPCODE_SHIFT 24
|
|
|
|
#define IB_BTH_PAD_MASK 3
|
|
|
|
#define IB_BTH_PKEY_MASK 0xffff
|
|
|
|
#define IB_BTH_PAD_SHIFT 20
|
2017-05-12 23:19:55 +07:00
|
|
|
#define IB_BTH_A_MASK 1
|
|
|
|
#define IB_BTH_A_SHIFT 31
|
|
|
|
#define IB_BTH_M_MASK 1
|
|
|
|
#define IB_BTH_M_SHIFT 22
|
|
|
|
#define IB_BTH_SE_MASK 1
|
|
|
|
#define IB_BTH_SE_SHIFT 23
|
|
|
|
#define IB_BTH_TVER_MASK 0xf
|
|
|
|
#define IB_BTH_TVER_SHIFT 16
|
2017-04-10 00:16:22 +07:00
|
|
|
|
|
|
|
static inline u8 ib_bth_get_pad(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return ((be32_to_cpu(ohdr->bth[0]) >> IB_BTH_PAD_SHIFT) &
|
|
|
|
IB_BTH_PAD_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u16 ib_bth_get_pkey(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return (be32_to_cpu(ohdr->bth[0]) & IB_BTH_PKEY_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u8 ib_bth_get_opcode(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return ((be32_to_cpu(ohdr->bth[0]) >> IB_BTH_OPCODE_SHIFT) &
|
|
|
|
IB_BTH_OPCODE_MASK);
|
|
|
|
}
|
|
|
|
|
2017-05-12 23:19:55 +07:00
|
|
|
static inline u8 ib_bth_get_ackreq(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return (u8)((be32_to_cpu(ohdr->bth[2]) >> IB_BTH_A_SHIFT) &
|
|
|
|
IB_BTH_A_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u8 ib_bth_get_migreq(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return (u8)((be32_to_cpu(ohdr->bth[0]) >> IB_BTH_M_SHIFT) &
|
|
|
|
IB_BTH_M_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u8 ib_bth_get_se(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return (u8)((be32_to_cpu(ohdr->bth[0]) >> IB_BTH_SE_SHIFT) &
|
|
|
|
IB_BTH_SE_MASK);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 ib_bth_get_psn(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return (u32)(be32_to_cpu(ohdr->bth[2]));
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 ib_bth_get_qpn(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return (u32)((be32_to_cpu(ohdr->bth[1])) & IB_QPN_MASK);
|
|
|
|
}
|
|
|
|
|
2018-02-02 01:46:38 +07:00
|
|
|
static inline bool ib_bth_get_becn(struct ib_other_headers *ohdr)
|
2017-05-12 23:19:55 +07:00
|
|
|
{
|
2018-02-02 01:46:38 +07:00
|
|
|
return (ohdr->bth[1]) & cpu_to_be32(IB_BECN_SMASK);
|
2017-05-12 23:19:55 +07:00
|
|
|
}
|
|
|
|
|
2018-02-02 01:46:38 +07:00
|
|
|
static inline bool ib_bth_get_fecn(struct ib_other_headers *ohdr)
|
2017-05-12 23:19:55 +07:00
|
|
|
{
|
2018-02-02 01:46:38 +07:00
|
|
|
return (ohdr->bth[1]) & cpu_to_be32(IB_FECN_SMASK);
|
2017-05-12 23:19:55 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline u8 ib_bth_get_tver(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return (u8)((be32_to_cpu(ohdr->bth[0]) >> IB_BTH_TVER_SHIFT) &
|
|
|
|
IB_BTH_TVER_MASK);
|
|
|
|
}
|
|
|
|
|
2018-02-02 01:46:15 +07:00
|
|
|
static inline bool ib_bth_is_solicited(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return ohdr->bth[0] & cpu_to_be32(IB_BTH_SOLICITED);
|
|
|
|
}
|
2018-02-02 01:46:23 +07:00
|
|
|
|
|
|
|
static inline bool ib_bth_is_migration(struct ib_other_headers *ohdr)
|
|
|
|
{
|
|
|
|
return ohdr->bth[0] & cpu_to_be32(IB_BTH_MIG_REQ);
|
|
|
|
}
|
2016-09-06 18:34:49 +07:00
|
|
|
#endif /* IB_HDRS_H */
|