> > Done via > > $ git ls-files | xargs sed -i -e 's/htonll/htobe64/g' > > Signed-off-by: Jason Gunthorpe <jgunthorpe@xxxxxxxxxxxxxxxxxxxx> > --- > ibacm/include/acm_mad.h | 84 +++++++++++++++++++++---------------------- > ibacm/prov/acmp/src/acmp.c | 10 +++--- > ibacm/src/acm.c | 6 ++-- > providers/cxgb3/cxio_wr.h | 2 +- > providers/cxgb3/qp.c | 12 +++---- > providers/cxgb4/t4.h | 2 +- > providers/mlx4/qp.c | 12 +++---- > providers/mlx4/srq.c | 2 +- > providers/mlx5/qp.c | 24 ++++++------- > providers/mlx5/srq.c | 2 +- > providers/mthca/qp.c | 34 +++++++++--------- > providers/mthca/srq.c | 4 +-- > providers/qedr/qelr_verbs.c | 2 +- > srp_daemon/srp_daemon.c | 16 ++++----- > srp_daemon/srp_handle_traps.c | 4 +-- > 15 files changed, 108 insertions(+), 108 deletions(-) > > diff --git a/ibacm/include/acm_mad.h b/ibacm/include/acm_mad.h > index 68e231bc6680ca..f237ef6fde563c 100644 > --- a/ibacm/include/acm_mad.h > +++ b/ibacm/include/acm_mad.h > @@ -121,52 +121,52 @@ struct ib_sa_mad { > > #define IB_SA_ATTR_PATH_REC htons(0x0035) > > -#define IB_COMP_MASK_PR_SERVICE_ID (htonll(1 << 0) | \ > - htonll(1 << 1)) > -#define IB_COMP_MASK_PR_DGID htonll(1 << 2) > -#define IB_COMP_MASK_PR_SGID htonll(1 << 3) > -#define IB_COMP_MASK_PR_DLID htonll(1 << 4) > -#define IB_COMP_MASK_PR_SLID htonll(1 << 5) > -#define IB_COMP_MASK_PR_RAW_TRAFFIC htonll(1 << 6) > -/* RESERVED htonll(1 << 7) */ > -#define IB_COMP_MASK_PR_FLOW_LABEL htonll(1 << 8) > -#define IB_COMP_MASK_PR_HOP_LIMIT htonll(1 << 9) > -#define IB_COMP_MASK_PR_TCLASS htonll(1 << 10) > -#define IB_COMP_MASK_PR_REVERSIBLE htonll(1 << 11) > -#define IB_COMP_MASK_PR_NUM_PATH htonll(1 << 12) > -#define IB_COMP_MASK_PR_PKEY htonll(1 << 13) > -#define IB_COMP_MASK_PR_QOS_CLASS htonll(1 << 14) > -#define IB_COMP_MASK_PR_SL htonll(1 << 15) > -#define IB_COMP_MASK_PR_MTU_SELECTOR htonll(1 << 16) > -#define IB_COMP_MASK_PR_MTU htonll(1 << 17) > -#define IB_COMP_MASK_PR_RATE_SELECTOR htonll(1 << 18) > -#define IB_COMP_MASK_PR_RATE htonll(1 << 19) > -#define IB_COMP_MASK_PR_PACKET_LIFETIME_SELECTOR htonll(1 << 20) > -#define IB_COMP_MASK_PR_PACKET_LIFETIME htonll(1 << 21) > -#define IB_COMP_MASK_PR_PREFERENCE htonll(1 << 22) > -/* RESERVED htonll(1 << 23) */ > +#define IB_COMP_MASK_PR_SERVICE_ID (htobe64(1 << 0) | \ > + htobe64(1 << 1)) > +#define IB_COMP_MASK_PR_DGID htobe64(1 << 2) > +#define IB_COMP_MASK_PR_SGID htobe64(1 << 3) > +#define IB_COMP_MASK_PR_DLID htobe64(1 << 4) > +#define IB_COMP_MASK_PR_SLID htobe64(1 << 5) > +#define IB_COMP_MASK_PR_RAW_TRAFFIC htobe64(1 << 6) > +/* RESERVED htobe64(1 << 7) */ > +#define IB_COMP_MASK_PR_FLOW_LABEL htobe64(1 << 8) > +#define IB_COMP_MASK_PR_HOP_LIMIT htobe64(1 << 9) > +#define IB_COMP_MASK_PR_TCLASS htobe64(1 << 10) > +#define IB_COMP_MASK_PR_REVERSIBLE htobe64(1 << 11) > +#define IB_COMP_MASK_PR_NUM_PATH htobe64(1 << 12) > +#define IB_COMP_MASK_PR_PKEY htobe64(1 << 13) > +#define IB_COMP_MASK_PR_QOS_CLASS htobe64(1 << 14) > +#define IB_COMP_MASK_PR_SL htobe64(1 << 15) > +#define IB_COMP_MASK_PR_MTU_SELECTOR htobe64(1 << 16) > +#define IB_COMP_MASK_PR_MTU htobe64(1 << 17) > +#define IB_COMP_MASK_PR_RATE_SELECTOR htobe64(1 << 18) > +#define IB_COMP_MASK_PR_RATE htobe64(1 << 19) > +#define IB_COMP_MASK_PR_PACKET_LIFETIME_SELECTOR htobe64(1 << 20) > +#define IB_COMP_MASK_PR_PACKET_LIFETIME htobe64(1 << 21) > +#define IB_COMP_MASK_PR_PREFERENCE htobe64(1 << 22) > +/* RESERVED htobe64(1 << 23) */ > > #define IB_MC_QPN 0xffffff > #define IB_SA_ATTR_MC_MEMBER_REC htons(0x0038) > > -#define IB_COMP_MASK_MC_MGID htonll(1 << 0) > -#define IB_COMP_MASK_MC_PORT_GID htonll(1 << 1) > -#define IB_COMP_MASK_MC_QKEY htonll(1 << 2) > -#define IB_COMP_MASK_MC_MLID htonll(1 << 3) > -#define IB_COMP_MASK_MC_MTU_SEL htonll(1 << 4) > -#define IB_COMP_MASK_MC_MTU htonll(1 << 5) > -#define IB_COMP_MASK_MC_TCLASS htonll(1 << 6) > -#define IB_COMP_MASK_MC_PKEY htonll(1 << 7) > -#define IB_COMP_MASK_MC_RATE_SEL htonll(1 << 8) > -#define IB_COMP_MASK_MC_RATE htonll(1 << 9) > -#define IB_COMP_MASK_MC_PACKET_LIFETIME_SEL htonll(1 << 10) > -#define IB_COMP_MASK_MC_PACKET_LIFETIME htonll(1 << 11) > -#define IB_COMP_MASK_MC_SL htonll(1 << 12) > -#define IB_COMP_MASK_MC_FLOW htonll(1 << 13) > -#define IB_COMP_MASK_MC_HOP htonll(1 << 14) > -#define IB_COMP_MASK_MC_SCOPE htonll(1 << 15) > -#define IB_COMP_MASK_MC_JOIN_STATE htonll(1 << 16) > -#define IB_COMP_MASK_MC_PROXY_JOIN htonll(1 << 17) > +#define IB_COMP_MASK_MC_MGID htobe64(1 << 0) > +#define IB_COMP_MASK_MC_PORT_GID htobe64(1 << 1) > +#define IB_COMP_MASK_MC_QKEY htobe64(1 << 2) > +#define IB_COMP_MASK_MC_MLID htobe64(1 << 3) > +#define IB_COMP_MASK_MC_MTU_SEL htobe64(1 << 4) > +#define IB_COMP_MASK_MC_MTU htobe64(1 << 5) > +#define IB_COMP_MASK_MC_TCLASS htobe64(1 << 6) > +#define IB_COMP_MASK_MC_PKEY htobe64(1 << 7) > +#define IB_COMP_MASK_MC_RATE_SEL htobe64(1 << 8) > +#define IB_COMP_MASK_MC_RATE htobe64(1 << 9) > +#define IB_COMP_MASK_MC_PACKET_LIFETIME_SEL htobe64(1 << 10) > +#define IB_COMP_MASK_MC_PACKET_LIFETIME htobe64(1 << 11) > +#define IB_COMP_MASK_MC_SL htobe64(1 << 12) > +#define IB_COMP_MASK_MC_FLOW htobe64(1 << 13) > +#define IB_COMP_MASK_MC_HOP htobe64(1 << 14) > +#define IB_COMP_MASK_MC_SCOPE htobe64(1 << 15) > +#define IB_COMP_MASK_MC_JOIN_STATE htobe64(1 << 16) > +#define IB_COMP_MASK_MC_PROXY_JOIN htobe64(1 << 17) > > struct ib_mc_member_rec { > union ibv_gid mgid; > diff --git a/ibacm/prov/acmp/src/acmp.c b/ibacm/prov/acmp/src/acmp.c > index fb6684f2eac1f5..6b7cbb8c203478 100644 > --- a/ibacm/prov/acmp/src/acmp.c > +++ b/ibacm/prov/acmp/src/acmp.c > @@ -780,7 +780,7 @@ static void acmp_init_path_query(struct ib_sa_mad *mad) > mad->mgmt_class = IB_MGMT_CLASS_SA; > mad->class_version = 2; > mad->method = IB_METHOD_GET; > - mad->tid = htonll((uint64_t) atomic_inc(&g_tid)); > + mad->tid = htobe64((uint64_t) atomic_inc(&g_tid)); > mad->attr_id = IB_SA_ATTR_PATH_REC; > } > > @@ -1367,7 +1367,7 @@ static void acmp_init_join(struct ib_sa_mad *mad, union ibv_gid *port_gid, > mad->mgmt_class = IB_MGMT_CLASS_SA; > mad->class_version = 2; > mad->method = IB_METHOD_SET; > - mad->tid = htonll((uint64_t) atomic_inc(&g_tid)); > + mad->tid = htobe64((uint64_t) atomic_inc(&g_tid)); > mad->attr_id = IB_SA_ATTR_MC_MEMBER_REC; > mad->comp_mask = > IB_COMP_MASK_MC_MGID | IB_COMP_MASK_MC_PORT_GID | > @@ -1673,7 +1673,7 @@ acmp_send_resolve(struct acmp_ep *ep, struct acmp_dest *dest, > mad->class_version = 1; > mad->method = IB_METHOD_GET; > mad->control = ACM_CTRL_RESOLVE; > - mad->tid = htonll((uint64_t) atomic_inc(&g_tid)); > + mad->tid = htobe64((uint64_t) atomic_inc(&g_tid)); > > rec = (struct acm_resolve_rec *) mad->data; > rec->src_type = (uint8_t) saddr->type; > @@ -1957,7 +1957,7 @@ static void acmp_query_perf(void *ep_context, uint64_t *values, uint8_t *cnt) > int i; > > for (i = 0; i < ACM_MAX_COUNTER; i++) > - values[i] = htonll((uint64_t) atomic_get(&ep->counters[i])); > + values[i] = htobe64((uint64_t) atomic_get(&ep->counters[i])); > *cnt = ACM_MAX_COUNTER; > } > > @@ -2082,7 +2082,7 @@ static void acmp_parse_osm_fullv1_lid2guid(FILE *f, uint64_t *lid2guid) > if (lid2guid[lid]) > acm_log(0, "ERROR - duplicate lid %u\n", lid); > else > - lid2guid[lid] = htonll(guid); > + lid2guid[lid] = htobe64(guid); > } > } > > diff --git a/ibacm/src/acm.c b/ibacm/src/acm.c > index fd06a2e81fbe76..2cf1ea8b2148c0 100644 > --- a/ibacm/src/acm.c > +++ b/ibacm/src/acm.c > @@ -1000,7 +1000,7 @@ static int acm_svr_perf_query(struct acmc_client *client, struct acm_msg *msg) > ((ntohs(msg->hdr.length) >= (ACM_MSG_HDR_LENGTH + ACM_MSG_EP_LENGTH) > && !(msg->resolve_data[0].flags & ACM_EP_FLAG_SOURCE)))) { > for (i = 0; i < ACM_MAX_COUNTER; i++) > - msg->perf_data[i] = htonll((uint64_t) atomic_get(&counter[i])); > + msg->perf_data[i] = htobe64((uint64_t) atomic_get(&counter[i])); > > msg->hdr.data[0] = ACM_MAX_COUNTER; > len = ACM_MSG_HDR_LENGTH + (ACM_MAX_COUNTER * sizeof(uint64_t)); > @@ -1426,7 +1426,7 @@ static int acm_nl_parse_path_attr(struct nlattr *attr, > sid = (uint64_t *) NLA_DATA(attr); > if (NLA_LEN(attr) == sizeof(*sid)) { > acm_log(2, "service_id 0x%" PRIx64 "\n", *sid); > - path->service_id = htonll(*sid); > + path->service_id = htobe64(*sid); > } else { > ret = -1; > } > @@ -2499,7 +2499,7 @@ static void acm_load_prov_config(void) > acm_log(2, "provider %s subnet_prefix 0x%" PRIx64 "\n", > prov_name, prefix); > /* Convert it into network byte order */ > - prefix = htonll(prefix); > + prefix = htobe64(prefix); > > list_for_each(&provider_list, prov, entry) { > if (!strcasecmp(prov->prov->name, prov_name)) { > diff --git a/providers/cxgb3/cxio_wr.h b/providers/cxgb3/cxio_wr.h > index d56c5298ebaad5..735b64918a15c8 100644 > --- a/providers/cxgb3/cxio_wr.h > +++ b/providers/cxgb3/cxio_wr.h > @@ -351,7 +351,7 @@ static inline void build_fw_riwrh(struct fw_riwrh *wqe, enum t3_wr_opcode op, > wqe->gen_tid_len = htonl(V_FW_RIWR_GEN(genbit) | V_FW_RIWR_TID(tid) | > V_FW_RIWR_LEN(len)); > /* 2nd gen bit... */ > - ((union t3_wr *)wqe)->flit[15] = htonll(genbit); > + ((union t3_wr *)wqe)->flit[15] = htobe64(genbit); > } > > /* > diff --git a/providers/cxgb3/qp.c b/providers/cxgb3/qp.c > index 30dd898f6e874b..b0cf10364e9a6c 100644 > --- a/providers/cxgb3/qp.c > +++ b/providers/cxgb3/qp.c > @@ -84,7 +84,7 @@ static inline int iwch_build_rdma_send(union t3_wr *wqe, struct ibv_send_wr *wr, > htonl(wr->sg_list[i].lkey); > wqe->send.sgl[i].len = > htonl(wr->sg_list[i].length); > - wqe->send.sgl[i].to = htonll(wr->sg_list[i].addr); > + wqe->send.sgl[i].to = htobe64(wr->sg_list[i].addr); > } > wqe->send.plen = htonl(wqe->send.plen); > wqe->send.num_sgle = htonl(wr->num_sge); > @@ -104,7 +104,7 @@ static inline int iwch_build_rdma_write(union t3_wr *wqe, > wqe->write.rdmaop = T3_RDMA_WRITE; > wqe->write.reserved = 0; > wqe->write.stag_sink = htonl(wr->wr.rdma.rkey); > - wqe->write.to_sink = htonll(wr->wr.rdma.remote_addr); > + wqe->write.to_sink = htobe64(wr->wr.rdma.remote_addr); > > wqe->write.num_sgle = wr->num_sge; > > @@ -139,7 +139,7 @@ static inline int iwch_build_rdma_write(union t3_wr *wqe, > wqe->write.sgl[i].len = > htonl(wr->sg_list[i].length); > wqe->write.sgl[i].to = > - htonll(wr->sg_list[i].addr); > + htobe64(wr->sg_list[i].addr); > } > wqe->write.plen = htonl(wqe->write.plen); > wqe->write.num_sgle = htonl(wr->num_sge); > @@ -157,10 +157,10 @@ static inline int iwch_build_rdma_read(union t3_wr *wqe, struct ibv_send_wr *wr, > wqe->read.reserved = 0; > if (wr->num_sge == 1 && wr->sg_list[0].length > 0) { > wqe->read.rem_stag = htonl(wr->wr.rdma.rkey); > - wqe->read.rem_to = htonll(wr->wr.rdma.remote_addr); > + wqe->read.rem_to = htobe64(wr->wr.rdma.remote_addr); > wqe->read.local_stag = htonl(wr->sg_list[0].lkey); > wqe->read.local_len = htonl(wr->sg_list[0].length); > - wqe->read.local_to = htonll(wr->sg_list[0].addr); > + wqe->read.local_to = htobe64(wr->sg_list[0].addr); > } else { > > /* build passable 0B read request */ > @@ -295,7 +295,7 @@ static inline int iwch_build_rdma_recv(struct iwch_device *rhp, > for (i = 0; i < wr->num_sge; i++) { > wqe->recv.sgl[i].stag = htonl(wr->sg_list[i].lkey); > wqe->recv.sgl[i].len = htonl(wr->sg_list[i].length); > - wqe->recv.sgl[i].to = htonll(wr->sg_list[i].addr); > + wqe->recv.sgl[i].to = htobe64(wr->sg_list[i].addr); > } > for (; i < T3_MAX_SGE; i++) { > wqe->recv.sgl[i].stag = 0; > diff --git a/providers/cxgb4/t4.h b/providers/cxgb4/t4.h > index cad675c3fdb193..59dbc65d4e164b 100644 > --- a/providers/cxgb4/t4.h > +++ b/providers/cxgb4/t4.h > @@ -55,7 +55,7 @@ > #define __iomem > #define cpu_to_be16 htons > #define cpu_to_be32 htonl > -#define cpu_to_be64 htonll > +#define cpu_to_be64 htobe64 > #define be16_to_cpu ntohs > #define be32_to_cpu ntohl > #define be64_to_cpu ntohll > diff --git a/providers/mlx4/qp.c b/providers/mlx4/qp.c > index 268fb7dc83dd11..a607326c7c452c 100644 > --- a/providers/mlx4/qp.c > +++ b/providers/mlx4/qp.c > @@ -157,7 +157,7 @@ static inline void set_local_inv_seg(struct mlx4_wqe_local_inval_seg *iseg, > static inline void set_raddr_seg(struct mlx4_wqe_raddr_seg *rseg, > uint64_t remote_addr, uint32_t rkey) > { > - rseg->raddr = htonll(remote_addr); > + rseg->raddr = htobe64(remote_addr); > rseg->rkey = htonl(rkey); > rseg->reserved = 0; > } > @@ -165,10 +165,10 @@ static inline void set_raddr_seg(struct mlx4_wqe_raddr_seg *rseg, > static void set_atomic_seg(struct mlx4_wqe_atomic_seg *aseg, struct ibv_send_wr *wr) > { > if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) { > - aseg->swap_add = htonll(wr->wr.atomic.swap); > - aseg->compare = htonll(wr->wr.atomic.compare_add); > + aseg->swap_add = htobe64(wr->wr.atomic.swap); > + aseg->compare = htobe64(wr->wr.atomic.compare_add); > } else { > - aseg->swap_add = htonll(wr->wr.atomic.compare_add); > + aseg->swap_add = htobe64(wr->wr.atomic.compare_add); > aseg->compare = 0; > } > > @@ -188,13 +188,13 @@ static void __set_data_seg(struct mlx4_wqe_data_seg *dseg, struct ibv_sge *sg) > { > dseg->byte_count = htonl(sg->length); > dseg->lkey = htonl(sg->lkey); > - dseg->addr = htonll(sg->addr); > + dseg->addr = htobe64(sg->addr); > } > > static void set_data_seg(struct mlx4_wqe_data_seg *dseg, struct ibv_sge *sg) > { > dseg->lkey = htonl(sg->lkey); > - dseg->addr = htonll(sg->addr); > + dseg->addr = htobe64(sg->addr); > > /* > * Need a barrier here before writing the byte_count field to > diff --git a/providers/mlx4/srq.c b/providers/mlx4/srq.c > index c0e028671828d8..4f90efdf927209 100644 > --- a/providers/mlx4/srq.c > +++ b/providers/mlx4/srq.c > @@ -96,7 +96,7 @@ int mlx4_post_srq_recv(struct ibv_srq *ibsrq, > for (i = 0; i < wr->num_sge; ++i) { > scat[i].byte_count = htonl(wr->sg_list[i].length); > scat[i].lkey = htonl(wr->sg_list[i].lkey); > - scat[i].addr = htonll(wr->sg_list[i].addr); > + scat[i].addr = htobe64(wr->sg_list[i].addr); > } > > if (i < srq->max_gs) { > diff --git a/providers/mlx5/qp.c b/providers/mlx5/qp.c > index e82b1a0bebc3f9..7d034880f4caf6 100644 > --- a/providers/mlx5/qp.c > +++ b/providers/mlx5/qp.c > @@ -190,7 +190,7 @@ static int mlx5_wq_overflow(struct mlx5_wq *wq, int nreq, struct mlx5_cq *cq) > static inline void set_raddr_seg(struct mlx5_wqe_raddr_seg *rseg, > uint64_t remote_addr, uint32_t rkey) > { > - rseg->raddr = htonll(remote_addr); > + rseg->raddr = htobe64(remote_addr); > rseg->rkey = htonl(rkey); > rseg->reserved = 0; > } > @@ -201,10 +201,10 @@ static void set_atomic_seg(struct mlx5_wqe_atomic_seg *aseg, > uint64_t compare_add) > { > if (opcode == IBV_WR_ATOMIC_CMP_AND_SWP) { > - aseg->swap_add = htonll(swap); > - aseg->compare = htonll(compare_add); > + aseg->swap_add = htobe64(swap); > + aseg->compare = htobe64(compare_add); > } else { > - aseg->swap_add = htonll(compare_add); > + aseg->swap_add = htobe64(compare_add); > } > } > > @@ -221,7 +221,7 @@ static void set_data_ptr_seg(struct mlx5_wqe_data_seg *dseg, struct ibv_sge *sg, > { > dseg->byte_count = htonl(sg->length - offset); > dseg->lkey = htonl(sg->lkey); > - dseg->addr = htonll(sg->addr + offset); > + dseg->addr = htobe64(sg->addr + offset); > } > > static void set_data_ptr_seg_atomic(struct mlx5_wqe_data_seg *dseg, > @@ -229,7 +229,7 @@ static void set_data_ptr_seg_atomic(struct mlx5_wqe_data_seg *dseg, > { > dseg->byte_count = htonl(MLX5_ATOMIC_SIZE); > dseg->lkey = htonl(sg->lkey); > - dseg->addr = htonll(sg->addr); > + dseg->addr = htobe64(sg->addr); > } > > /* > @@ -430,7 +430,7 @@ static void set_umr_data_seg(struct mlx5_qp *qp, enum ibv_mw_type type, > > data->klm.byte_count = htonl(bind_info->length); > data->klm.mkey = htonl(bind_info->mr->lkey); > - data->klm.address = htonll(bind_info->addr); > + data->klm.address = htobe64(bind_info->addr); > > memset(&data->klm + 1, 0, sizeof(data->reserved) - > sizeof(data->klm)); > @@ -467,8 +467,8 @@ static void set_umr_mkey_seg(struct mlx5_qp *qp, enum ibv_mw_type type, > if (bind_info->mw_access_flags & IBV_ACCESS_ZERO_BASED) > mkey->start_addr = 0; > else > - mkey->start_addr = htonll(bind_info->addr); > - mkey->len = htonll(bind_info->length); > + mkey->start_addr = htobe64(bind_info->addr); > + mkey->len = htobe64(bind_info->length); > } else { > mkey->free = MLX5_WQE_MKEY_CONTEXT_FREE; > } > @@ -485,20 +485,20 @@ static inline void set_umr_control_seg(struct mlx5_qp *qp, enum ibv_mw_type type > > ctrl->flags = MLX5_WQE_UMR_CTRL_FLAG_TRNSLATION_OFFSET | > MLX5_WQE_UMR_CTRL_FLAG_INLINE; > - ctrl->mkey_mask = htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_FREE | > + ctrl->mkey_mask = htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_FREE | > MLX5_WQE_UMR_CTRL_MKEY_MASK_MKEY); > ctrl->translation_offset = 0; > memset(ctrl->rsvd0, 0, sizeof(ctrl->rsvd0)); > memset(ctrl->rsvd1, 0, sizeof(ctrl->rsvd1)); > > if (type == IBV_MW_TYPE_2) > - ctrl->mkey_mask |= htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_QPN); > + ctrl->mkey_mask |= htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_QPN); > > if (bind_info->length) { > ctrl->klm_octowords = get_klm_octo(1); > if (type == IBV_MW_TYPE_2) > ctrl->flags |= MLX5_WQE_UMR_CTRL_FLAG_CHECK_FREE; > - ctrl->mkey_mask |= htonll(MLX5_WQE_UMR_CTRL_MKEY_MASK_LEN | > + ctrl->mkey_mask |= htobe64(MLX5_WQE_UMR_CTRL_MKEY_MASK_LEN | > MLX5_WQE_UMR_CTRL_MKEY_MASK_START_ADDR | > MLX5_WQE_UMR_CTRL_MKEY_MASK_ACCESS_LOCAL_WRITE | > MLX5_WQE_UMR_CTRL_MKEY_MASK_ACCESS_REMOTE_READ | > diff --git a/providers/mlx5/srq.c b/providers/mlx5/srq.c > index a06afa3a5931b6..b362ec85724961 100644 > --- a/providers/mlx5/srq.c > +++ b/providers/mlx5/srq.c > @@ -120,7 +120,7 @@ int mlx5_post_srq_recv(struct ibv_srq *ibsrq, > for (i = 0; i < wr->num_sge; ++i) { > scat[i].byte_count = htonl(wr->sg_list[i].length); > scat[i].lkey = htonl(wr->sg_list[i].lkey); > - scat[i].addr = htonll(wr->sg_list[i].addr); > + scat[i].addr = htobe64(wr->sg_list[i].addr); > } > > if (i < srq->max_gs) { > diff --git a/providers/mthca/qp.c b/providers/mthca/qp.c > index 129e6c74ef5382..d221bb19bfa67c 100644 > --- a/providers/mthca/qp.c > +++ b/providers/mthca/qp.c > @@ -147,7 +147,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > case IBV_WR_ATOMIC_CMP_AND_SWP: > case IBV_WR_ATOMIC_FETCH_AND_ADD: > ((struct mthca_raddr_seg *) wqe)->raddr = > - htonll(wr->wr.atomic.remote_addr); > + htobe64(wr->wr.atomic.remote_addr); > ((struct mthca_raddr_seg *) wqe)->rkey = > htonl(wr->wr.atomic.rkey); > ((struct mthca_raddr_seg *) wqe)->reserved = 0; > @@ -156,12 +156,12 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > > if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) { > ((struct mthca_atomic_seg *) wqe)->swap_add = > - htonll(wr->wr.atomic.swap); > + htobe64(wr->wr.atomic.swap); > ((struct mthca_atomic_seg *) wqe)->compare = > - htonll(wr->wr.atomic.compare_add); > + htobe64(wr->wr.atomic.compare_add); > } else { > ((struct mthca_atomic_seg *) wqe)->swap_add = > - htonll(wr->wr.atomic.compare_add); > + htobe64(wr->wr.atomic.compare_add); > ((struct mthca_atomic_seg *) wqe)->compare = 0; > } > > @@ -174,7 +174,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > case IBV_WR_RDMA_WRITE_WITH_IMM: > case IBV_WR_RDMA_READ: > ((struct mthca_raddr_seg *) wqe)->raddr = > - htonll(wr->wr.rdma.remote_addr); > + htobe64(wr->wr.rdma.remote_addr); > ((struct mthca_raddr_seg *) wqe)->rkey = > htonl(wr->wr.rdma.rkey); > ((struct mthca_raddr_seg *) wqe)->reserved = 0; > @@ -194,7 +194,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > case IBV_WR_RDMA_WRITE: > case IBV_WR_RDMA_WRITE_WITH_IMM: > ((struct mthca_raddr_seg *) wqe)->raddr = > - htonll(wr->wr.rdma.remote_addr); > + htobe64(wr->wr.rdma.remote_addr); > ((struct mthca_raddr_seg *) wqe)->rkey = > htonl(wr->wr.rdma.rkey); > ((struct mthca_raddr_seg *) wqe)->reserved = 0; > @@ -213,7 +213,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > ((struct mthca_tavor_ud_seg *) wqe)->lkey = > htonl(to_mah(wr->wr.ud.ah)->key); > ((struct mthca_tavor_ud_seg *) wqe)->av_addr = > - htonll((uintptr_t) to_mah(wr->wr.ud.ah)->av); > + htobe64((uintptr_t) to_mah(wr->wr.ud.ah)->av); > ((struct mthca_tavor_ud_seg *) wqe)->dqpn = > htonl(wr->wr.ud.remote_qpn); > ((struct mthca_tavor_ud_seg *) wqe)->qkey = > @@ -265,7 +265,7 @@ int mthca_tavor_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > seg = wqe; > seg->byte_count = htonl(wr->sg_list[i].length); > seg->lkey = htonl(wr->sg_list[i].lkey); > - seg->addr = htonll(wr->sg_list[i].addr); > + seg->addr = htobe64(wr->sg_list[i].addr); > wqe += sizeof *seg; > } > > @@ -372,7 +372,7 @@ int mthca_tavor_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr, > ((struct mthca_data_seg *) wqe)->lkey = > htonl(wr->sg_list[i].lkey); > ((struct mthca_data_seg *) wqe)->addr = > - htonll(wr->sg_list[i].addr); > + htobe64(wr->sg_list[i].addr); > wqe += sizeof (struct mthca_data_seg); > size += sizeof (struct mthca_data_seg) / 16; > } > @@ -508,7 +508,7 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > case IBV_WR_ATOMIC_CMP_AND_SWP: > case IBV_WR_ATOMIC_FETCH_AND_ADD: > ((struct mthca_raddr_seg *) wqe)->raddr = > - htonll(wr->wr.atomic.remote_addr); > + htobe64(wr->wr.atomic.remote_addr); > ((struct mthca_raddr_seg *) wqe)->rkey = > htonl(wr->wr.atomic.rkey); > ((struct mthca_raddr_seg *) wqe)->reserved = 0; > @@ -517,12 +517,12 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > > if (wr->opcode == IBV_WR_ATOMIC_CMP_AND_SWP) { > ((struct mthca_atomic_seg *) wqe)->swap_add = > - htonll(wr->wr.atomic.swap); > + htobe64(wr->wr.atomic.swap); > ((struct mthca_atomic_seg *) wqe)->compare = > - htonll(wr->wr.atomic.compare_add); > + htobe64(wr->wr.atomic.compare_add); > } else { > ((struct mthca_atomic_seg *) wqe)->swap_add = > - htonll(wr->wr.atomic.compare_add); > + htobe64(wr->wr.atomic.compare_add); > ((struct mthca_atomic_seg *) wqe)->compare = 0; > } > > @@ -535,7 +535,7 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > case IBV_WR_RDMA_WRITE_WITH_IMM: > case IBV_WR_RDMA_READ: > ((struct mthca_raddr_seg *) wqe)->raddr = > - htonll(wr->wr.rdma.remote_addr); > + htobe64(wr->wr.rdma.remote_addr); > ((struct mthca_raddr_seg *) wqe)->rkey = > htonl(wr->wr.rdma.rkey); > ((struct mthca_raddr_seg *) wqe)->reserved = 0; > @@ -555,7 +555,7 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > case IBV_WR_RDMA_WRITE: > case IBV_WR_RDMA_WRITE_WITH_IMM: > ((struct mthca_raddr_seg *) wqe)->raddr = > - htonll(wr->wr.rdma.remote_addr); > + htobe64(wr->wr.rdma.remote_addr); > ((struct mthca_raddr_seg *) wqe)->rkey = > htonl(wr->wr.rdma.rkey); > ((struct mthca_raddr_seg *) wqe)->reserved = 0; > @@ -624,7 +624,7 @@ int mthca_arbel_post_send(struct ibv_qp *ibqp, struct ibv_send_wr *wr, > seg = wqe; > seg->byte_count = htonl(wr->sg_list[i].length); > seg->lkey = htonl(wr->sg_list[i].lkey); > - seg->addr = htonll(wr->sg_list[i].addr); > + seg->addr = htobe64(wr->sg_list[i].addr); > wqe += sizeof *seg; > } > > @@ -730,7 +730,7 @@ int mthca_arbel_post_recv(struct ibv_qp *ibqp, struct ibv_recv_wr *wr, > ((struct mthca_data_seg *) wqe)->lkey = > htonl(wr->sg_list[i].lkey); > ((struct mthca_data_seg *) wqe)->addr = > - htonll(wr->sg_list[i].addr); > + htobe64(wr->sg_list[i].addr); > wqe += sizeof (struct mthca_data_seg); > } > > diff --git a/providers/mthca/srq.c b/providers/mthca/srq.c > index ff6f532733f4ae..66ac924a720c84 100644 > --- a/providers/mthca/srq.c > +++ b/providers/mthca/srq.c > @@ -126,7 +126,7 @@ int mthca_tavor_post_srq_recv(struct ibv_srq *ibsrq, > ((struct mthca_data_seg *) wqe)->lkey = > htonl(wr->sg_list[i].lkey); > ((struct mthca_data_seg *) wqe)->addr = > - htonll(wr->sg_list[i].addr); > + htobe64(wr->sg_list[i].addr); > wqe += sizeof (struct mthca_data_seg); > } > > @@ -219,7 +219,7 @@ int mthca_arbel_post_srq_recv(struct ibv_srq *ibsrq, > ((struct mthca_data_seg *) wqe)->lkey = > htonl(wr->sg_list[i].lkey); > ((struct mthca_data_seg *) wqe)->addr = > - htonll(wr->sg_list[i].addr); > + htobe64(wr->sg_list[i].addr); > wqe += sizeof (struct mthca_data_seg); > } > > diff --git a/providers/qedr/qelr_verbs.c b/providers/qedr/qelr_verbs.c > index 823ed8754c09d7..932115c8ab7f39 100644 > --- a/providers/qedr/qelr_verbs.c > +++ b/providers/qedr/qelr_verbs.c > @@ -900,7 +900,7 @@ static inline void qelr_edpm_set_rdma_ext(struct qelr_qp *qp, > if (!qp->edpm.is_edpm) > return; > > - qp->edpm.rdma_ext->remote_va = htonll(remote_addr); > + qp->edpm.rdma_ext->remote_va = htobe64(remote_addr); > qp->edpm.rdma_ext->remote_key = htonl(rkey); > qp->edpm.dpm_payload_offset += sizeof(*qp->edpm.rdma_ext); > qp->edpm.dpm_payload_size += sizeof(*qp->edpm.rdma_ext); > diff --git a/srp_daemon/srp_daemon.c b/srp_daemon/srp_daemon.c > index e0e550d81ab3c2..ff6118acf611ca 100644 > --- a/srp_daemon/srp_daemon.c > +++ b/srp_daemon/srp_daemon.c > @@ -462,9 +462,9 @@ static int add_non_exist_target(struct target_details *target) > if (srpd_sys_read_gid(scsi_host_dir, "dgid", dgid_val)) > continue; > } > - if (htonll(target->subnet_prefix) != *((uint64_t *) dgid_val)) > + if (htobe64(target->subnet_prefix) != *((uint64_t *) dgid_val)) > continue; > - if (htonll(target->h_guid) != *((uint64_t *) (dgid_val+8))) > + if (htobe64(target->h_guid) != *((uint64_t *) (dgid_val+8))) > continue; > > /* If there is no local_ib_device in the scsi host dir (old kernel module), assumes it is equal */ > @@ -603,7 +603,7 @@ static int send_and_get(int portid, int agent, srp_ib_user_mad_t *out_mad, > /* Skip tid 0 because OpenSM ignores it. */ > if (++tid == 0) > ++tid; > - out_dm_mad->tid = htonll(tid); > + out_dm_mad->tid = htobe64(tid); > > ret = umad_send(portid, agent, out_mad, MAD_BLOCK_SIZE, > config->timeout, 0); > @@ -1033,7 +1033,7 @@ int get_node(struct umad_resources *umad_res, uint16_t dlid, uint64_t *guid) > init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid, > SRP_SA_ATTR_NODE, 0); > > - out_sa_mad->comp_mask = htonll(1); /* LID */ > + out_sa_mad->comp_mask = htobe64(1); /* LID */ > node = (void *) out_sa_mad->data; > node->lid = htons(dlid); > > @@ -1059,7 +1059,7 @@ static int get_port_info(struct umad_resources *umad_res, uint16_t dlid, > init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid, > SRP_SA_ATTR_PORT_INFO, 0); > > - out_sa_mad->comp_mask = htonll(1); /* LID */ > + out_sa_mad->comp_mask = htobe64(1); /* LID */ > port_info = (void *) out_sa_mad->data; > port_info->endport_lid = htons(dlid); > > @@ -1134,7 +1134,7 @@ static int get_shared_pkeys(struct resources *res, > continue; > > /* Mark components: DLID, SLID, PKEY */ > - out_sa_mad->comp_mask = htonll(1 << 4 | 1 << 5 | 1 << 13); > + out_sa_mad->comp_mask = htobe64(1 << 4 | 1 << 5 | 1 << 13); > out_sa_mad->rmpp_version = 1; > out_sa_mad->rmpp_type = 1; > path_rec = (ib_path_rec_t *)out_sa_mad->data; > @@ -1194,7 +1194,7 @@ static int do_dm_port_list(struct resources *res) > SRP_SA_ATTR_PORT_INFO, SRP_SM_CAP_MASK_MATCH_ATTR_MOD); > > out_sa_mad->method = SRP_SA_METHOD_GET_TABLE; > - out_sa_mad->comp_mask = htonll(1 << 7); /* Capability mask */ > + out_sa_mad->comp_mask = htobe64(1 << 7); /* Capability mask */ > out_sa_mad->rmpp_version = 1; > out_sa_mad->rmpp_type = 1; > port_info = (void *) out_sa_mad->data; > @@ -2330,7 +2330,7 @@ static int get_lid(struct umad_resources *umad_res, ib_gid_t *gid, uint16_t *lid > init_srp_sa_mad(&out_mad, umad_res->agent, umad_res->sm_lid, > SRP_SA_ATTR_PATH_REC, 0); > > - out_sa_mad->comp_mask = htonll( 4 | 8 | 64 | 512 | 4096 ); > + out_sa_mad->comp_mask = htobe64( 4 | 8 | 64 | 512 | 4096 ); > > path_rec->sgid = *gid; > path_rec->dgid = *gid; > diff --git a/srp_daemon/srp_handle_traps.c b/srp_daemon/srp_handle_traps.c > index 01cc5445e8352b..6a7a2d1ab479a1 100644 > --- a/srp_daemon/srp_handle_traps.c > +++ b/srp_daemon/srp_handle_traps.c > @@ -606,14 +606,14 @@ static int register_to_trap(struct sync_resources *sync_res, > comp_mask |= SRP_INFORMINFO_QPN_COMP; > } > > - p_sa_mad->comp_mask = htonll(comp_mask); > + p_sa_mad->comp_mask = htobe64(comp_mask); > pr_debug("comp_mask: %llx\n", comp_mask); > > do { > pthread_mutex_lock(res->mad_buffer_mutex); > res->mad_buffer->base_ver = 0; // flag that the buffer is empty > pthread_mutex_unlock(res->mad_buffer_mutex); > - mad_hdr->trans_id = htonll(trans_id); > + mad_hdr->trans_id = htobe64(trans_id); > trans_id++; > > ret = ibv_post_send(res->qp, &sr, bad_wr); > -- > 2.7.4 > Looks good. Reviewed-by: Ram Amrani <Ram.Amrani@xxxxxxxxxx> -- To unsubscribe from this list: send the line "unsubscribe linux-rdma" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html