blob: 4f1fb0cefe516d45c0b47ac4029c62f2badec5ba [file] [log] [blame]
// SPDX-License-Identifier: GPL-2.0-only
/* Copyright (C) 2024 Intel Corporation */
#include "idpf.h"
#include "idpf_ptp.h"
#include "idpf_virtchnl.h"
/**
* idpf_ptp_get_caps - Send virtchnl get ptp capabilities message
* @adapter: Driver specific private structure
*
* Send virtchnl get PTP capabilities message.
*
* Return: 0 on success, -errno on failure.
*/
int idpf_ptp_get_caps(struct idpf_adapter *adapter)
{
struct virtchnl2_ptp_get_caps *recv_ptp_caps_msg __free(kfree) = NULL;
struct virtchnl2_ptp_get_caps send_ptp_caps_msg = {
.caps = cpu_to_le32(VIRTCHNL2_CAP_PTP_GET_DEVICE_CLK_TIME |
VIRTCHNL2_CAP_PTP_GET_DEVICE_CLK_TIME_MB |
VIRTCHNL2_CAP_PTP_GET_CROSS_TIME |
VIRTCHNL2_CAP_PTP_SET_DEVICE_CLK_TIME_MB |
VIRTCHNL2_CAP_PTP_ADJ_DEVICE_CLK_MB |
VIRTCHNL2_CAP_PTP_TX_TSTAMPS_MB)
};
struct idpf_vc_xn_params xn_params = {
.vc_op = VIRTCHNL2_OP_PTP_GET_CAPS,
.send_buf.iov_base = &send_ptp_caps_msg,
.send_buf.iov_len = sizeof(send_ptp_caps_msg),
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
};
struct virtchnl2_ptp_cross_time_reg_offsets cross_tstamp_offsets;
struct virtchnl2_ptp_clk_adj_reg_offsets clk_adj_offsets;
struct virtchnl2_ptp_clk_reg_offsets clock_offsets;
struct idpf_ptp_secondary_mbx *scnd_mbx;
struct idpf_ptp *ptp = adapter->ptp;
enum idpf_ptp_access access_type;
u32 temp_offset;
int reply_sz;
recv_ptp_caps_msg = kzalloc(sizeof(struct virtchnl2_ptp_get_caps),
GFP_KERNEL);
if (!recv_ptp_caps_msg)
return -ENOMEM;
xn_params.recv_buf.iov_base = recv_ptp_caps_msg;
xn_params.recv_buf.iov_len = sizeof(*recv_ptp_caps_msg);
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
if (reply_sz < 0)
return reply_sz;
else if (reply_sz != sizeof(*recv_ptp_caps_msg))
return -EIO;
ptp->caps = le32_to_cpu(recv_ptp_caps_msg->caps);
ptp->base_incval = le64_to_cpu(recv_ptp_caps_msg->base_incval);
ptp->max_adj = le32_to_cpu(recv_ptp_caps_msg->max_adj);
scnd_mbx = &ptp->secondary_mbx;
scnd_mbx->peer_mbx_q_id = le16_to_cpu(recv_ptp_caps_msg->peer_mbx_q_id);
/* if the ptp_mb_q_id holds invalid value (0xffff), the secondary
* mailbox is not supported.
*/
scnd_mbx->valid = scnd_mbx->peer_mbx_q_id != 0xffff;
if (scnd_mbx->valid)
scnd_mbx->peer_id = recv_ptp_caps_msg->peer_id;
/* Determine the access type for the PTP features */
idpf_ptp_get_features_access(adapter);
access_type = ptp->get_dev_clk_time_access;
if (access_type != IDPF_PTP_DIRECT)
goto cross_tstamp;
clock_offsets = recv_ptp_caps_msg->clk_offsets;
temp_offset = le32_to_cpu(clock_offsets.dev_clk_ns_l);
ptp->dev_clk_regs.dev_clk_ns_l = idpf_get_reg_addr(adapter,
temp_offset);
temp_offset = le32_to_cpu(clock_offsets.dev_clk_ns_h);
ptp->dev_clk_regs.dev_clk_ns_h = idpf_get_reg_addr(adapter,
temp_offset);
temp_offset = le32_to_cpu(clock_offsets.phy_clk_ns_l);
ptp->dev_clk_regs.phy_clk_ns_l = idpf_get_reg_addr(adapter,
temp_offset);
temp_offset = le32_to_cpu(clock_offsets.phy_clk_ns_h);
ptp->dev_clk_regs.phy_clk_ns_h = idpf_get_reg_addr(adapter,
temp_offset);
temp_offset = le32_to_cpu(clock_offsets.cmd_sync_trigger);
ptp->dev_clk_regs.cmd_sync = idpf_get_reg_addr(adapter, temp_offset);
cross_tstamp:
access_type = ptp->get_cross_tstamp_access;
if (access_type != IDPF_PTP_DIRECT)
goto discipline_clock;
cross_tstamp_offsets = recv_ptp_caps_msg->cross_time_offsets;
temp_offset = le32_to_cpu(cross_tstamp_offsets.sys_time_ns_l);
ptp->dev_clk_regs.sys_time_ns_l = idpf_get_reg_addr(adapter,
temp_offset);
temp_offset = le32_to_cpu(cross_tstamp_offsets.sys_time_ns_h);
ptp->dev_clk_regs.sys_time_ns_h = idpf_get_reg_addr(adapter,
temp_offset);
temp_offset = le32_to_cpu(cross_tstamp_offsets.cmd_sync_trigger);
ptp->dev_clk_regs.cmd_sync = idpf_get_reg_addr(adapter, temp_offset);
discipline_clock:
access_type = ptp->adj_dev_clk_time_access;
if (access_type != IDPF_PTP_DIRECT)
return 0;
clk_adj_offsets = recv_ptp_caps_msg->clk_adj_offsets;
/* Device clock offsets */
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_cmd_type);
ptp->dev_clk_regs.cmd = idpf_get_reg_addr(adapter, temp_offset);
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_incval_l);
ptp->dev_clk_regs.incval_l = idpf_get_reg_addr(adapter, temp_offset);
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_incval_h);
ptp->dev_clk_regs.incval_h = idpf_get_reg_addr(adapter, temp_offset);
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_shadj_l);
ptp->dev_clk_regs.shadj_l = idpf_get_reg_addr(adapter, temp_offset);
temp_offset = le32_to_cpu(clk_adj_offsets.dev_clk_shadj_h);
ptp->dev_clk_regs.shadj_h = idpf_get_reg_addr(adapter, temp_offset);
/* PHY clock offsets */
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_cmd_type);
ptp->dev_clk_regs.phy_cmd = idpf_get_reg_addr(adapter, temp_offset);
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_incval_l);
ptp->dev_clk_regs.phy_incval_l = idpf_get_reg_addr(adapter,
temp_offset);
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_incval_h);
ptp->dev_clk_regs.phy_incval_h = idpf_get_reg_addr(adapter,
temp_offset);
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_shadj_l);
ptp->dev_clk_regs.phy_shadj_l = idpf_get_reg_addr(adapter, temp_offset);
temp_offset = le32_to_cpu(clk_adj_offsets.phy_clk_shadj_h);
ptp->dev_clk_regs.phy_shadj_h = idpf_get_reg_addr(adapter, temp_offset);
return 0;
}
/**
* idpf_ptp_get_dev_clk_time - Send virtchnl get device clk time message
* @adapter: Driver specific private structure
* @dev_clk_time: Pointer to the device clock structure where the value is set
*
* Send virtchnl get time message to get the time of the clock.
*
* Return: 0 on success, -errno otherwise.
*/
int idpf_ptp_get_dev_clk_time(struct idpf_adapter *adapter,
struct idpf_ptp_dev_timers *dev_clk_time)
{
struct virtchnl2_ptp_get_dev_clk_time get_dev_clk_time_msg;
struct idpf_vc_xn_params xn_params = {
.vc_op = VIRTCHNL2_OP_PTP_GET_DEV_CLK_TIME,
.send_buf.iov_base = &get_dev_clk_time_msg,
.send_buf.iov_len = sizeof(get_dev_clk_time_msg),
.recv_buf.iov_base = &get_dev_clk_time_msg,
.recv_buf.iov_len = sizeof(get_dev_clk_time_msg),
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
};
int reply_sz;
u64 dev_time;
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
if (reply_sz < 0)
return reply_sz;
if (reply_sz != sizeof(get_dev_clk_time_msg))
return -EIO;
dev_time = le64_to_cpu(get_dev_clk_time_msg.dev_time_ns);
dev_clk_time->dev_clk_time_ns = dev_time;
return 0;
}
/**
* idpf_ptp_get_cross_time - Send virtchnl get cross time message
* @adapter: Driver specific private structure
* @cross_time: Pointer to the device clock structure where the value is set
*
* Send virtchnl get cross time message to get the time of the clock and the
* system time.
*
* Return: 0 on success, -errno otherwise.
*/
int idpf_ptp_get_cross_time(struct idpf_adapter *adapter,
struct idpf_ptp_dev_timers *cross_time)
{
struct virtchnl2_ptp_get_cross_time cross_time_msg;
struct idpf_vc_xn_params xn_params = {
.vc_op = VIRTCHNL2_OP_PTP_GET_CROSS_TIME,
.send_buf.iov_base = &cross_time_msg,
.send_buf.iov_len = sizeof(cross_time_msg),
.recv_buf.iov_base = &cross_time_msg,
.recv_buf.iov_len = sizeof(cross_time_msg),
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
};
int reply_sz;
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
if (reply_sz < 0)
return reply_sz;
if (reply_sz != sizeof(cross_time_msg))
return -EIO;
cross_time->dev_clk_time_ns = le64_to_cpu(cross_time_msg.dev_time_ns);
cross_time->sys_time_ns = le64_to_cpu(cross_time_msg.sys_time_ns);
return 0;
}
/**
* idpf_ptp_set_dev_clk_time - Send virtchnl set device time message
* @adapter: Driver specific private structure
* @time: New time value
*
* Send virtchnl set time message to set the time of the clock.
*
* Return: 0 on success, -errno otherwise.
*/
int idpf_ptp_set_dev_clk_time(struct idpf_adapter *adapter, u64 time)
{
struct virtchnl2_ptp_set_dev_clk_time set_dev_clk_time_msg = {
.dev_time_ns = cpu_to_le64(time),
};
struct idpf_vc_xn_params xn_params = {
.vc_op = VIRTCHNL2_OP_PTP_SET_DEV_CLK_TIME,
.send_buf.iov_base = &set_dev_clk_time_msg,
.send_buf.iov_len = sizeof(set_dev_clk_time_msg),
.recv_buf.iov_base = &set_dev_clk_time_msg,
.recv_buf.iov_len = sizeof(set_dev_clk_time_msg),
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
};
int reply_sz;
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
if (reply_sz < 0)
return reply_sz;
if (reply_sz != sizeof(set_dev_clk_time_msg))
return -EIO;
return 0;
}
/**
* idpf_ptp_adj_dev_clk_time - Send virtchnl adj device clock time message
* @adapter: Driver specific private structure
* @delta: Offset in nanoseconds to adjust the time by
*
* Send virtchnl adj time message to adjust the clock by the indicated delta.
*
* Return: 0 on success, -errno otherwise.
*/
int idpf_ptp_adj_dev_clk_time(struct idpf_adapter *adapter, s64 delta)
{
struct virtchnl2_ptp_adj_dev_clk_time adj_dev_clk_time_msg = {
.delta = cpu_to_le64(delta),
};
struct idpf_vc_xn_params xn_params = {
.vc_op = VIRTCHNL2_OP_PTP_ADJ_DEV_CLK_TIME,
.send_buf.iov_base = &adj_dev_clk_time_msg,
.send_buf.iov_len = sizeof(adj_dev_clk_time_msg),
.recv_buf.iov_base = &adj_dev_clk_time_msg,
.recv_buf.iov_len = sizeof(adj_dev_clk_time_msg),
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
};
int reply_sz;
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
if (reply_sz < 0)
return reply_sz;
if (reply_sz != sizeof(adj_dev_clk_time_msg))
return -EIO;
return 0;
}
/**
* idpf_ptp_adj_dev_clk_fine - Send virtchnl adj time message
* @adapter: Driver specific private structure
* @incval: Source timer increment value per clock cycle
*
* Send virtchnl adj fine message to adjust the frequency of the clock by
* incval.
*
* Return: 0 on success, -errno otherwise.
*/
int idpf_ptp_adj_dev_clk_fine(struct idpf_adapter *adapter, u64 incval)
{
struct virtchnl2_ptp_adj_dev_clk_fine adj_dev_clk_fine_msg = {
.incval = cpu_to_le64(incval),
};
struct idpf_vc_xn_params xn_params = {
.vc_op = VIRTCHNL2_OP_PTP_ADJ_DEV_CLK_FINE,
.send_buf.iov_base = &adj_dev_clk_fine_msg,
.send_buf.iov_len = sizeof(adj_dev_clk_fine_msg),
.recv_buf.iov_base = &adj_dev_clk_fine_msg,
.recv_buf.iov_len = sizeof(adj_dev_clk_fine_msg),
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
};
int reply_sz;
reply_sz = idpf_vc_xn_exec(adapter, &xn_params);
if (reply_sz < 0)
return reply_sz;
if (reply_sz != sizeof(adj_dev_clk_fine_msg))
return -EIO;
return 0;
}
/**
* idpf_ptp_get_vport_tstamps_caps - Send virtchnl to get tstamps caps for vport
* @vport: Virtual port structure
*
* Send virtchnl get vport tstamps caps message to receive the set of tstamp
* capabilities per vport.
*
* Return: 0 on success, -errno otherwise.
*/
int idpf_ptp_get_vport_tstamps_caps(struct idpf_vport *vport)
{
struct virtchnl2_ptp_get_vport_tx_tstamp_caps send_tx_tstamp_caps;
struct virtchnl2_ptp_get_vport_tx_tstamp_caps *rcv_tx_tstamp_caps;
struct virtchnl2_ptp_tx_tstamp_latch_caps tx_tstamp_latch_caps;
struct idpf_ptp_vport_tx_tstamp_caps *tstamp_caps;
struct idpf_ptp_tx_tstamp *ptp_tx_tstamp, *tmp;
struct idpf_vc_xn_params xn_params = {
.vc_op = VIRTCHNL2_OP_PTP_GET_VPORT_TX_TSTAMP_CAPS,
.send_buf.iov_base = &send_tx_tstamp_caps,
.send_buf.iov_len = sizeof(send_tx_tstamp_caps),
.recv_buf.iov_len = IDPF_CTLQ_MAX_BUF_LEN,
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
};
enum idpf_ptp_access tstamp_access, get_dev_clk_access;
struct idpf_ptp *ptp = vport->adapter->ptp;
struct list_head *head;
int err = 0, reply_sz;
u16 num_latches;
u32 size;
if (!ptp)
return -EOPNOTSUPP;
tstamp_access = ptp->tx_tstamp_access;
get_dev_clk_access = ptp->get_dev_clk_time_access;
if (tstamp_access == IDPF_PTP_NONE ||
get_dev_clk_access == IDPF_PTP_NONE)
return -EOPNOTSUPP;
rcv_tx_tstamp_caps = kzalloc(IDPF_CTLQ_MAX_BUF_LEN, GFP_KERNEL);
if (!rcv_tx_tstamp_caps)
return -ENOMEM;
send_tx_tstamp_caps.vport_id = cpu_to_le32(vport->vport_id);
xn_params.recv_buf.iov_base = rcv_tx_tstamp_caps;
reply_sz = idpf_vc_xn_exec(vport->adapter, &xn_params);
if (reply_sz < 0) {
err = reply_sz;
goto get_tstamp_caps_out;
}
num_latches = le16_to_cpu(rcv_tx_tstamp_caps->num_latches);
size = struct_size(rcv_tx_tstamp_caps, tstamp_latches, num_latches);
if (reply_sz != size) {
err = -EIO;
goto get_tstamp_caps_out;
}
size = struct_size(tstamp_caps, tx_tstamp_status, num_latches);
tstamp_caps = kzalloc(size, GFP_KERNEL);
if (!tstamp_caps) {
err = -ENOMEM;
goto get_tstamp_caps_out;
}
tstamp_caps->access = true;
tstamp_caps->num_entries = num_latches;
INIT_LIST_HEAD(&tstamp_caps->latches_in_use);
INIT_LIST_HEAD(&tstamp_caps->latches_free);
spin_lock_init(&tstamp_caps->latches_lock);
spin_lock_init(&tstamp_caps->status_lock);
tstamp_caps->tstamp_ns_lo_bit = rcv_tx_tstamp_caps->tstamp_ns_lo_bit;
for (u16 i = 0; i < tstamp_caps->num_entries; i++) {
__le32 offset_l, offset_h;
ptp_tx_tstamp = kzalloc(sizeof(*ptp_tx_tstamp), GFP_KERNEL);
if (!ptp_tx_tstamp) {
err = -ENOMEM;
goto err_free_ptp_tx_stamp_list;
}
tx_tstamp_latch_caps = rcv_tx_tstamp_caps->tstamp_latches[i];
if (tstamp_access != IDPF_PTP_DIRECT)
goto skip_offsets;
offset_l = tx_tstamp_latch_caps.tx_latch_reg_offset_l;
offset_h = tx_tstamp_latch_caps.tx_latch_reg_offset_h;
ptp_tx_tstamp->tx_latch_reg_offset_l = le32_to_cpu(offset_l);
ptp_tx_tstamp->tx_latch_reg_offset_h = le32_to_cpu(offset_h);
skip_offsets:
ptp_tx_tstamp->idx = tx_tstamp_latch_caps.index;
list_add(&ptp_tx_tstamp->list_member,
&tstamp_caps->latches_free);
tstamp_caps->tx_tstamp_status[i].state = IDPF_PTP_FREE;
}
vport->tx_tstamp_caps = tstamp_caps;
kfree(rcv_tx_tstamp_caps);
return 0;
err_free_ptp_tx_stamp_list:
head = &tstamp_caps->latches_free;
list_for_each_entry_safe(ptp_tx_tstamp, tmp, head, list_member) {
list_del(&ptp_tx_tstamp->list_member);
kfree(ptp_tx_tstamp);
}
kfree(tstamp_caps);
get_tstamp_caps_out:
kfree(rcv_tx_tstamp_caps);
return err;
}
/**
* idpf_ptp_update_tstamp_tracker - Update the Tx timestamp tracker based on
* the skb compatibility.
* @caps: Tx timestamp capabilities that monitor the latch status
* @skb: skb for which the tstamp value is returned through virtchnl message
* @current_state: Current state of the Tx timestamp latch
* @expected_state: Expected state of the Tx timestamp latch
*
* Find a proper skb tracker for which the Tx timestamp is received and change
* the state to expected value.
*
* Return: true if the tracker has been found and updated, false otherwise.
*/
static bool
idpf_ptp_update_tstamp_tracker(struct idpf_ptp_vport_tx_tstamp_caps *caps,
struct sk_buff *skb,
enum idpf_ptp_tx_tstamp_state current_state,
enum idpf_ptp_tx_tstamp_state expected_state)
{
bool updated = false;
spin_lock(&caps->status_lock);
for (u16 i = 0; i < caps->num_entries; i++) {
struct idpf_ptp_tx_tstamp_status *status;
status = &caps->tx_tstamp_status[i];
if (skb == status->skb && status->state == current_state) {
status->state = expected_state;
updated = true;
break;
}
}
spin_unlock(&caps->status_lock);
return updated;
}
/**
* idpf_ptp_get_tstamp_value - Get the Tx timestamp value and provide it
* back to the skb.
* @vport: Virtual port structure
* @tstamp_latch: Tx timestamp latch structure fulfilled by the Control Plane
* @ptp_tx_tstamp: Tx timestamp latch to add to the free list
*
* Read the value of the Tx timestamp for a given latch received from the
* Control Plane, extend it to 64 bit and provide back to the skb.
*
* Return: 0 on success, -errno otherwise.
*/
static int
idpf_ptp_get_tstamp_value(struct idpf_vport *vport,
struct virtchnl2_ptp_tx_tstamp_latch *tstamp_latch,
struct idpf_ptp_tx_tstamp *ptp_tx_tstamp)
{
struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps;
struct skb_shared_hwtstamps shhwtstamps;
bool state_upd = false;
u8 tstamp_ns_lo_bit;
u64 tstamp;
tx_tstamp_caps = vport->tx_tstamp_caps;
tstamp_ns_lo_bit = tx_tstamp_caps->tstamp_ns_lo_bit;
ptp_tx_tstamp->tstamp = le64_to_cpu(tstamp_latch->tstamp);
ptp_tx_tstamp->tstamp >>= tstamp_ns_lo_bit;
state_upd = idpf_ptp_update_tstamp_tracker(tx_tstamp_caps,
ptp_tx_tstamp->skb,
IDPF_PTP_READ_VALUE,
IDPF_PTP_FREE);
if (!state_upd)
return -EINVAL;
tstamp = idpf_ptp_extend_ts(vport, ptp_tx_tstamp->tstamp);
shhwtstamps.hwtstamp = ns_to_ktime(tstamp);
skb_tstamp_tx(ptp_tx_tstamp->skb, &shhwtstamps);
consume_skb(ptp_tx_tstamp->skb);
list_add(&ptp_tx_tstamp->list_member,
&tx_tstamp_caps->latches_free);
return 0;
}
/**
* idpf_ptp_get_tx_tstamp_async_handler - Async callback for getting Tx tstamps
* @adapter: Driver specific private structure
* @xn: transaction for message
* @ctlq_msg: received message
*
* Read the tstamps Tx tstamp values from a received message and put them
* directly to the skb. The number of timestamps to read is specified by
* the virtchnl message.
*
* Return: 0 on success, -errno otherwise.
*/
static int
idpf_ptp_get_tx_tstamp_async_handler(struct idpf_adapter *adapter,
struct idpf_vc_xn *xn,
const struct idpf_ctlq_msg *ctlq_msg)
{
struct virtchnl2_ptp_get_vport_tx_tstamp_latches *recv_tx_tstamp_msg;
struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps;
struct virtchnl2_ptp_tx_tstamp_latch tstamp_latch;
struct idpf_ptp_tx_tstamp *tx_tstamp, *tmp;
struct idpf_vport *tstamp_vport = NULL;
struct list_head *head;
u16 num_latches;
u32 vport_id;
int err = 0;
recv_tx_tstamp_msg = ctlq_msg->ctx.indirect.payload->va;
vport_id = le32_to_cpu(recv_tx_tstamp_msg->vport_id);
idpf_for_each_vport(adapter, vport) {
if (!vport)
continue;
if (vport->vport_id == vport_id) {
tstamp_vport = vport;
break;
}
}
if (!tstamp_vport || !tstamp_vport->tx_tstamp_caps)
return -EINVAL;
tx_tstamp_caps = tstamp_vport->tx_tstamp_caps;
num_latches = le16_to_cpu(recv_tx_tstamp_msg->num_latches);
spin_lock_bh(&tx_tstamp_caps->latches_lock);
head = &tx_tstamp_caps->latches_in_use;
for (u16 i = 0; i < num_latches; i++) {
tstamp_latch = recv_tx_tstamp_msg->tstamp_latches[i];
if (!tstamp_latch.valid)
continue;
if (list_empty(head)) {
err = -ENOBUFS;
goto unlock;
}
list_for_each_entry_safe(tx_tstamp, tmp, head, list_member) {
if (tstamp_latch.index == tx_tstamp->idx) {
list_del(&tx_tstamp->list_member);
err = idpf_ptp_get_tstamp_value(tstamp_vport,
&tstamp_latch,
tx_tstamp);
if (err)
goto unlock;
break;
}
}
}
unlock:
spin_unlock_bh(&tx_tstamp_caps->latches_lock);
return err;
}
/**
* idpf_ptp_get_tx_tstamp - Send virtchnl get Tx timestamp latches message
* @vport: Virtual port structure
*
* Send virtchnl get Tx tstamp message to read the value of the HW timestamp.
* The message contains a list of indexes set in the Tx descriptors.
*
* Return: 0 on success, -errno otherwise.
*/
int idpf_ptp_get_tx_tstamp(struct idpf_vport *vport)
{
struct virtchnl2_ptp_get_vport_tx_tstamp_latches *send_tx_tstamp_msg;
struct idpf_ptp_vport_tx_tstamp_caps *tx_tstamp_caps;
struct idpf_vc_xn_params xn_params = {
.vc_op = VIRTCHNL2_OP_PTP_GET_VPORT_TX_TSTAMP,
.timeout_ms = IDPF_VC_XN_DEFAULT_TIMEOUT_MSEC,
.async = true,
.async_handler = idpf_ptp_get_tx_tstamp_async_handler,
};
struct idpf_ptp_tx_tstamp *ptp_tx_tstamp;
int reply_sz, size, msg_size;
struct list_head *head;
bool state_upd;
u16 id = 0;
tx_tstamp_caps = vport->tx_tstamp_caps;
head = &tx_tstamp_caps->latches_in_use;
size = struct_size(send_tx_tstamp_msg, tstamp_latches,
tx_tstamp_caps->num_entries);
send_tx_tstamp_msg = kzalloc(size, GFP_KERNEL);
if (!send_tx_tstamp_msg)
return -ENOMEM;
spin_lock_bh(&tx_tstamp_caps->latches_lock);
list_for_each_entry(ptp_tx_tstamp, head, list_member) {
u8 idx;
state_upd = idpf_ptp_update_tstamp_tracker(tx_tstamp_caps,
ptp_tx_tstamp->skb,
IDPF_PTP_REQUEST,
IDPF_PTP_READ_VALUE);
if (!state_upd)
continue;
idx = ptp_tx_tstamp->idx;
send_tx_tstamp_msg->tstamp_latches[id].index = idx;
id++;
}
spin_unlock_bh(&tx_tstamp_caps->latches_lock);
msg_size = struct_size(send_tx_tstamp_msg, tstamp_latches, id);
send_tx_tstamp_msg->vport_id = cpu_to_le32(vport->vport_id);
send_tx_tstamp_msg->num_latches = cpu_to_le16(id);
xn_params.send_buf.iov_base = send_tx_tstamp_msg;
xn_params.send_buf.iov_len = msg_size;
reply_sz = idpf_vc_xn_exec(vport->adapter, &xn_params);
kfree(send_tx_tstamp_msg);
return min(reply_sz, 0);
}