| // SPDX-License-Identifier: GPL-2.0-only | 
 |  | 
 | #include <net/netdev_queues.h> | 
 |  | 
 | #include "netlink.h" | 
 | #include "common.h" | 
 |  | 
 | struct rings_req_info { | 
 | 	struct ethnl_req_info		base; | 
 | }; | 
 |  | 
 | struct rings_reply_data { | 
 | 	struct ethnl_reply_data		base; | 
 | 	struct ethtool_ringparam	ringparam; | 
 | 	struct kernel_ethtool_ringparam	kernel_ringparam; | 
 | 	u32				supported_ring_params; | 
 | }; | 
 |  | 
 | #define RINGS_REPDATA(__reply_base) \ | 
 | 	container_of(__reply_base, struct rings_reply_data, base) | 
 |  | 
 | const struct nla_policy ethnl_rings_get_policy[] = { | 
 | 	[ETHTOOL_A_RINGS_HEADER]		= | 
 | 		NLA_POLICY_NESTED(ethnl_header_policy), | 
 | }; | 
 |  | 
 | static int rings_prepare_data(const struct ethnl_req_info *req_base, | 
 | 			      struct ethnl_reply_data *reply_base, | 
 | 			      const struct genl_info *info) | 
 | { | 
 | 	struct rings_reply_data *data = RINGS_REPDATA(reply_base); | 
 | 	struct net_device *dev = reply_base->dev; | 
 | 	int ret; | 
 |  | 
 | 	if (!dev->ethtool_ops->get_ringparam) | 
 | 		return -EOPNOTSUPP; | 
 |  | 
 | 	data->supported_ring_params = dev->ethtool_ops->supported_ring_params; | 
 | 	ret = ethnl_ops_begin(dev); | 
 | 	if (ret < 0) | 
 | 		return ret; | 
 |  | 
 | 	data->kernel_ringparam.tcp_data_split = dev->cfg->hds_config; | 
 | 	data->kernel_ringparam.hds_thresh = dev->cfg->hds_thresh; | 
 |  | 
 | 	dev->ethtool_ops->get_ringparam(dev, &data->ringparam, | 
 | 					&data->kernel_ringparam, info->extack); | 
 | 	ethnl_ops_complete(dev); | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | static int rings_reply_size(const struct ethnl_req_info *req_base, | 
 | 			    const struct ethnl_reply_data *reply_base) | 
 | { | 
 | 	return nla_total_size(sizeof(u32)) +	/* _RINGS_RX_MAX */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_RX_MINI_MAX */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_RX_JUMBO_MAX */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_TX_MAX */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_RX */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_RX_MINI */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_RX_JUMBO */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_TX */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_RX_BUF_LEN */ | 
 | 	       nla_total_size(sizeof(u8))  +	/* _RINGS_TCP_DATA_SPLIT */ | 
 | 	       nla_total_size(sizeof(u32)  +	/* _RINGS_CQE_SIZE */ | 
 | 	       nla_total_size(sizeof(u8))  +	/* _RINGS_TX_PUSH */ | 
 | 	       nla_total_size(sizeof(u8))) +	/* _RINGS_RX_PUSH */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_TX_PUSH_BUF_LEN */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_TX_PUSH_BUF_LEN_MAX */ | 
 | 	       nla_total_size(sizeof(u32)) +	/* _RINGS_HDS_THRESH */ | 
 | 	       nla_total_size(sizeof(u32));	/* _RINGS_HDS_THRESH_MAX*/ | 
 | } | 
 |  | 
 | static int rings_fill_reply(struct sk_buff *skb, | 
 | 			    const struct ethnl_req_info *req_base, | 
 | 			    const struct ethnl_reply_data *reply_base) | 
 | { | 
 | 	const struct rings_reply_data *data = RINGS_REPDATA(reply_base); | 
 | 	const struct kernel_ethtool_ringparam *kr = &data->kernel_ringparam; | 
 | 	const struct ethtool_ringparam *ringparam = &data->ringparam; | 
 | 	u32 supported_ring_params = data->supported_ring_params; | 
 |  | 
 | 	WARN_ON(kr->tcp_data_split > ETHTOOL_TCP_DATA_SPLIT_ENABLED); | 
 |  | 
 | 	if ((ringparam->rx_max_pending && | 
 | 	     (nla_put_u32(skb, ETHTOOL_A_RINGS_RX_MAX, | 
 | 			  ringparam->rx_max_pending) || | 
 | 	      nla_put_u32(skb, ETHTOOL_A_RINGS_RX, | 
 | 			  ringparam->rx_pending))) || | 
 | 	    (ringparam->rx_mini_max_pending && | 
 | 	     (nla_put_u32(skb, ETHTOOL_A_RINGS_RX_MINI_MAX, | 
 | 			  ringparam->rx_mini_max_pending) || | 
 | 	      nla_put_u32(skb, ETHTOOL_A_RINGS_RX_MINI, | 
 | 			  ringparam->rx_mini_pending))) || | 
 | 	    (ringparam->rx_jumbo_max_pending && | 
 | 	     (nla_put_u32(skb, ETHTOOL_A_RINGS_RX_JUMBO_MAX, | 
 | 			  ringparam->rx_jumbo_max_pending) || | 
 | 	      nla_put_u32(skb, ETHTOOL_A_RINGS_RX_JUMBO, | 
 | 			  ringparam->rx_jumbo_pending))) || | 
 | 	    (ringparam->tx_max_pending && | 
 | 	     (nla_put_u32(skb, ETHTOOL_A_RINGS_TX_MAX, | 
 | 			  ringparam->tx_max_pending) || | 
 | 	      nla_put_u32(skb, ETHTOOL_A_RINGS_TX, | 
 | 			  ringparam->tx_pending)))  || | 
 | 	    (kr->rx_buf_len && | 
 | 	     (nla_put_u32(skb, ETHTOOL_A_RINGS_RX_BUF_LEN, kr->rx_buf_len))) || | 
 | 	    (kr->tcp_data_split && | 
 | 	     (nla_put_u8(skb, ETHTOOL_A_RINGS_TCP_DATA_SPLIT, | 
 | 			 kr->tcp_data_split))) || | 
 | 	    (kr->cqe_size && | 
 | 	     (nla_put_u32(skb, ETHTOOL_A_RINGS_CQE_SIZE, kr->cqe_size))) || | 
 | 	    nla_put_u8(skb, ETHTOOL_A_RINGS_TX_PUSH, !!kr->tx_push) || | 
 | 	    nla_put_u8(skb, ETHTOOL_A_RINGS_RX_PUSH, !!kr->rx_push) || | 
 | 	    ((supported_ring_params & ETHTOOL_RING_USE_TX_PUSH_BUF_LEN) && | 
 | 	     (nla_put_u32(skb, ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN_MAX, | 
 | 			  kr->tx_push_buf_max_len) || | 
 | 	      nla_put_u32(skb, ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN, | 
 | 			  kr->tx_push_buf_len))) || | 
 | 	    ((supported_ring_params & ETHTOOL_RING_USE_HDS_THRS) && | 
 | 	     (nla_put_u32(skb, ETHTOOL_A_RINGS_HDS_THRESH, | 
 | 			  kr->hds_thresh) || | 
 | 	      nla_put_u32(skb, ETHTOOL_A_RINGS_HDS_THRESH_MAX, | 
 | 			  kr->hds_thresh_max)))) | 
 | 		return -EMSGSIZE; | 
 |  | 
 | 	return 0; | 
 | } | 
 |  | 
 | /* RINGS_SET */ | 
 |  | 
 | const struct nla_policy ethnl_rings_set_policy[] = { | 
 | 	[ETHTOOL_A_RINGS_HEADER]		= | 
 | 		NLA_POLICY_NESTED(ethnl_header_policy), | 
 | 	[ETHTOOL_A_RINGS_RX]			= { .type = NLA_U32 }, | 
 | 	[ETHTOOL_A_RINGS_RX_MINI]		= { .type = NLA_U32 }, | 
 | 	[ETHTOOL_A_RINGS_RX_JUMBO]		= { .type = NLA_U32 }, | 
 | 	[ETHTOOL_A_RINGS_TX]			= { .type = NLA_U32 }, | 
 | 	[ETHTOOL_A_RINGS_RX_BUF_LEN]            = NLA_POLICY_MIN(NLA_U32, 1), | 
 | 	[ETHTOOL_A_RINGS_TCP_DATA_SPLIT]	= | 
 | 		NLA_POLICY_MAX(NLA_U8, ETHTOOL_TCP_DATA_SPLIT_ENABLED), | 
 | 	[ETHTOOL_A_RINGS_CQE_SIZE]		= NLA_POLICY_MIN(NLA_U32, 1), | 
 | 	[ETHTOOL_A_RINGS_TX_PUSH]		= NLA_POLICY_MAX(NLA_U8, 1), | 
 | 	[ETHTOOL_A_RINGS_RX_PUSH]		= NLA_POLICY_MAX(NLA_U8, 1), | 
 | 	[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN]	= { .type = NLA_U32 }, | 
 | 	[ETHTOOL_A_RINGS_HDS_THRESH]		= { .type = NLA_U32 }, | 
 | }; | 
 |  | 
 | static int | 
 | ethnl_set_rings_validate(struct ethnl_req_info *req_info, | 
 | 			 struct genl_info *info) | 
 | { | 
 | 	const struct ethtool_ops *ops = req_info->dev->ethtool_ops; | 
 | 	struct nlattr **tb = info->attrs; | 
 |  | 
 | 	if (tb[ETHTOOL_A_RINGS_RX_BUF_LEN] && | 
 | 	    !(ops->supported_ring_params & ETHTOOL_RING_USE_RX_BUF_LEN)) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, | 
 | 				    tb[ETHTOOL_A_RINGS_RX_BUF_LEN], | 
 | 				    "setting rx buf len not supported"); | 
 | 		return -EOPNOTSUPP; | 
 | 	} | 
 |  | 
 | 	if (tb[ETHTOOL_A_RINGS_TCP_DATA_SPLIT] && | 
 | 	    !(ops->supported_ring_params & ETHTOOL_RING_USE_TCP_DATA_SPLIT)) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, | 
 | 				    tb[ETHTOOL_A_RINGS_TCP_DATA_SPLIT], | 
 | 				    "setting TCP data split is not supported"); | 
 | 		return -EOPNOTSUPP; | 
 | 	} | 
 |  | 
 | 	if (tb[ETHTOOL_A_RINGS_HDS_THRESH] && | 
 | 	    !(ops->supported_ring_params & ETHTOOL_RING_USE_HDS_THRS)) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, | 
 | 				    tb[ETHTOOL_A_RINGS_HDS_THRESH], | 
 | 				    "setting hds-thresh is not supported"); | 
 | 		return -EOPNOTSUPP; | 
 | 	} | 
 |  | 
 | 	if (tb[ETHTOOL_A_RINGS_CQE_SIZE] && | 
 | 	    !(ops->supported_ring_params & ETHTOOL_RING_USE_CQE_SIZE)) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, | 
 | 				    tb[ETHTOOL_A_RINGS_CQE_SIZE], | 
 | 				    "setting cqe size not supported"); | 
 | 		return -EOPNOTSUPP; | 
 | 	} | 
 |  | 
 | 	if (tb[ETHTOOL_A_RINGS_TX_PUSH] && | 
 | 	    !(ops->supported_ring_params & ETHTOOL_RING_USE_TX_PUSH)) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, | 
 | 				    tb[ETHTOOL_A_RINGS_TX_PUSH], | 
 | 				    "setting tx push not supported"); | 
 | 		return -EOPNOTSUPP; | 
 | 	} | 
 |  | 
 | 	if (tb[ETHTOOL_A_RINGS_RX_PUSH] && | 
 | 	    !(ops->supported_ring_params & ETHTOOL_RING_USE_RX_PUSH)) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, | 
 | 				    tb[ETHTOOL_A_RINGS_RX_PUSH], | 
 | 				    "setting rx push not supported"); | 
 | 		return -EOPNOTSUPP; | 
 | 	} | 
 |  | 
 | 	if (tb[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN] && | 
 | 	    !(ops->supported_ring_params & ETHTOOL_RING_USE_TX_PUSH_BUF_LEN)) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, | 
 | 				    tb[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN], | 
 | 				    "setting tx push buf len is not supported"); | 
 | 		return -EOPNOTSUPP; | 
 | 	} | 
 |  | 
 | 	return ops->get_ringparam && ops->set_ringparam ? 1 : -EOPNOTSUPP; | 
 | } | 
 |  | 
 | static int | 
 | ethnl_set_rings(struct ethnl_req_info *req_info, struct genl_info *info) | 
 | { | 
 | 	struct kernel_ethtool_ringparam kernel_ringparam; | 
 | 	struct net_device *dev = req_info->dev; | 
 | 	struct ethtool_ringparam ringparam; | 
 | 	struct nlattr **tb = info->attrs; | 
 | 	const struct nlattr *err_attr; | 
 | 	bool mod = false; | 
 | 	int ret; | 
 |  | 
 | 	ethtool_ringparam_get_cfg(dev, &ringparam, &kernel_ringparam, | 
 | 				  info->extack); | 
 |  | 
 | 	ethnl_update_u32(&ringparam.rx_pending, tb[ETHTOOL_A_RINGS_RX], &mod); | 
 | 	ethnl_update_u32(&ringparam.rx_mini_pending, | 
 | 			 tb[ETHTOOL_A_RINGS_RX_MINI], &mod); | 
 | 	ethnl_update_u32(&ringparam.rx_jumbo_pending, | 
 | 			 tb[ETHTOOL_A_RINGS_RX_JUMBO], &mod); | 
 | 	ethnl_update_u32(&ringparam.tx_pending, tb[ETHTOOL_A_RINGS_TX], &mod); | 
 | 	ethnl_update_u32(&kernel_ringparam.rx_buf_len, | 
 | 			 tb[ETHTOOL_A_RINGS_RX_BUF_LEN], &mod); | 
 | 	ethnl_update_u8(&kernel_ringparam.tcp_data_split, | 
 | 			tb[ETHTOOL_A_RINGS_TCP_DATA_SPLIT], &mod); | 
 | 	ethnl_update_u32(&kernel_ringparam.cqe_size, | 
 | 			 tb[ETHTOOL_A_RINGS_CQE_SIZE], &mod); | 
 | 	ethnl_update_u8(&kernel_ringparam.tx_push, | 
 | 			tb[ETHTOOL_A_RINGS_TX_PUSH], &mod); | 
 | 	ethnl_update_u8(&kernel_ringparam.rx_push, | 
 | 			tb[ETHTOOL_A_RINGS_RX_PUSH], &mod); | 
 | 	ethnl_update_u32(&kernel_ringparam.tx_push_buf_len, | 
 | 			 tb[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN], &mod); | 
 | 	ethnl_update_u32(&kernel_ringparam.hds_thresh, | 
 | 			 tb[ETHTOOL_A_RINGS_HDS_THRESH], &mod); | 
 | 	if (!mod) | 
 | 		return 0; | 
 |  | 
 | 	if (kernel_ringparam.tcp_data_split == ETHTOOL_TCP_DATA_SPLIT_ENABLED && | 
 | 	    dev_xdp_sb_prog_count(dev)) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, | 
 | 				    tb[ETHTOOL_A_RINGS_TCP_DATA_SPLIT], | 
 | 				    "tcp-data-split can not be enabled with single buffer XDP"); | 
 | 		return -EINVAL; | 
 | 	} | 
 |  | 
 | 	if (dev_get_min_mp_channel_count(dev)) { | 
 | 		if (kernel_ringparam.tcp_data_split != | 
 | 		    ETHTOOL_TCP_DATA_SPLIT_ENABLED) { | 
 | 			NL_SET_ERR_MSG(info->extack, | 
 | 				       "can't disable tcp-data-split while device has memory provider enabled"); | 
 | 			return -EINVAL; | 
 | 		} else if (kernel_ringparam.hds_thresh) { | 
 | 			NL_SET_ERR_MSG(info->extack, | 
 | 				       "can't set non-zero hds_thresh while device is memory provider enabled"); | 
 | 			return -EINVAL; | 
 | 		} | 
 | 	} | 
 |  | 
 | 	/* ensure new ring parameters are within limits */ | 
 | 	if (ringparam.rx_pending > ringparam.rx_max_pending) | 
 | 		err_attr = tb[ETHTOOL_A_RINGS_RX]; | 
 | 	else if (ringparam.rx_mini_pending > ringparam.rx_mini_max_pending) | 
 | 		err_attr = tb[ETHTOOL_A_RINGS_RX_MINI]; | 
 | 	else if (ringparam.rx_jumbo_pending > ringparam.rx_jumbo_max_pending) | 
 | 		err_attr = tb[ETHTOOL_A_RINGS_RX_JUMBO]; | 
 | 	else if (ringparam.tx_pending > ringparam.tx_max_pending) | 
 | 		err_attr = tb[ETHTOOL_A_RINGS_TX]; | 
 | 	else if (kernel_ringparam.hds_thresh > kernel_ringparam.hds_thresh_max) | 
 | 		err_attr = tb[ETHTOOL_A_RINGS_HDS_THRESH]; | 
 | 	else | 
 | 		err_attr = NULL; | 
 | 	if (err_attr) { | 
 | 		NL_SET_ERR_MSG_ATTR(info->extack, err_attr, | 
 | 				    "requested ring size exceeds maximum"); | 
 | 		return -EINVAL; | 
 | 	} | 
 |  | 
 | 	if (kernel_ringparam.tx_push_buf_len > kernel_ringparam.tx_push_buf_max_len) { | 
 | 		NL_SET_ERR_MSG_ATTR_FMT(info->extack, tb[ETHTOOL_A_RINGS_TX_PUSH_BUF_LEN], | 
 | 					"Requested TX push buffer exceeds the maximum of %u", | 
 | 					kernel_ringparam.tx_push_buf_max_len); | 
 |  | 
 | 		return -EINVAL; | 
 | 	} | 
 |  | 
 | 	dev->cfg_pending->hds_config = kernel_ringparam.tcp_data_split; | 
 | 	dev->cfg_pending->hds_thresh = kernel_ringparam.hds_thresh; | 
 |  | 
 | 	ret = dev->ethtool_ops->set_ringparam(dev, &ringparam, | 
 | 					      &kernel_ringparam, info->extack); | 
 | 	return ret < 0 ? ret : 1; | 
 | } | 
 |  | 
 | const struct ethnl_request_ops ethnl_rings_request_ops = { | 
 | 	.request_cmd		= ETHTOOL_MSG_RINGS_GET, | 
 | 	.reply_cmd		= ETHTOOL_MSG_RINGS_GET_REPLY, | 
 | 	.hdr_attr		= ETHTOOL_A_RINGS_HEADER, | 
 | 	.req_info_size		= sizeof(struct rings_req_info), | 
 | 	.reply_data_size	= sizeof(struct rings_reply_data), | 
 |  | 
 | 	.prepare_data		= rings_prepare_data, | 
 | 	.reply_size		= rings_reply_size, | 
 | 	.fill_reply		= rings_fill_reply, | 
 |  | 
 | 	.set_validate		= ethnl_set_rings_validate, | 
 | 	.set			= ethnl_set_rings, | 
 | 	.set_ntf_cmd		= ETHTOOL_MSG_RINGS_NTF, | 
 | }; |