1557 lines
		
	
	
		
			42 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			1557 lines
		
	
	
		
			42 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| // SPDX-License-Identifier: GPL-2.0-only
 | |
| /*
 | |
|  * Copyright (C) 2020-21 Intel Corporation.
 | |
|  */
 | |
| 
 | |
| #include <linux/nospec.h>
 | |
| 
 | |
| #include "iosm_ipc_imem_ops.h"
 | |
| #include "iosm_ipc_mux_codec.h"
 | |
| #include "iosm_ipc_task_queue.h"
 | |
| 
 | |
| /* Test the link power state and send a MUX command in blocking mode. */
 | |
| static int ipc_mux_tq_cmd_send(struct iosm_imem *ipc_imem, int arg, void *msg,
 | |
| 			       size_t size)
 | |
| {
 | |
| 	struct iosm_mux *ipc_mux = ipc_imem->mux;
 | |
| 	const struct mux_acb *acb = msg;
 | |
| 
 | |
| 	skb_queue_tail(&ipc_mux->channel->ul_list, acb->skb);
 | |
| 	ipc_imem_ul_send(ipc_mux->imem);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int ipc_mux_acb_send(struct iosm_mux *ipc_mux, bool blocking)
 | |
| {
 | |
| 	struct completion *completion = &ipc_mux->channel->ul_sem;
 | |
| 	int ret = ipc_task_queue_send_task(ipc_mux->imem, ipc_mux_tq_cmd_send,
 | |
| 					   0, &ipc_mux->acb,
 | |
| 					   sizeof(ipc_mux->acb), false);
 | |
| 	if (ret) {
 | |
| 		dev_err(ipc_mux->dev, "unable to send mux command");
 | |
| 		return ret;
 | |
| 	}
 | |
| 
 | |
| 	/* if blocking, suspend the app and wait for irq in the flash or
 | |
| 	 * crash phase. return false on timeout to indicate failure.
 | |
| 	 */
 | |
| 	if (blocking) {
 | |
| 		u32 wait_time_milliseconds = IPC_MUX_CMD_RUN_DEFAULT_TIMEOUT;
 | |
| 
 | |
| 		reinit_completion(completion);
 | |
| 
 | |
| 		if (wait_for_completion_interruptible_timeout
 | |
| 		   (completion, msecs_to_jiffies(wait_time_milliseconds)) ==
 | |
| 		   0) {
 | |
| 			dev_err(ipc_mux->dev, "ch[%d] timeout",
 | |
| 				ipc_mux->channel_id);
 | |
| 			ipc_uevent_send(ipc_mux->imem->dev, UEVENT_MDM_TIMEOUT);
 | |
| 			return -ETIMEDOUT;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* Initialize the command header. */
 | |
| static void ipc_mux_acb_init(struct iosm_mux *ipc_mux)
 | |
| {
 | |
| 	struct mux_acb *acb = &ipc_mux->acb;
 | |
| 	struct mux_acbh *header;
 | |
| 
 | |
| 	header = (struct mux_acbh *)(acb->skb)->data;
 | |
| 	header->block_length = cpu_to_le32(sizeof(struct mux_acbh));
 | |
| 	header->first_cmd_index = header->block_length;
 | |
| 	header->signature = cpu_to_le32(IOSM_AGGR_MUX_SIG_ACBH);
 | |
| 	header->sequence_nr = cpu_to_le16(ipc_mux->acb_tx_sequence_nr++);
 | |
| }
 | |
| 
 | |
| /* Add a command to the ACB. */
 | |
| static struct mux_cmdh *ipc_mux_acb_add_cmd(struct iosm_mux *ipc_mux, u32 cmd,
 | |
| 					    void *param, u32 param_size)
 | |
| {
 | |
| 	struct mux_acbh *header;
 | |
| 	struct mux_cmdh *cmdh;
 | |
| 	struct mux_acb *acb;
 | |
| 
 | |
| 	acb = &ipc_mux->acb;
 | |
| 	header = (struct mux_acbh *)(acb->skb)->data;
 | |
| 	cmdh = (struct mux_cmdh *)
 | |
| 		((acb->skb)->data + le32_to_cpu(header->block_length));
 | |
| 
 | |
| 	cmdh->signature = cpu_to_le32(MUX_SIG_CMDH);
 | |
| 	cmdh->command_type = cpu_to_le32(cmd);
 | |
| 	cmdh->if_id = acb->if_id;
 | |
| 
 | |
| 	acb->cmd = cmd;
 | |
| 	cmdh->cmd_len = cpu_to_le16(offsetof(struct mux_cmdh, param) +
 | |
| 				    param_size);
 | |
| 	cmdh->transaction_id = cpu_to_le32(ipc_mux->tx_transaction_id++);
 | |
| 	if (param)
 | |
| 		memcpy(&cmdh->param, param, param_size);
 | |
| 
 | |
| 	skb_put(acb->skb, le32_to_cpu(header->block_length) +
 | |
| 					le16_to_cpu(cmdh->cmd_len));
 | |
| 
 | |
| 	return cmdh;
 | |
| }
 | |
| 
 | |
| /* Prepare mux Command */
 | |
| static struct mux_lite_cmdh *ipc_mux_lite_add_cmd(struct iosm_mux *ipc_mux,
 | |
| 						  u32 cmd, struct mux_acb *acb,
 | |
| 						  void *param, u32 param_size)
 | |
| {
 | |
| 	struct mux_lite_cmdh *cmdh = (struct mux_lite_cmdh *)acb->skb->data;
 | |
| 
 | |
| 	cmdh->signature = cpu_to_le32(MUX_SIG_CMDH);
 | |
| 	cmdh->command_type = cpu_to_le32(cmd);
 | |
| 	cmdh->if_id = acb->if_id;
 | |
| 
 | |
| 	acb->cmd = cmd;
 | |
| 
 | |
| 	cmdh->cmd_len = cpu_to_le16(offsetof(struct mux_lite_cmdh, param) +
 | |
| 				    param_size);
 | |
| 	cmdh->transaction_id = cpu_to_le32(ipc_mux->tx_transaction_id++);
 | |
| 
 | |
| 	if (param)
 | |
| 		memcpy(&cmdh->param, param, param_size);
 | |
| 
 | |
| 	skb_put(acb->skb, le16_to_cpu(cmdh->cmd_len));
 | |
| 
 | |
| 	return cmdh;
 | |
| }
 | |
| 
 | |
| static int ipc_mux_acb_alloc(struct iosm_mux *ipc_mux)
 | |
| {
 | |
| 	struct mux_acb *acb = &ipc_mux->acb;
 | |
| 	struct sk_buff *skb;
 | |
| 	dma_addr_t mapping;
 | |
| 
 | |
| 	/* Allocate skb memory for the uplink buffer. */
 | |
| 	skb = ipc_pcie_alloc_skb(ipc_mux->pcie, MUX_MAX_UL_ACB_BUF_SIZE,
 | |
| 				 GFP_ATOMIC, &mapping, DMA_TO_DEVICE, 0);
 | |
| 	if (!skb)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	/* Save the skb address. */
 | |
| 	acb->skb = skb;
 | |
| 
 | |
| 	memset(skb->data, 0, MUX_MAX_UL_ACB_BUF_SIZE);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int ipc_mux_dl_acb_send_cmds(struct iosm_mux *ipc_mux, u32 cmd_type, u8 if_id,
 | |
| 			     u32 transaction_id, union mux_cmd_param *param,
 | |
| 			     size_t res_size, bool blocking, bool respond)
 | |
| {
 | |
| 	struct mux_acb *acb = &ipc_mux->acb;
 | |
| 	union mux_type_cmdh cmdh;
 | |
| 	int ret = 0;
 | |
| 
 | |
| 	acb->if_id = if_id;
 | |
| 	ret = ipc_mux_acb_alloc(ipc_mux);
 | |
| 	if (ret)
 | |
| 		return ret;
 | |
| 
 | |
| 	if (ipc_mux->protocol == MUX_LITE) {
 | |
| 		cmdh.ack_lite = ipc_mux_lite_add_cmd(ipc_mux, cmd_type, acb,
 | |
| 						     param, res_size);
 | |
| 
 | |
| 		if (respond)
 | |
| 			cmdh.ack_lite->transaction_id =
 | |
| 					cpu_to_le32(transaction_id);
 | |
| 	} else {
 | |
| 		/* Initialize the ACB header. */
 | |
| 		ipc_mux_acb_init(ipc_mux);
 | |
| 		cmdh.ack_aggr = ipc_mux_acb_add_cmd(ipc_mux, cmd_type, param,
 | |
| 						    res_size);
 | |
| 
 | |
| 		if (respond)
 | |
| 			cmdh.ack_aggr->transaction_id =
 | |
| 					cpu_to_le32(transaction_id);
 | |
| 	}
 | |
| 	ret = ipc_mux_acb_send(ipc_mux, blocking);
 | |
| 
 | |
| 	return ret;
 | |
| }
 | |
| 
 | |
| void ipc_mux_netif_tx_flowctrl(struct mux_session *session, int idx, bool on)
 | |
| {
 | |
| 	/* Inform the network interface to start/stop flow ctrl */
 | |
| 	ipc_wwan_tx_flowctrl(session->wwan, idx, on);
 | |
| }
 | |
| 
 | |
| static int ipc_mux_dl_cmdresps_decode_process(struct iosm_mux *ipc_mux,
 | |
| 					      union mux_cmd_param param,
 | |
| 					      __le32 command_type, u8 if_id,
 | |
| 					      __le32 transaction_id)
 | |
| {
 | |
| 	struct mux_acb *acb = &ipc_mux->acb;
 | |
| 
 | |
| 	switch (le32_to_cpu(command_type)) {
 | |
| 	case MUX_CMD_OPEN_SESSION_RESP:
 | |
| 	case MUX_CMD_CLOSE_SESSION_RESP:
 | |
| 		/* Resume the control application. */
 | |
| 		acb->got_param = param;
 | |
| 		break;
 | |
| 
 | |
| 	case MUX_LITE_CMD_FLOW_CTL_ACK:
 | |
| 		/* This command type is not expected as response for
 | |
| 		 * Aggregation version of the protocol. So return non-zero.
 | |
| 		 */
 | |
| 		if (ipc_mux->protocol != MUX_LITE)
 | |
| 			return -EINVAL;
 | |
| 
 | |
| 		dev_dbg(ipc_mux->dev, "if_id %u FLOW_CTL_ACK %u received",
 | |
| 			if_id, le32_to_cpu(transaction_id));
 | |
| 		break;
 | |
| 
 | |
| 	case IOSM_AGGR_MUX_CMD_FLOW_CTL_ACK:
 | |
| 		/* This command type is not expected as response for
 | |
| 		 * Lite version of the protocol. So return non-zero.
 | |
| 		 */
 | |
| 		if (ipc_mux->protocol == MUX_LITE)
 | |
| 			return -EINVAL;
 | |
| 		break;
 | |
| 
 | |
| 	default:
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 
 | |
| 	acb->wanted_response = MUX_CMD_INVALID;
 | |
| 	acb->got_response = le32_to_cpu(command_type);
 | |
| 	complete(&ipc_mux->channel->ul_sem);
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static int ipc_mux_dl_cmds_decode_process(struct iosm_mux *ipc_mux,
 | |
| 					  union mux_cmd_param *param,
 | |
| 					  __le32 command_type, u8 if_id,
 | |
| 					  __le16 cmd_len, int size)
 | |
| {
 | |
| 	struct mux_session *session;
 | |
| 	struct hrtimer *adb_timer;
 | |
| 
 | |
| 	dev_dbg(ipc_mux->dev, "if_id[%d]: dlcmds decode process %d",
 | |
| 		if_id, le32_to_cpu(command_type));
 | |
| 
 | |
| 	switch (le32_to_cpu(command_type)) {
 | |
| 	case MUX_LITE_CMD_FLOW_CTL:
 | |
| 	case IOSM_AGGR_MUX_CMD_FLOW_CTL_DISABLE:
 | |
| 
 | |
| 		if (if_id >= IPC_MEM_MUX_IP_SESSION_ENTRIES) {
 | |
| 			dev_err(ipc_mux->dev, "if_id [%d] not valid",
 | |
| 				if_id);
 | |
| 			return -EINVAL; /* No session interface id. */
 | |
| 		}
 | |
| 
 | |
| 		session = &ipc_mux->session[if_id];
 | |
| 		adb_timer = &ipc_mux->imem->adb_timer;
 | |
| 
 | |
| 		if (param->flow_ctl.mask == cpu_to_le32(0xFFFFFFFF)) {
 | |
| 			/* Backward Compatibility */
 | |
| 			if (cmd_len == cpu_to_le16(size))
 | |
| 				session->flow_ctl_mask =
 | |
| 					le32_to_cpu(param->flow_ctl.mask);
 | |
| 			else
 | |
| 				session->flow_ctl_mask = ~0;
 | |
| 			/* if CP asks for FLOW CTRL Enable
 | |
| 			 * then set our internal flow control Tx flag
 | |
| 			 * to limit uplink session queueing
 | |
| 			 */
 | |
| 			session->net_tx_stop = true;
 | |
| 
 | |
| 			/* We have to call Finish ADB here.
 | |
| 			 * Otherwise any already queued data
 | |
| 			 * will be sent to CP when ADB is full
 | |
| 			 * for some other sessions.
 | |
| 			 */
 | |
| 			if (ipc_mux->protocol == MUX_AGGREGATION) {
 | |
| 				ipc_mux_ul_adb_finish(ipc_mux);
 | |
| 				ipc_imem_hrtimer_stop(adb_timer);
 | |
| 			}
 | |
| 			/* Update the stats */
 | |
| 			session->flow_ctl_en_cnt++;
 | |
| 		} else if (param->flow_ctl.mask == 0) {
 | |
| 			/* Just reset the Flow control mask and let
 | |
| 			 * mux_flow_ctrl_low_thre_b take control on
 | |
| 			 * our internal Tx flag and enabling kernel
 | |
| 			 * flow control
 | |
| 			 */
 | |
| 			dev_dbg(ipc_mux->dev, "if_id[%u] flow_ctl mask 0x%08X",
 | |
| 				if_id, le32_to_cpu(param->flow_ctl.mask));
 | |
| 			/* Backward Compatibility */
 | |
| 			if (cmd_len == cpu_to_le16(size))
 | |
| 				session->flow_ctl_mask =
 | |
| 					le32_to_cpu(param->flow_ctl.mask);
 | |
| 			else
 | |
| 				session->flow_ctl_mask = 0;
 | |
| 			/* Update the stats */
 | |
| 			session->flow_ctl_dis_cnt++;
 | |
| 		} else {
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 		ipc_mux->acc_adb_size = 0;
 | |
| 		ipc_mux->acc_payload_size = 0;
 | |
| 
 | |
| 		dev_dbg(ipc_mux->dev, "if_id[%u] FLOW CTRL 0x%08X", if_id,
 | |
| 			le32_to_cpu(param->flow_ctl.mask));
 | |
| 		break;
 | |
| 
 | |
| 	case MUX_LITE_CMD_LINK_STATUS_REPORT:
 | |
| 		break;
 | |
| 
 | |
| 	default:
 | |
| 		return -EINVAL;
 | |
| 	}
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* Decode and Send appropriate response to a command block. */
 | |
| static void ipc_mux_dl_cmd_decode(struct iosm_mux *ipc_mux, struct sk_buff *skb)
 | |
| {
 | |
| 	struct mux_lite_cmdh *cmdh = (struct mux_lite_cmdh *)skb->data;
 | |
| 	__le32 trans_id = cmdh->transaction_id;
 | |
| 	int size;
 | |
| 
 | |
| 	if (ipc_mux_dl_cmdresps_decode_process(ipc_mux, cmdh->param,
 | |
| 					       cmdh->command_type, cmdh->if_id,
 | |
| 					       cmdh->transaction_id)) {
 | |
| 		/* Unable to decode command response indicates the cmd_type
 | |
| 		 * may be a command instead of response. So try to decoding it.
 | |
| 		 */
 | |
| 		size = offsetof(struct mux_lite_cmdh, param) +
 | |
| 				sizeof(cmdh->param.flow_ctl);
 | |
| 		if (!ipc_mux_dl_cmds_decode_process(ipc_mux, &cmdh->param,
 | |
| 						    cmdh->command_type,
 | |
| 						    cmdh->if_id,
 | |
| 						    cmdh->cmd_len, size)) {
 | |
| 			/* Decoded command may need a response. Give the
 | |
| 			 * response according to the command type.
 | |
| 			 */
 | |
| 			union mux_cmd_param *mux_cmd = NULL;
 | |
| 			size_t size = 0;
 | |
| 			u32 cmd = MUX_LITE_CMD_LINK_STATUS_REPORT_RESP;
 | |
| 
 | |
| 			if (cmdh->command_type ==
 | |
| 			    cpu_to_le32(MUX_LITE_CMD_LINK_STATUS_REPORT)) {
 | |
| 				mux_cmd = &cmdh->param;
 | |
| 				mux_cmd->link_status_resp.response =
 | |
| 					cpu_to_le32(MUX_CMD_RESP_SUCCESS);
 | |
| 				/* response field is u32 */
 | |
| 				size = sizeof(u32);
 | |
| 			} else if (cmdh->command_type ==
 | |
| 				   cpu_to_le32(MUX_LITE_CMD_FLOW_CTL)) {
 | |
| 				cmd = MUX_LITE_CMD_FLOW_CTL_ACK;
 | |
| 			} else {
 | |
| 				return;
 | |
| 			}
 | |
| 
 | |
| 			if (ipc_mux_dl_acb_send_cmds(ipc_mux, cmd, cmdh->if_id,
 | |
| 						     le32_to_cpu(trans_id),
 | |
| 						     mux_cmd, size, false,
 | |
| 						     true))
 | |
| 				dev_err(ipc_mux->dev,
 | |
| 					"if_id %d: cmd send failed",
 | |
| 					cmdh->if_id);
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* Pass the DL packet to the netif layer. */
 | |
| static int ipc_mux_net_receive(struct iosm_mux *ipc_mux, int if_id,
 | |
| 			       struct iosm_wwan *wwan, u32 offset,
 | |
| 			       u8 service_class, struct sk_buff *skb,
 | |
| 			       u32 pkt_len)
 | |
| {
 | |
| 	struct sk_buff *dest_skb = skb_clone(skb, GFP_ATOMIC);
 | |
| 
 | |
| 	if (!dest_skb)
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	skb_pull(dest_skb, offset);
 | |
| 	skb_trim(dest_skb, pkt_len);
 | |
| 	/* Pass the packet to the netif layer. */
 | |
| 	dest_skb->priority = service_class;
 | |
| 
 | |
| 	return ipc_wwan_receive(wwan, dest_skb, false, if_id);
 | |
| }
 | |
| 
 | |
| /* Decode Flow Credit Table in the block */
 | |
| static void ipc_mux_dl_fcth_decode(struct iosm_mux *ipc_mux,
 | |
| 				   unsigned char *block)
 | |
| {
 | |
| 	struct ipc_mem_lite_gen_tbl *fct = (struct ipc_mem_lite_gen_tbl *)block;
 | |
| 	struct iosm_wwan *wwan;
 | |
| 	int ul_credits;
 | |
| 	int if_id;
 | |
| 
 | |
| 	if (fct->vfl_length != sizeof(fct->vfl.nr_of_bytes)) {
 | |
| 		dev_err(ipc_mux->dev, "unexpected FCT length: %d",
 | |
| 			fct->vfl_length);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if_id = fct->if_id;
 | |
| 	if (if_id >= IPC_MEM_MUX_IP_SESSION_ENTRIES) {
 | |
| 		dev_err(ipc_mux->dev, "not supported if_id: %d", if_id);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/* Is the session active ? */
 | |
| 	if_id = array_index_nospec(if_id, IPC_MEM_MUX_IP_SESSION_ENTRIES);
 | |
| 	wwan = ipc_mux->session[if_id].wwan;
 | |
| 	if (!wwan) {
 | |
| 		dev_err(ipc_mux->dev, "session Net ID is NULL");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	ul_credits = le32_to_cpu(fct->vfl.nr_of_bytes);
 | |
| 
 | |
| 	dev_dbg(ipc_mux->dev, "Flow_Credit:: if_id[%d] Old: %d Grants: %d",
 | |
| 		if_id, ipc_mux->session[if_id].ul_flow_credits, ul_credits);
 | |
| 
 | |
| 	/* Update the Flow Credit information from ADB */
 | |
| 	ipc_mux->session[if_id].ul_flow_credits += ul_credits;
 | |
| 
 | |
| 	/* Check whether the TX can be started */
 | |
| 	if (ipc_mux->session[if_id].ul_flow_credits > 0) {
 | |
| 		ipc_mux->session[if_id].net_tx_stop = false;
 | |
| 		ipc_mux_netif_tx_flowctrl(&ipc_mux->session[if_id],
 | |
| 					  ipc_mux->session[if_id].if_id, false);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* Decode non-aggregated datagram */
 | |
| static void ipc_mux_dl_adgh_decode(struct iosm_mux *ipc_mux,
 | |
| 				   struct sk_buff *skb)
 | |
| {
 | |
| 	u32 pad_len, packet_offset, adgh_len;
 | |
| 	struct iosm_wwan *wwan;
 | |
| 	struct mux_adgh *adgh;
 | |
| 	u8 *block = skb->data;
 | |
| 	int rc = 0;
 | |
| 	u8 if_id;
 | |
| 
 | |
| 	adgh = (struct mux_adgh *)block;
 | |
| 
 | |
| 	if (adgh->signature != cpu_to_le32(IOSM_AGGR_MUX_SIG_ADGH)) {
 | |
| 		dev_err(ipc_mux->dev, "invalid ADGH signature received");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	if_id = adgh->if_id;
 | |
| 	if (if_id >= IPC_MEM_MUX_IP_SESSION_ENTRIES) {
 | |
| 		dev_err(ipc_mux->dev, "invalid if_id while decoding %d", if_id);
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/* Is the session active ? */
 | |
| 	if_id = array_index_nospec(if_id, IPC_MEM_MUX_IP_SESSION_ENTRIES);
 | |
| 	wwan = ipc_mux->session[if_id].wwan;
 | |
| 	if (!wwan) {
 | |
| 		dev_err(ipc_mux->dev, "session Net ID is NULL");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	/* Store the pad len for the corresponding session
 | |
| 	 * Pad bytes as negotiated in the open session less the header size
 | |
| 	 * (see session management chapter for details).
 | |
| 	 * If resulting padding is zero or less, the additional head padding is
 | |
| 	 * omitted. For e.g., if HEAD_PAD_LEN = 16 or less, this field is
 | |
| 	 * omitted if HEAD_PAD_LEN = 20, then this field will have 4 bytes
 | |
| 	 * set to zero
 | |
| 	 */
 | |
| 	pad_len =
 | |
| 		ipc_mux->session[if_id].dl_head_pad_len - IPC_MEM_DL_ETH_OFFSET;
 | |
| 	packet_offset = sizeof(*adgh) + pad_len;
 | |
| 
 | |
| 	if_id += ipc_mux->wwan_q_offset;
 | |
| 	adgh_len = le16_to_cpu(adgh->length);
 | |
| 
 | |
| 	/* Pass the packet to the netif layer */
 | |
| 	rc = ipc_mux_net_receive(ipc_mux, if_id, wwan, packet_offset,
 | |
| 				 adgh->service_class, skb,
 | |
| 				 adgh_len - packet_offset);
 | |
| 	if (rc) {
 | |
| 		dev_err(ipc_mux->dev, "mux adgh decoding error");
 | |
| 		return;
 | |
| 	}
 | |
| 	ipc_mux->session[if_id].flush = 1;
 | |
| }
 | |
| 
 | |
| static void ipc_mux_dl_acbcmd_decode(struct iosm_mux *ipc_mux,
 | |
| 				     struct mux_cmdh *cmdh, int size)
 | |
| {
 | |
| 	u32 link_st  = IOSM_AGGR_MUX_CMD_LINK_STATUS_REPORT_RESP;
 | |
| 	u32 fctl_dis = IOSM_AGGR_MUX_CMD_FLOW_CTL_DISABLE;
 | |
| 	u32 fctl_ena = IOSM_AGGR_MUX_CMD_FLOW_CTL_ENABLE;
 | |
| 	u32 fctl_ack = IOSM_AGGR_MUX_CMD_FLOW_CTL_ACK;
 | |
| 	union mux_cmd_param *cmd_p = NULL;
 | |
| 	u32 cmd = link_st;
 | |
| 	u32 trans_id;
 | |
| 
 | |
| 	if (!ipc_mux_dl_cmds_decode_process(ipc_mux, &cmdh->param,
 | |
| 					    cmdh->command_type, cmdh->if_id,
 | |
| 					    cmdh->cmd_len, size)) {
 | |
| 		size = 0;
 | |
| 		if (cmdh->command_type == cpu_to_le32(link_st)) {
 | |
| 			cmd_p = &cmdh->param;
 | |
| 			cmd_p->link_status_resp.response = MUX_CMD_RESP_SUCCESS;
 | |
| 		} else if ((cmdh->command_type == cpu_to_le32(fctl_ena)) ||
 | |
| 				(cmdh->command_type == cpu_to_le32(fctl_dis))) {
 | |
| 			cmd = fctl_ack;
 | |
| 		} else {
 | |
| 			return;
 | |
| 			}
 | |
| 		trans_id = le32_to_cpu(cmdh->transaction_id);
 | |
| 		ipc_mux_dl_acb_send_cmds(ipc_mux, cmd, cmdh->if_id,
 | |
| 					 trans_id, cmd_p, size, false, true);
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* Decode an aggregated command block. */
 | |
| static void ipc_mux_dl_acb_decode(struct iosm_mux *ipc_mux, struct sk_buff *skb)
 | |
| {
 | |
| 	struct mux_acbh *acbh;
 | |
| 	struct mux_cmdh *cmdh;
 | |
| 	u32 next_cmd_index;
 | |
| 	u8 *block;
 | |
| 	int size;
 | |
| 
 | |
| 	acbh = (struct mux_acbh *)(skb->data);
 | |
| 	block = (u8 *)(skb->data);
 | |
| 
 | |
| 	next_cmd_index = le32_to_cpu(acbh->first_cmd_index);
 | |
| 	next_cmd_index = array_index_nospec(next_cmd_index,
 | |
| 					    sizeof(struct mux_cmdh));
 | |
| 
 | |
| 	while (next_cmd_index != 0) {
 | |
| 		cmdh = (struct mux_cmdh *)&block[next_cmd_index];
 | |
| 		next_cmd_index = le32_to_cpu(cmdh->next_cmd_index);
 | |
| 		if (ipc_mux_dl_cmdresps_decode_process(ipc_mux, cmdh->param,
 | |
| 						       cmdh->command_type,
 | |
| 						       cmdh->if_id,
 | |
| 						       cmdh->transaction_id)) {
 | |
| 			size = offsetof(struct mux_cmdh, param) +
 | |
| 				sizeof(cmdh->param.flow_ctl);
 | |
| 			ipc_mux_dl_acbcmd_decode(ipc_mux, cmdh, size);
 | |
| 		}
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* process datagram */
 | |
| static int mux_dl_process_dg(struct iosm_mux *ipc_mux, struct mux_adbh *adbh,
 | |
| 			     struct mux_adth_dg *dg, struct sk_buff *skb,
 | |
| 			     int if_id, int nr_of_dg)
 | |
| {
 | |
| 	u32 dl_head_pad_len = ipc_mux->session[if_id].dl_head_pad_len;
 | |
| 	u32 packet_offset, i, rc, dg_len;
 | |
| 
 | |
| 	for (i = 0; i < nr_of_dg; i++, dg++) {
 | |
| 		if (le32_to_cpu(dg->datagram_index)
 | |
| 				< sizeof(struct mux_adbh))
 | |
| 			goto dg_error;
 | |
| 
 | |
| 		/* Is the packet inside of the ADB */
 | |
| 		if (le32_to_cpu(dg->datagram_index) >=
 | |
| 					le32_to_cpu(adbh->block_length)) {
 | |
| 			goto dg_error;
 | |
| 		} else {
 | |
| 			packet_offset =
 | |
| 				le32_to_cpu(dg->datagram_index) +
 | |
| 				dl_head_pad_len;
 | |
| 			dg_len = le16_to_cpu(dg->datagram_length);
 | |
| 			/* Pass the packet to the netif layer. */
 | |
| 			rc = ipc_mux_net_receive(ipc_mux, if_id, ipc_mux->wwan,
 | |
| 						 packet_offset,
 | |
| 						 dg->service_class, skb,
 | |
| 						 dg_len - dl_head_pad_len);
 | |
| 			if (rc)
 | |
| 				goto dg_error;
 | |
| 		}
 | |
| 	}
 | |
| 	return 0;
 | |
| dg_error:
 | |
| 	return -1;
 | |
| }
 | |
| 
 | |
| /* Decode an aggregated data block. */
 | |
| static void mux_dl_adb_decode(struct iosm_mux *ipc_mux,
 | |
| 			      struct sk_buff *skb)
 | |
| {
 | |
| 	struct mux_adth_dg *dg;
 | |
| 	struct iosm_wwan *wwan;
 | |
| 	struct mux_adbh *adbh;
 | |
| 	struct mux_adth *adth;
 | |
| 	int nr_of_dg, if_id;
 | |
| 	u32 adth_index;
 | |
| 	u8 *block;
 | |
| 
 | |
| 	block = skb->data;
 | |
| 	adbh = (struct mux_adbh *)block;
 | |
| 
 | |
| 	/* Process the aggregated datagram tables. */
 | |
| 	adth_index = le32_to_cpu(adbh->first_table_index);
 | |
| 
 | |
| 	/* Has CP sent an empty ADB ? */
 | |
| 	if (adth_index < 1) {
 | |
| 		dev_err(ipc_mux->dev, "unexpected empty ADB");
 | |
| 		goto adb_decode_err;
 | |
| 	}
 | |
| 
 | |
| 	/* Loop through mixed session tables. */
 | |
| 	while (adth_index) {
 | |
| 		/* Get the reference to the table header. */
 | |
| 		adth = (struct mux_adth *)(block + adth_index);
 | |
| 
 | |
| 		/* Get the interface id and map it to the netif id. */
 | |
| 		if_id = adth->if_id;
 | |
| 		if (if_id >= IPC_MEM_MUX_IP_SESSION_ENTRIES)
 | |
| 			goto adb_decode_err;
 | |
| 
 | |
| 		if_id = array_index_nospec(if_id,
 | |
| 					   IPC_MEM_MUX_IP_SESSION_ENTRIES);
 | |
| 
 | |
| 		/* Is the session active ? */
 | |
| 		wwan = ipc_mux->session[if_id].wwan;
 | |
| 		if (!wwan)
 | |
| 			goto adb_decode_err;
 | |
| 
 | |
| 		/* Consistency checks for aggregated datagram table. */
 | |
| 		if (adth->signature != cpu_to_le32(IOSM_AGGR_MUX_SIG_ADTH))
 | |
| 			goto adb_decode_err;
 | |
| 
 | |
| 		if (le16_to_cpu(adth->table_length) < (sizeof(struct mux_adth) -
 | |
| 				sizeof(struct mux_adth_dg)))
 | |
| 			goto adb_decode_err;
 | |
| 
 | |
| 		/* Calculate the number of datagrams. */
 | |
| 		nr_of_dg = (le16_to_cpu(adth->table_length) -
 | |
| 					sizeof(struct mux_adth) +
 | |
| 					sizeof(struct mux_adth_dg)) /
 | |
| 					sizeof(struct mux_adth_dg);
 | |
| 
 | |
| 		/* Is the datagram table empty ? */
 | |
| 		if (nr_of_dg < 1) {
 | |
| 			dev_err(ipc_mux->dev,
 | |
| 				"adthidx=%u,nr_of_dg=%d,next_tblidx=%u",
 | |
| 				adth_index, nr_of_dg,
 | |
| 				le32_to_cpu(adth->next_table_index));
 | |
| 
 | |
| 			/* Move to the next aggregated datagram table. */
 | |
| 			adth_index = le32_to_cpu(adth->next_table_index);
 | |
| 			continue;
 | |
| 		}
 | |
| 
 | |
| 		/* New aggregated datagram table. */
 | |
| 		dg = &adth->dg;
 | |
| 		if (mux_dl_process_dg(ipc_mux, adbh, dg, skb, if_id,
 | |
| 				      nr_of_dg) < 0)
 | |
| 			goto adb_decode_err;
 | |
| 
 | |
| 		/* mark session for final flush */
 | |
| 		ipc_mux->session[if_id].flush = 1;
 | |
| 
 | |
| 		/* Move to the next aggregated datagram table. */
 | |
| 		adth_index = le32_to_cpu(adth->next_table_index);
 | |
| 	}
 | |
| 
 | |
| adb_decode_err:
 | |
| 	return;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * ipc_mux_dl_decode -  Route the DL packet through the IP MUX layer
 | |
|  *                      depending on Header.
 | |
|  * @ipc_mux:            Pointer to MUX data-struct
 | |
|  * @skb:                Pointer to ipc_skb.
 | |
|  */
 | |
| void ipc_mux_dl_decode(struct iosm_mux *ipc_mux, struct sk_buff *skb)
 | |
| {
 | |
| 	u32 signature;
 | |
| 
 | |
| 	if (!skb->data)
 | |
| 		return;
 | |
| 
 | |
| 	/* Decode the MUX header type. */
 | |
| 	signature = le32_to_cpup((__le32 *)skb->data);
 | |
| 
 | |
| 	switch (signature) {
 | |
| 	case IOSM_AGGR_MUX_SIG_ADBH:	/* Aggregated Data Block Header */
 | |
| 		mux_dl_adb_decode(ipc_mux, skb);
 | |
| 		break;
 | |
| 	case IOSM_AGGR_MUX_SIG_ADGH:
 | |
| 		ipc_mux_dl_adgh_decode(ipc_mux, skb);
 | |
| 		break;
 | |
| 	case MUX_SIG_FCTH:
 | |
| 		ipc_mux_dl_fcth_decode(ipc_mux, skb->data);
 | |
| 		break;
 | |
| 	case IOSM_AGGR_MUX_SIG_ACBH:	/* Aggregated Command Block Header */
 | |
| 		ipc_mux_dl_acb_decode(ipc_mux, skb);
 | |
| 		break;
 | |
| 	case MUX_SIG_CMDH:
 | |
| 		ipc_mux_dl_cmd_decode(ipc_mux, skb);
 | |
| 		break;
 | |
| 
 | |
| 	default:
 | |
| 		dev_err(ipc_mux->dev, "invalid ABH signature");
 | |
| 	}
 | |
| 
 | |
| 	ipc_pcie_kfree_skb(ipc_mux->pcie, skb);
 | |
| }
 | |
| 
 | |
| static int ipc_mux_ul_skb_alloc(struct iosm_mux *ipc_mux,
 | |
| 				struct mux_adb *ul_adb, u32 type)
 | |
| {
 | |
| 	/* Take the first element of the free list. */
 | |
| 	struct sk_buff *skb = skb_dequeue(&ul_adb->free_list);
 | |
| 	u32 no_if = IPC_MEM_MUX_IP_SESSION_ENTRIES;
 | |
| 	u32 *next_tb_id;
 | |
| 	int qlt_size;
 | |
| 	u32 if_id;
 | |
| 
 | |
| 	if (!skb)
 | |
| 		return -EBUSY; /* Wait for a free ADB skb. */
 | |
| 
 | |
| 	/* Mark it as UL ADB to select the right free operation. */
 | |
| 	IPC_CB(skb)->op_type = (u8)UL_MUX_OP_ADB;
 | |
| 
 | |
| 	switch (type) {
 | |
| 	case IOSM_AGGR_MUX_SIG_ADBH:
 | |
| 		/* Save the ADB memory settings. */
 | |
| 		ul_adb->dest_skb = skb;
 | |
| 		ul_adb->buf = skb->data;
 | |
| 		ul_adb->size = IPC_MEM_MAX_ADB_BUF_SIZE;
 | |
| 
 | |
| 		/* reset statistic counter */
 | |
| 		ul_adb->if_cnt = 0;
 | |
| 		ul_adb->payload_size = 0;
 | |
| 		ul_adb->dg_cnt_total = 0;
 | |
| 
 | |
| 		/* Initialize the ADBH. */
 | |
| 		ul_adb->adbh = (struct mux_adbh *)ul_adb->buf;
 | |
| 		memset(ul_adb->adbh, 0, sizeof(struct mux_adbh));
 | |
| 		ul_adb->adbh->signature = cpu_to_le32(IOSM_AGGR_MUX_SIG_ADBH);
 | |
| 		ul_adb->adbh->block_length =
 | |
| 					cpu_to_le32(sizeof(struct mux_adbh));
 | |
| 		next_tb_id = (unsigned int *)&ul_adb->adbh->first_table_index;
 | |
| 		ul_adb->next_table_index = next_tb_id;
 | |
| 
 | |
| 		/* Clear the local copy of DGs for new ADB */
 | |
| 		memset(ul_adb->dg, 0, sizeof(ul_adb->dg));
 | |
| 
 | |
| 		/* Clear the DG count and QLT updated status for new ADB */
 | |
| 		for (if_id = 0; if_id < no_if; if_id++) {
 | |
| 			ul_adb->dg_count[if_id] = 0;
 | |
| 			ul_adb->qlt_updated[if_id] = 0;
 | |
| 		}
 | |
| 		break;
 | |
| 
 | |
| 	case IOSM_AGGR_MUX_SIG_ADGH:
 | |
| 		/* Save the ADB memory settings. */
 | |
| 		ul_adb->dest_skb = skb;
 | |
| 		ul_adb->buf = skb->data;
 | |
| 		ul_adb->size = IPC_MEM_MAX_DL_MUX_LITE_BUF_SIZE;
 | |
| 		/* reset statistic counter */
 | |
| 		ul_adb->if_cnt = 0;
 | |
| 		ul_adb->payload_size = 0;
 | |
| 		ul_adb->dg_cnt_total = 0;
 | |
| 
 | |
| 		ul_adb->adgh = (struct mux_adgh *)skb->data;
 | |
| 		memset(ul_adb->adgh, 0, sizeof(struct mux_adgh));
 | |
| 		break;
 | |
| 
 | |
| 	case MUX_SIG_QLTH:
 | |
| 		qlt_size = offsetof(struct ipc_mem_lite_gen_tbl, vfl) +
 | |
| 			   (MUX_QUEUE_LEVEL * sizeof(struct mux_lite_vfl));
 | |
| 
 | |
| 		if (qlt_size > IPC_MEM_MAX_DL_MUX_LITE_BUF_SIZE) {
 | |
| 			dev_err(ipc_mux->dev,
 | |
| 				"can't support. QLT size:%d SKB size: %d",
 | |
| 				qlt_size, IPC_MEM_MAX_DL_MUX_LITE_BUF_SIZE);
 | |
| 			return -ERANGE;
 | |
| 		}
 | |
| 
 | |
| 		ul_adb->qlth_skb = skb;
 | |
| 		memset((ul_adb->qlth_skb)->data, 0, qlt_size);
 | |
| 		skb_put(skb, qlt_size);
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| static void ipc_mux_ul_adgh_finish(struct iosm_mux *ipc_mux)
 | |
| {
 | |
| 	struct mux_adb *ul_adb = &ipc_mux->ul_adb;
 | |
| 	u16 adgh_len;
 | |
| 	long long bytes;
 | |
| 	char *str;
 | |
| 
 | |
| 	if (!ul_adb->dest_skb) {
 | |
| 		dev_err(ipc_mux->dev, "no dest skb");
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	adgh_len = le16_to_cpu(ul_adb->adgh->length);
 | |
| 	skb_put(ul_adb->dest_skb, adgh_len);
 | |
| 	skb_queue_tail(&ipc_mux->channel->ul_list, ul_adb->dest_skb);
 | |
| 	ul_adb->dest_skb = NULL;
 | |
| 
 | |
| 	if (ipc_mux->ul_flow == MUX_UL_ON_CREDITS) {
 | |
| 		struct mux_session *session;
 | |
| 
 | |
| 		session = &ipc_mux->session[ul_adb->adgh->if_id];
 | |
| 		str = "available_credits";
 | |
| 		bytes = (long long)session->ul_flow_credits;
 | |
| 
 | |
| 	} else {
 | |
| 		str = "pend_bytes";
 | |
| 		bytes = ipc_mux->ul_data_pend_bytes;
 | |
| 		ipc_mux->ul_data_pend_bytes = ipc_mux->ul_data_pend_bytes +
 | |
| 					      adgh_len;
 | |
| 	}
 | |
| 
 | |
| 	dev_dbg(ipc_mux->dev, "UL ADGH: size=%u, if_id=%d, payload=%d, %s=%lld",
 | |
| 		adgh_len, ul_adb->adgh->if_id, ul_adb->payload_size,
 | |
| 		str, bytes);
 | |
| }
 | |
| 
 | |
| static void ipc_mux_ul_encode_adth(struct iosm_mux *ipc_mux,
 | |
| 				   struct mux_adb *ul_adb, int *out_offset)
 | |
| {
 | |
| 	int i, qlt_size, offset = *out_offset;
 | |
| 	struct mux_qlth *p_adb_qlt;
 | |
| 	struct mux_adth_dg *dg;
 | |
| 	struct mux_adth *adth;
 | |
| 	u16 adth_dg_size;
 | |
| 	u32 *next_tb_id;
 | |
| 
 | |
| 	qlt_size = offsetof(struct mux_qlth, ql) +
 | |
| 			MUX_QUEUE_LEVEL * sizeof(struct mux_qlth_ql);
 | |
| 
 | |
| 	for (i = 0; i < ipc_mux->nr_sessions; i++) {
 | |
| 		if (ul_adb->dg_count[i] > 0) {
 | |
| 			adth_dg_size = offsetof(struct mux_adth, dg) +
 | |
| 					ul_adb->dg_count[i] * sizeof(*dg);
 | |
| 
 | |
| 			*ul_adb->next_table_index = offset;
 | |
| 			adth = (struct mux_adth *)&ul_adb->buf[offset];
 | |
| 			next_tb_id = (unsigned int *)&adth->next_table_index;
 | |
| 			ul_adb->next_table_index = next_tb_id;
 | |
| 			offset += adth_dg_size;
 | |
| 			adth->signature = cpu_to_le32(IOSM_AGGR_MUX_SIG_ADTH);
 | |
| 			adth->if_id = i;
 | |
| 			adth->table_length = cpu_to_le16(adth_dg_size);
 | |
| 			adth_dg_size -= offsetof(struct mux_adth, dg);
 | |
| 			memcpy(&adth->dg, ul_adb->dg[i], adth_dg_size);
 | |
| 			ul_adb->if_cnt++;
 | |
| 		}
 | |
| 
 | |
| 		if (ul_adb->qlt_updated[i]) {
 | |
| 			*ul_adb->next_table_index = offset;
 | |
| 			p_adb_qlt = (struct mux_qlth *)&ul_adb->buf[offset];
 | |
| 			ul_adb->next_table_index =
 | |
| 				(u32 *)&p_adb_qlt->next_table_index;
 | |
| 			memcpy(p_adb_qlt, ul_adb->pp_qlt[i], qlt_size);
 | |
| 			offset += qlt_size;
 | |
| 		}
 | |
| 	}
 | |
| 	*out_offset = offset;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * ipc_mux_ul_adb_finish - Add the TD of the aggregated session packets to TDR.
 | |
|  * @ipc_mux:               Pointer to MUX data-struct.
 | |
|  */
 | |
| void ipc_mux_ul_adb_finish(struct iosm_mux *ipc_mux)
 | |
| {
 | |
| 	bool ul_data_pend = false;
 | |
| 	struct mux_adb *ul_adb;
 | |
| 	unsigned long flags;
 | |
| 	int offset;
 | |
| 
 | |
| 	ul_adb = &ipc_mux->ul_adb;
 | |
| 	if (!ul_adb->dest_skb)
 | |
| 		return;
 | |
| 
 | |
| 	offset = *ul_adb->next_table_index;
 | |
| 	ipc_mux_ul_encode_adth(ipc_mux, ul_adb, &offset);
 | |
| 	ul_adb->adbh->block_length = cpu_to_le32(offset);
 | |
| 
 | |
| 	if (le32_to_cpu(ul_adb->adbh->block_length) > ul_adb->size) {
 | |
| 		ul_adb->dest_skb = NULL;
 | |
| 		return;
 | |
| 	}
 | |
| 
 | |
| 	*ul_adb->next_table_index = 0;
 | |
| 	ul_adb->adbh->sequence_nr = cpu_to_le16(ipc_mux->adb_tx_sequence_nr++);
 | |
| 	skb_put(ul_adb->dest_skb, le32_to_cpu(ul_adb->adbh->block_length));
 | |
| 
 | |
| 	spin_lock_irqsave(&(&ipc_mux->channel->ul_list)->lock, flags);
 | |
| 	__skb_queue_tail(&ipc_mux->channel->ul_list,  ul_adb->dest_skb);
 | |
| 	spin_unlock_irqrestore(&(&ipc_mux->channel->ul_list)->lock, flags);
 | |
| 
 | |
| 	ul_adb->dest_skb = NULL;
 | |
| 	/* Updates the TDs with ul_list */
 | |
| 	ul_data_pend = ipc_imem_ul_write_td(ipc_mux->imem);
 | |
| 
 | |
| 	/* Delay the doorbell irq */
 | |
| 	if (ul_data_pend)
 | |
| 		ipc_imem_td_update_timer_start(ipc_mux->imem);
 | |
| 
 | |
| 	ipc_mux->acc_adb_size +=  le32_to_cpu(ul_adb->adbh->block_length);
 | |
| 	ipc_mux->acc_payload_size += ul_adb->payload_size;
 | |
| 	ipc_mux->ul_data_pend_bytes += ul_adb->payload_size;
 | |
| }
 | |
| 
 | |
| /* Allocates an ADB from the free list and initializes it with ADBH  */
 | |
| static bool ipc_mux_ul_adb_allocate(struct iosm_mux *ipc_mux,
 | |
| 				    struct mux_adb *adb, int *size_needed,
 | |
| 				    u32 type)
 | |
| {
 | |
| 	bool ret_val = false;
 | |
| 	int status;
 | |
| 
 | |
| 	if (!adb->dest_skb) {
 | |
| 		/* Allocate memory for the ADB including of the
 | |
| 		 * datagram table header.
 | |
| 		 */
 | |
| 		status = ipc_mux_ul_skb_alloc(ipc_mux, adb, type);
 | |
| 		if (status)
 | |
| 			/* Is a pending ADB available ? */
 | |
| 			ret_val = true; /* None. */
 | |
| 
 | |
| 		/* Update size need to zero only for new ADB memory */
 | |
| 		*size_needed = 0;
 | |
| 	}
 | |
| 
 | |
| 	return ret_val;
 | |
| }
 | |
| 
 | |
| /* Informs the network stack to stop sending further packets for all opened
 | |
|  * sessions
 | |
|  */
 | |
| static void ipc_mux_stop_tx_for_all_sessions(struct iosm_mux *ipc_mux)
 | |
| {
 | |
| 	struct mux_session *session;
 | |
| 	int idx;
 | |
| 
 | |
| 	for (idx = 0; idx < IPC_MEM_MUX_IP_SESSION_ENTRIES; idx++) {
 | |
| 		session = &ipc_mux->session[idx];
 | |
| 
 | |
| 		if (!session->wwan)
 | |
| 			continue;
 | |
| 
 | |
| 		session->net_tx_stop = true;
 | |
| 	}
 | |
| }
 | |
| 
 | |
| /* Sends Queue Level Table of all opened sessions */
 | |
| static bool ipc_mux_lite_send_qlt(struct iosm_mux *ipc_mux)
 | |
| {
 | |
| 	struct ipc_mem_lite_gen_tbl *qlt;
 | |
| 	struct mux_session *session;
 | |
| 	bool qlt_updated = false;
 | |
| 	int i;
 | |
| 	int qlt_size;
 | |
| 
 | |
| 	if (!ipc_mux->initialized || ipc_mux->state != MUX_S_ACTIVE)
 | |
| 		return qlt_updated;
 | |
| 
 | |
| 	qlt_size = offsetof(struct ipc_mem_lite_gen_tbl, vfl) +
 | |
| 		   MUX_QUEUE_LEVEL * sizeof(struct mux_lite_vfl);
 | |
| 
 | |
| 	for (i = 0; i < IPC_MEM_MUX_IP_SESSION_ENTRIES; i++) {
 | |
| 		session = &ipc_mux->session[i];
 | |
| 
 | |
| 		if (!session->wwan || session->flow_ctl_mask)
 | |
| 			continue;
 | |
| 
 | |
| 		if (ipc_mux_ul_skb_alloc(ipc_mux, &ipc_mux->ul_adb,
 | |
| 					 MUX_SIG_QLTH)) {
 | |
| 			dev_err(ipc_mux->dev,
 | |
| 				"no reserved mem to send QLT of if_id: %d", i);
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 		/* Prepare QLT */
 | |
| 		qlt = (struct ipc_mem_lite_gen_tbl *)(ipc_mux->ul_adb.qlth_skb)
 | |
| 			      ->data;
 | |
| 		qlt->signature = cpu_to_le32(MUX_SIG_QLTH);
 | |
| 		qlt->length = cpu_to_le16(qlt_size);
 | |
| 		qlt->if_id = i;
 | |
| 		qlt->vfl_length = MUX_QUEUE_LEVEL * sizeof(struct mux_lite_vfl);
 | |
| 		qlt->reserved[0] = 0;
 | |
| 		qlt->reserved[1] = 0;
 | |
| 
 | |
| 		qlt->vfl.nr_of_bytes = cpu_to_le32(session->ul_list.qlen);
 | |
| 
 | |
| 		/* Add QLT to the transfer list. */
 | |
| 		skb_queue_tail(&ipc_mux->channel->ul_list,
 | |
| 			       ipc_mux->ul_adb.qlth_skb);
 | |
| 
 | |
| 		qlt_updated = true;
 | |
| 		ipc_mux->ul_adb.qlth_skb = NULL;
 | |
| 	}
 | |
| 
 | |
| 	if (qlt_updated)
 | |
| 		/* Updates the TDs with ul_list */
 | |
| 		(void)ipc_imem_ul_write_td(ipc_mux->imem);
 | |
| 
 | |
| 	return qlt_updated;
 | |
| }
 | |
| 
 | |
| /* Checks the available credits for the specified session and returns
 | |
|  * number of packets for which credits are available.
 | |
|  */
 | |
| static int ipc_mux_ul_bytes_credits_check(struct iosm_mux *ipc_mux,
 | |
| 					  struct mux_session *session,
 | |
| 					  struct sk_buff_head *ul_list,
 | |
| 					  int max_nr_of_pkts)
 | |
| {
 | |
| 	int pkts_to_send = 0;
 | |
| 	struct sk_buff *skb;
 | |
| 	int credits = 0;
 | |
| 
 | |
| 	if (ipc_mux->ul_flow == MUX_UL_ON_CREDITS) {
 | |
| 		credits = session->ul_flow_credits;
 | |
| 		if (credits <= 0) {
 | |
| 			dev_dbg(ipc_mux->dev,
 | |
| 				"FC::if_id[%d] Insuff.Credits/Qlen:%d/%u",
 | |
| 				session->if_id, session->ul_flow_credits,
 | |
| 				session->ul_list.qlen); /* nr_of_bytes */
 | |
| 			return 0;
 | |
| 		}
 | |
| 	} else {
 | |
| 		credits = IPC_MEM_MUX_UL_FLOWCTRL_HIGH_B -
 | |
| 			  ipc_mux->ul_data_pend_bytes;
 | |
| 		if (credits <= 0) {
 | |
| 			ipc_mux_stop_tx_for_all_sessions(ipc_mux);
 | |
| 
 | |
| 			dev_dbg(ipc_mux->dev,
 | |
| 				"if_id[%d] encod. fail Bytes: %llu, thresh: %d",
 | |
| 				session->if_id, ipc_mux->ul_data_pend_bytes,
 | |
| 				IPC_MEM_MUX_UL_FLOWCTRL_HIGH_B);
 | |
| 			return 0;
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	/* Check if there are enough credits/bytes available to send the
 | |
| 	 * requested max_nr_of_pkts. Otherwise restrict the nr_of_pkts
 | |
| 	 * depending on available credits.
 | |
| 	 */
 | |
| 	skb_queue_walk(ul_list, skb)
 | |
| 	{
 | |
| 		if (!(credits >= skb->len && pkts_to_send < max_nr_of_pkts))
 | |
| 			break;
 | |
| 		credits -= skb->len;
 | |
| 		pkts_to_send++;
 | |
| 	}
 | |
| 
 | |
| 	return pkts_to_send;
 | |
| }
 | |
| 
 | |
| /* Encode the UL IP packet according to Lite spec. */
 | |
| static int ipc_mux_ul_adgh_encode(struct iosm_mux *ipc_mux, int session_id,
 | |
| 				  struct mux_session *session,
 | |
| 				  struct sk_buff_head *ul_list,
 | |
| 				  struct mux_adb *adb, int nr_of_pkts)
 | |
| {
 | |
| 	int offset = sizeof(struct mux_adgh);
 | |
| 	int adb_updated = -EINVAL;
 | |
| 	struct sk_buff *src_skb;
 | |
| 	int aligned_size = 0;
 | |
| 	int nr_of_skb = 0;
 | |
| 	u32 pad_len = 0;
 | |
| 
 | |
| 	/* Re-calculate the number of packets depending on number of bytes to be
 | |
| 	 * processed/available credits.
 | |
| 	 */
 | |
| 	nr_of_pkts = ipc_mux_ul_bytes_credits_check(ipc_mux, session, ul_list,
 | |
| 						    nr_of_pkts);
 | |
| 
 | |
| 	/* If calculated nr_of_pkts from available credits is <= 0
 | |
| 	 * then nothing to do.
 | |
| 	 */
 | |
| 	if (nr_of_pkts <= 0)
 | |
| 		return 0;
 | |
| 
 | |
| 	/* Read configured UL head_pad_length for session.*/
 | |
| 	if (session->ul_head_pad_len > IPC_MEM_DL_ETH_OFFSET)
 | |
| 		pad_len = session->ul_head_pad_len - IPC_MEM_DL_ETH_OFFSET;
 | |
| 
 | |
| 	/* Process all pending UL packets for this session
 | |
| 	 * depending on the allocated datagram table size.
 | |
| 	 */
 | |
| 	while (nr_of_pkts > 0) {
 | |
| 		/* get destination skb allocated */
 | |
| 		if (ipc_mux_ul_adb_allocate(ipc_mux, adb, &ipc_mux->size_needed,
 | |
| 					    IOSM_AGGR_MUX_SIG_ADGH)) {
 | |
| 			dev_err(ipc_mux->dev, "no reserved memory for ADGH");
 | |
| 			return -ENOMEM;
 | |
| 		}
 | |
| 
 | |
| 		/* Peek at the head of the list. */
 | |
| 		src_skb = skb_peek(ul_list);
 | |
| 		if (!src_skb) {
 | |
| 			dev_err(ipc_mux->dev,
 | |
| 				"skb peek return NULL with count : %d",
 | |
| 				nr_of_pkts);
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 		/* Calculate the memory value. */
 | |
| 		aligned_size = ALIGN((pad_len + src_skb->len), 4);
 | |
| 
 | |
| 		ipc_mux->size_needed = sizeof(struct mux_adgh) + aligned_size;
 | |
| 
 | |
| 		if (ipc_mux->size_needed > adb->size) {
 | |
| 			dev_dbg(ipc_mux->dev, "size needed %d, adgh size %d",
 | |
| 				ipc_mux->size_needed, adb->size);
 | |
| 			/* Return 1 if any IP packet is added to the transfer
 | |
| 			 * list.
 | |
| 			 */
 | |
| 			return nr_of_skb ? 1 : 0;
 | |
| 		}
 | |
| 
 | |
| 		/* Add buffer (without head padding to next pending transfer) */
 | |
| 		memcpy(adb->buf + offset + pad_len, src_skb->data,
 | |
| 		       src_skb->len);
 | |
| 
 | |
| 		adb->adgh->signature = cpu_to_le32(IOSM_AGGR_MUX_SIG_ADGH);
 | |
| 		adb->adgh->if_id = session_id;
 | |
| 		adb->adgh->length =
 | |
| 			cpu_to_le16(sizeof(struct mux_adgh) + pad_len +
 | |
| 				    src_skb->len);
 | |
| 		adb->adgh->service_class = src_skb->priority;
 | |
| 		adb->adgh->next_count = --nr_of_pkts;
 | |
| 		adb->dg_cnt_total++;
 | |
| 		adb->payload_size += src_skb->len;
 | |
| 
 | |
| 		if (ipc_mux->ul_flow == MUX_UL_ON_CREDITS)
 | |
| 			/* Decrement the credit value as we are processing the
 | |
| 			 * datagram from the UL list.
 | |
| 			 */
 | |
| 			session->ul_flow_credits -= src_skb->len;
 | |
| 
 | |
| 		/* Remove the processed elements and free it. */
 | |
| 		src_skb = skb_dequeue(ul_list);
 | |
| 		dev_kfree_skb(src_skb);
 | |
| 		nr_of_skb++;
 | |
| 
 | |
| 		ipc_mux_ul_adgh_finish(ipc_mux);
 | |
| 	}
 | |
| 
 | |
| 	if (nr_of_skb) {
 | |
| 		/* Send QLT info to modem if pending bytes > high watermark
 | |
| 		 * in case of mux lite
 | |
| 		 */
 | |
| 		if (ipc_mux->ul_flow == MUX_UL_ON_CREDITS ||
 | |
| 		    ipc_mux->ul_data_pend_bytes >=
 | |
| 			    IPC_MEM_MUX_UL_FLOWCTRL_LOW_B)
 | |
| 			adb_updated = ipc_mux_lite_send_qlt(ipc_mux);
 | |
| 		else
 | |
| 			adb_updated = 1;
 | |
| 
 | |
| 		/* Updates the TDs with ul_list */
 | |
| 		(void)ipc_imem_ul_write_td(ipc_mux->imem);
 | |
| 	}
 | |
| 
 | |
| 	return adb_updated;
 | |
| }
 | |
| 
 | |
| /**
 | |
|  * ipc_mux_ul_adb_update_ql - Adds Queue Level Table and Queue Level to ADB
 | |
|  * @ipc_mux:            pointer to MUX instance data
 | |
|  * @p_adb:              pointer to UL aggegated data block
 | |
|  * @session_id:         session id
 | |
|  * @qlth_n_ql_size:     Length (in bytes) of the datagram table
 | |
|  * @ul_list:            pointer to skb buffer head
 | |
|  */
 | |
| void ipc_mux_ul_adb_update_ql(struct iosm_mux *ipc_mux, struct mux_adb *p_adb,
 | |
| 			      int session_id, int qlth_n_ql_size,
 | |
| 			      struct sk_buff_head *ul_list)
 | |
| {
 | |
| 	int qlevel = ul_list->qlen;
 | |
| 	struct mux_qlth *p_qlt;
 | |
| 
 | |
| 	p_qlt = (struct mux_qlth *)p_adb->pp_qlt[session_id];
 | |
| 
 | |
| 	/* Initialize QLTH if not been done */
 | |
| 	if (p_adb->qlt_updated[session_id] == 0) {
 | |
| 		p_qlt->signature = cpu_to_le32(MUX_SIG_QLTH);
 | |
| 		p_qlt->if_id = session_id;
 | |
| 		p_qlt->table_length = cpu_to_le16(qlth_n_ql_size);
 | |
| 		p_qlt->reserved = 0;
 | |
| 		p_qlt->reserved2 = 0;
 | |
| 	}
 | |
| 
 | |
| 	/* Update Queue Level information always */
 | |
| 	p_qlt->ql.nr_of_bytes = cpu_to_le32(qlevel);
 | |
| 	p_adb->qlt_updated[session_id] = 1;
 | |
| }
 | |
| 
 | |
| /* Update the next table index. */
 | |
| static int mux_ul_dg_update_tbl_index(struct iosm_mux *ipc_mux,
 | |
| 				      int session_id,
 | |
| 				      struct sk_buff_head *ul_list,
 | |
| 				      struct mux_adth_dg *dg,
 | |
| 				      int aligned_size,
 | |
| 				      u32 qlth_n_ql_size,
 | |
| 				      struct mux_adb *adb,
 | |
| 				      struct sk_buff *src_skb)
 | |
| {
 | |
| 	ipc_mux_ul_adb_update_ql(ipc_mux, adb, session_id,
 | |
| 				 qlth_n_ql_size, ul_list);
 | |
| 	ipc_mux_ul_adb_finish(ipc_mux);
 | |
| 	if (ipc_mux_ul_adb_allocate(ipc_mux, adb, &ipc_mux->size_needed,
 | |
| 				    IOSM_AGGR_MUX_SIG_ADBH))
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	ipc_mux->size_needed = le32_to_cpu(adb->adbh->block_length);
 | |
| 
 | |
| 	ipc_mux->size_needed += offsetof(struct mux_adth, dg);
 | |
| 	ipc_mux->size_needed += qlth_n_ql_size;
 | |
| 	ipc_mux->size_needed += sizeof(*dg) + aligned_size;
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| /* Process encode session UL data. */
 | |
| static int mux_ul_dg_encode(struct iosm_mux *ipc_mux, struct mux_adb *adb,
 | |
| 			    struct mux_adth_dg *dg,
 | |
| 			    struct sk_buff_head *ul_list,
 | |
| 			    struct sk_buff *src_skb, int session_id,
 | |
| 			    int pkt_to_send, u32 qlth_n_ql_size,
 | |
| 			    int *out_offset, int head_pad_len)
 | |
| {
 | |
| 	int aligned_size;
 | |
| 	int offset = *out_offset;
 | |
| 	unsigned long flags;
 | |
| 	int nr_of_skb = 0;
 | |
| 
 | |
| 	while (pkt_to_send > 0) {
 | |
| 		/* Peek at the head of the list. */
 | |
| 		src_skb = skb_peek(ul_list);
 | |
| 		if (!src_skb) {
 | |
| 			dev_err(ipc_mux->dev,
 | |
| 				"skb peek return NULL with count : %d",
 | |
| 				pkt_to_send);
 | |
| 			return -1;
 | |
| 		}
 | |
| 		aligned_size = ALIGN((head_pad_len + src_skb->len), 4);
 | |
| 		ipc_mux->size_needed += sizeof(*dg) + aligned_size;
 | |
| 
 | |
| 		if (ipc_mux->size_needed > adb->size ||
 | |
| 		    ((ipc_mux->size_needed + ipc_mux->ul_data_pend_bytes) >=
 | |
| 		      IPC_MEM_MUX_UL_FLOWCTRL_HIGH_B)) {
 | |
| 			*adb->next_table_index = offset;
 | |
| 			if (mux_ul_dg_update_tbl_index(ipc_mux, session_id,
 | |
| 						       ul_list, dg,
 | |
| 						       aligned_size,
 | |
| 						       qlth_n_ql_size, adb,
 | |
| 						       src_skb) < 0)
 | |
| 				return -ENOMEM;
 | |
| 			nr_of_skb = 0;
 | |
| 			offset = le32_to_cpu(adb->adbh->block_length);
 | |
| 			/* Load pointer to next available datagram entry */
 | |
| 			dg = adb->dg[session_id] + adb->dg_count[session_id];
 | |
| 		}
 | |
| 		/* Add buffer without head padding to next pending transfer. */
 | |
| 		memcpy(adb->buf + offset + head_pad_len,
 | |
| 		       src_skb->data, src_skb->len);
 | |
| 		/* Setup datagram entry. */
 | |
| 		dg->datagram_index = cpu_to_le32(offset);
 | |
| 		dg->datagram_length = cpu_to_le16(src_skb->len + head_pad_len);
 | |
| 		dg->service_class = (((struct sk_buff *)src_skb)->priority);
 | |
| 		dg->reserved = 0;
 | |
| 		adb->dg_cnt_total++;
 | |
| 		adb->payload_size += le16_to_cpu(dg->datagram_length);
 | |
| 		dg++;
 | |
| 		adb->dg_count[session_id]++;
 | |
| 		offset += aligned_size;
 | |
| 		/* Remove the processed elements and free it. */
 | |
| 		spin_lock_irqsave(&ul_list->lock, flags);
 | |
| 		src_skb = __skb_dequeue(ul_list);
 | |
| 		spin_unlock_irqrestore(&ul_list->lock, flags);
 | |
| 
 | |
| 		dev_kfree_skb(src_skb);
 | |
| 		nr_of_skb++;
 | |
| 		pkt_to_send--;
 | |
| 	}
 | |
| 	*out_offset = offset;
 | |
| 	return nr_of_skb;
 | |
| }
 | |
| 
 | |
| /* Process encode session UL data to ADB. */
 | |
| static int mux_ul_adb_encode(struct iosm_mux *ipc_mux, int session_id,
 | |
| 			     struct mux_session *session,
 | |
| 			     struct sk_buff_head *ul_list, struct mux_adb *adb,
 | |
| 			     int pkt_to_send)
 | |
| {
 | |
| 	int adb_updated = -EINVAL;
 | |
| 	int head_pad_len, offset;
 | |
| 	struct sk_buff *src_skb = NULL;
 | |
| 	struct mux_adth_dg *dg;
 | |
| 	u32 qlth_n_ql_size;
 | |
| 
 | |
| 	/* If any of the opened session has set Flow Control ON then limit the
 | |
| 	 * UL data to mux_flow_ctrl_high_thresh_b bytes
 | |
| 	 */
 | |
| 	if (ipc_mux->ul_data_pend_bytes >=
 | |
| 		IPC_MEM_MUX_UL_FLOWCTRL_HIGH_B) {
 | |
| 		ipc_mux_stop_tx_for_all_sessions(ipc_mux);
 | |
| 		return adb_updated;
 | |
| 	}
 | |
| 
 | |
| 	qlth_n_ql_size = offsetof(struct mux_qlth, ql) +
 | |
| 			 MUX_QUEUE_LEVEL * sizeof(struct mux_qlth_ql);
 | |
| 	head_pad_len = session->ul_head_pad_len;
 | |
| 
 | |
| 	if (session->ul_head_pad_len > IPC_MEM_DL_ETH_OFFSET)
 | |
| 		head_pad_len = session->ul_head_pad_len - IPC_MEM_DL_ETH_OFFSET;
 | |
| 
 | |
| 	if (ipc_mux_ul_adb_allocate(ipc_mux, adb, &ipc_mux->size_needed,
 | |
| 				    IOSM_AGGR_MUX_SIG_ADBH))
 | |
| 		return -ENOMEM;
 | |
| 
 | |
| 	offset = le32_to_cpu(adb->adbh->block_length);
 | |
| 
 | |
| 	if (ipc_mux->size_needed == 0)
 | |
| 		ipc_mux->size_needed = offset;
 | |
| 
 | |
| 	/* Calculate the size needed for ADTH, QLTH and QL*/
 | |
| 	if (adb->dg_count[session_id] == 0) {
 | |
| 		ipc_mux->size_needed += offsetof(struct mux_adth, dg);
 | |
| 		ipc_mux->size_needed += qlth_n_ql_size;
 | |
| 	}
 | |
| 
 | |
| 	dg = adb->dg[session_id] + adb->dg_count[session_id];
 | |
| 
 | |
| 	if (mux_ul_dg_encode(ipc_mux, adb, dg, ul_list, src_skb,
 | |
| 			     session_id, pkt_to_send, qlth_n_ql_size, &offset,
 | |
| 			     head_pad_len) > 0) {
 | |
| 		adb_updated = 1;
 | |
| 		*adb->next_table_index = offset;
 | |
| 		ipc_mux_ul_adb_update_ql(ipc_mux, adb, session_id,
 | |
| 					 qlth_n_ql_size, ul_list);
 | |
| 		adb->adbh->block_length = cpu_to_le32(offset);
 | |
| 	}
 | |
| 
 | |
| 	return adb_updated;
 | |
| }
 | |
| 
 | |
| bool ipc_mux_ul_data_encode(struct iosm_mux *ipc_mux)
 | |
| {
 | |
| 	struct sk_buff_head *ul_list;
 | |
| 	struct mux_session *session;
 | |
| 	int updated = 0;
 | |
| 	int session_id;
 | |
| 	int dg_n;
 | |
| 	int i;
 | |
| 
 | |
| 	if (!ipc_mux || ipc_mux->state != MUX_S_ACTIVE ||
 | |
| 	    ipc_mux->adb_prep_ongoing)
 | |
| 		return false;
 | |
| 
 | |
| 	ipc_mux->adb_prep_ongoing = true;
 | |
| 
 | |
| 	for (i = 0; i < IPC_MEM_MUX_IP_SESSION_ENTRIES; i++) {
 | |
| 		session_id = ipc_mux->rr_next_session;
 | |
| 		session = &ipc_mux->session[session_id];
 | |
| 
 | |
| 		/* Go to next handle rr_next_session overflow */
 | |
| 		ipc_mux->rr_next_session++;
 | |
| 		if (ipc_mux->rr_next_session >= IPC_MEM_MUX_IP_SESSION_ENTRIES)
 | |
| 			ipc_mux->rr_next_session = 0;
 | |
| 
 | |
| 		if (!session->wwan || session->flow_ctl_mask ||
 | |
| 		    session->net_tx_stop)
 | |
| 			continue;
 | |
| 
 | |
| 		ul_list = &session->ul_list;
 | |
| 
 | |
| 		/* Is something pending in UL and flow ctrl off */
 | |
| 		dg_n = skb_queue_len(ul_list);
 | |
| 		if (dg_n > MUX_MAX_UL_DG_ENTRIES)
 | |
| 			dg_n = MUX_MAX_UL_DG_ENTRIES;
 | |
| 
 | |
| 		if (dg_n == 0)
 | |
| 			/* Nothing to do for ipc_mux session
 | |
| 			 * -> try next session id.
 | |
| 			 */
 | |
| 			continue;
 | |
| 		if (ipc_mux->protocol == MUX_LITE)
 | |
| 			updated = ipc_mux_ul_adgh_encode(ipc_mux, session_id,
 | |
| 							 session, ul_list,
 | |
| 							 &ipc_mux->ul_adb,
 | |
| 							 dg_n);
 | |
| 		else
 | |
| 			updated = mux_ul_adb_encode(ipc_mux, session_id,
 | |
| 						    session, ul_list,
 | |
| 						    &ipc_mux->ul_adb,
 | |
| 						    dg_n);
 | |
| 	}
 | |
| 
 | |
| 	ipc_mux->adb_prep_ongoing = false;
 | |
| 	return updated == 1;
 | |
| }
 | |
| 
 | |
| /* Calculates the Payload from any given ADB. */
 | |
| static int ipc_mux_get_payload_from_adb(struct iosm_mux *ipc_mux,
 | |
| 					struct mux_adbh *p_adbh)
 | |
| {
 | |
| 	struct mux_adth_dg *dg;
 | |
| 	struct mux_adth *adth;
 | |
| 	u32 payload_size = 0;
 | |
| 	u32 next_table_idx;
 | |
| 	int nr_of_dg, i;
 | |
| 
 | |
| 	/* Process the aggregated datagram tables. */
 | |
| 	next_table_idx = le32_to_cpu(p_adbh->first_table_index);
 | |
| 
 | |
| 	if (next_table_idx < sizeof(struct mux_adbh)) {
 | |
| 		dev_err(ipc_mux->dev, "unexpected empty ADB");
 | |
| 		return payload_size;
 | |
| 	}
 | |
| 
 | |
| 	while (next_table_idx != 0) {
 | |
| 		/* Get the reference to the table header. */
 | |
| 		adth = (struct mux_adth *)((u8 *)p_adbh + next_table_idx);
 | |
| 
 | |
| 		if (adth->signature == cpu_to_le32(IOSM_AGGR_MUX_SIG_ADTH)) {
 | |
| 			nr_of_dg = (le16_to_cpu(adth->table_length) -
 | |
| 					sizeof(struct mux_adth) +
 | |
| 					sizeof(struct mux_adth_dg)) /
 | |
| 					sizeof(struct mux_adth_dg);
 | |
| 
 | |
| 			if (nr_of_dg <= 0)
 | |
| 				return payload_size;
 | |
| 
 | |
| 			dg = &adth->dg;
 | |
| 
 | |
| 			for (i = 0; i < nr_of_dg; i++, dg++) {
 | |
| 				if (le32_to_cpu(dg->datagram_index) <
 | |
| 					sizeof(struct mux_adbh)) {
 | |
| 					return payload_size;
 | |
| 				}
 | |
| 				payload_size +=
 | |
| 					le16_to_cpu(dg->datagram_length);
 | |
| 			}
 | |
| 		}
 | |
| 		next_table_idx = le32_to_cpu(adth->next_table_index);
 | |
| 	}
 | |
| 
 | |
| 	return payload_size;
 | |
| }
 | |
| 
 | |
| void ipc_mux_ul_encoded_process(struct iosm_mux *ipc_mux, struct sk_buff *skb)
 | |
| {
 | |
| 	union mux_type_header hr;
 | |
| 	u16 adgh_len;
 | |
| 	int payload;
 | |
| 
 | |
| 	if (ipc_mux->protocol == MUX_LITE) {
 | |
| 		hr.adgh = (struct mux_adgh *)skb->data;
 | |
| 		adgh_len = le16_to_cpu(hr.adgh->length);
 | |
| 		if (hr.adgh->signature == cpu_to_le32(IOSM_AGGR_MUX_SIG_ADGH) &&
 | |
| 		    ipc_mux->ul_flow == MUX_UL)
 | |
| 			ipc_mux->ul_data_pend_bytes =
 | |
| 					ipc_mux->ul_data_pend_bytes - adgh_len;
 | |
| 	} else {
 | |
| 		hr.adbh = (struct mux_adbh *)(skb->data);
 | |
| 		payload = ipc_mux_get_payload_from_adb(ipc_mux, hr.adbh);
 | |
| 		ipc_mux->ul_data_pend_bytes -= payload;
 | |
| 	}
 | |
| 
 | |
| 	if (ipc_mux->ul_flow == MUX_UL)
 | |
| 		dev_dbg(ipc_mux->dev, "ul_data_pend_bytes: %lld",
 | |
| 			ipc_mux->ul_data_pend_bytes);
 | |
| 
 | |
| 	/* Reset the skb settings. */
 | |
| 	skb_trim(skb, 0);
 | |
| 
 | |
| 	/* Add the consumed ADB to the free list. */
 | |
| 	skb_queue_tail((&ipc_mux->ul_adb.free_list), skb);
 | |
| }
 | |
| 
 | |
| /* Start the NETIF uplink send transfer in MUX mode. */
 | |
| static int ipc_mux_tq_ul_trigger_encode(struct iosm_imem *ipc_imem, int arg,
 | |
| 					void *msg, size_t size)
 | |
| {
 | |
| 	struct iosm_mux *ipc_mux = ipc_imem->mux;
 | |
| 	bool ul_data_pend = false;
 | |
| 
 | |
| 	/* Add session UL data to a ADB and ADGH */
 | |
| 	ul_data_pend = ipc_mux_ul_data_encode(ipc_mux);
 | |
| 	if (ul_data_pend) {
 | |
| 		if (ipc_mux->protocol == MUX_AGGREGATION)
 | |
| 			ipc_imem_adb_timer_start(ipc_mux->imem);
 | |
| 
 | |
| 		/* Delay the doorbell irq */
 | |
| 		ipc_imem_td_update_timer_start(ipc_mux->imem);
 | |
| 	}
 | |
| 	/* reset the debounce flag */
 | |
| 	ipc_mux->ev_mux_net_transmit_pending = false;
 | |
| 
 | |
| 	return 0;
 | |
| }
 | |
| 
 | |
| int ipc_mux_ul_trigger_encode(struct iosm_mux *ipc_mux, int if_id,
 | |
| 			      struct sk_buff *skb)
 | |
| {
 | |
| 	struct mux_session *session = &ipc_mux->session[if_id];
 | |
| 	int ret = -EINVAL;
 | |
| 
 | |
| 	if (ipc_mux->channel &&
 | |
| 	    ipc_mux->channel->state != IMEM_CHANNEL_ACTIVE) {
 | |
| 		dev_err(ipc_mux->dev,
 | |
| 			"channel state is not IMEM_CHANNEL_ACTIVE");
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	if (!session->wwan) {
 | |
| 		dev_err(ipc_mux->dev, "session net ID is NULL");
 | |
| 		ret = -EFAULT;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	/* Session is under flow control.
 | |
| 	 * Check if packet can be queued in session list, if not
 | |
| 	 * suspend net tx
 | |
| 	 */
 | |
| 	if (skb_queue_len(&session->ul_list) >=
 | |
| 	    (session->net_tx_stop ?
 | |
| 		     IPC_MEM_MUX_UL_SESS_FCON_THRESHOLD :
 | |
| 		     (IPC_MEM_MUX_UL_SESS_FCON_THRESHOLD *
 | |
| 		      IPC_MEM_MUX_UL_SESS_FCOFF_THRESHOLD_FACTOR))) {
 | |
| 		ipc_mux_netif_tx_flowctrl(session, session->if_id, true);
 | |
| 		ret = -EBUSY;
 | |
| 		goto out;
 | |
| 	}
 | |
| 
 | |
| 	/* Add skb to the uplink skb accumulator. */
 | |
| 	skb_queue_tail(&session->ul_list, skb);
 | |
| 
 | |
| 	/* Inform the IPC kthread to pass uplink IP packets to CP. */
 | |
| 	if (!ipc_mux->ev_mux_net_transmit_pending) {
 | |
| 		ipc_mux->ev_mux_net_transmit_pending = true;
 | |
| 		ret = ipc_task_queue_send_task(ipc_mux->imem,
 | |
| 					       ipc_mux_tq_ul_trigger_encode, 0,
 | |
| 					       NULL, 0, false);
 | |
| 		if (ret)
 | |
| 			goto out;
 | |
| 	}
 | |
| 	dev_dbg(ipc_mux->dev, "mux ul if[%d] qlen=%d/%u, len=%d/%d, prio=%d",
 | |
| 		if_id, skb_queue_len(&session->ul_list), session->ul_list.qlen,
 | |
| 		skb->len, skb->truesize, skb->priority);
 | |
| 	ret = 0;
 | |
| out:
 | |
| 	return ret;
 | |
| }
 |