| // SPDX-License-Identifier: GPL-2.0-only |
| /* |
| * Copyright (C) 2020-21 Intel Corporation. |
| */ |
| |
| #include <linux/nospec.h> |
| |
| #include "iosm_ipc_imem_ops.h" |
| #include "iosm_ipc_mux_codec.h" |
| #include "iosm_ipc_task_queue.h" |
| |
| /* Test the link power state and send a MUX command in blocking mode. */ |
| static int ipc_mux_tq_cmd_send(struct iosm_imem *ipc_imem, int arg, void *msg, |
| size_t size) |
| { |
| struct iosm_mux *ipc_mux = ipc_imem->mux; |
| const struct mux_acb *acb = msg; |
| |
| skb_queue_tail(&ipc_mux->channel->ul_list, acb->skb); |
| ipc_imem_ul_send(ipc_mux->imem); |
| |
| return 0; |
| } |
| |
| static int ipc_mux_acb_send(struct iosm_mux *ipc_mux, bool blocking) |
| { |
| struct completion *completion = &ipc_mux->channel->ul_sem; |
| int ret = ipc_task_queue_send_task(ipc_mux->imem, ipc_mux_tq_cmd_send, |
| 0, &ipc_mux->acb, |
| sizeof(ipc_mux->acb), false); |
| if (ret) { |
| dev_err(ipc_mux->dev, "unable to send mux command"); |
| return ret; |
| } |
| |
| /* if blocking, suspend the app and wait for irq in the flash or |
| * crash phase. return false on timeout to indicate failure. |
| */ |
| if (blocking) { |
| u32 wait_time_milliseconds = IPC_MUX_CMD_RUN_DEFAULT_TIMEOUT; |
| |
| reinit_completion(completion); |
| |
| if (wait_for_completion_interruptible_timeout |
| (completion, msecs_to_jiffies(wait_time_milliseconds)) == |
| 0) { |
| dev_err(ipc_mux->dev, "ch[%d] timeout", |
| ipc_mux->channel_id); |
| ipc_uevent_send(ipc_mux->imem->dev, UEVENT_MDM_TIMEOUT); |
| return -ETIMEDOUT; |
| } |
| } |
| |
| return 0; |
| } |
| |
| /* Initialize the command header. */ |
| static void ipc_mux_acb_init(struct iosm_mux *ipc_mux) |
| { |
| struct mux_acb *acb = &ipc_mux->acb; |
| struct mux_acbh *header; |
| |
| header = (struct mux_acbh *)(acb->skb)->data; |
| header->block_length = cpu_to_le32(sizeof(struct mux_acbh)); |
| header->first_cmd_index = header->block_length; |
| header->signature = cpu_to_le32(IOSM_AGGR_MUX_SIG_ACBH); |
| header->sequence_nr = cpu_to_le16(ipc_mux->acb_tx_sequence_nr++); |
| } |
| |
| /* Add a command to the ACB. */ |
| static struct mux_cmdh *ipc_mux_acb_add_cmd(struct iosm_mux *ipc_mux, u32 cmd, |
| void *param, u32 param_size) |
| { |
| struct mux_acbh *header; |
| struct mux_cmdh *cmdh; |
| struct mux_acb *acb; |
| |
| acb = &ipc_mux->acb; |
| header = (struct mux_acbh *)(acb->skb)->data; |
| cmdh = (struct mux_cmdh *) |
| ((acb->skb)->data + le32_to_cpu(header->block_length)); |
| |
| cmdh->signature = cpu_to_le32(MUX_SIG_CMDH); |
| cmdh->command_type = cpu_to_le32(cmd); |
| cmdh->if_id = acb->if_id; |
| |
| acb->cmd = cmd; |
| cmdh->cmd_len = cpu_to_le16(offsetof(struct mux_cmdh, param) + |
| param_size); |
| cmdh->transaction_id = cpu_to_le32(ipc_mux->tx_transaction_id++); |
| if (param) |
| memcpy(&cmdh->param, param, param_size); |
| |
| skb_put(acb->skb, le32_to_cpu(header->block_length) + |
| le16_to_cpu(cmdh->cmd_len)); |
| |
| return cmdh; |
| } |
| |
| /* Prepare mux Command */ |
| static struct mux_lite_cmdh *ipc_mux_lite_add_cmd(struct iosm_mux *ipc_mux, |
| u32 cmd, struct mux_acb *acb, |
| void *param, u32 param_size) |
| { |
| struct mux_lite_cmdh *cmdh = (struct mux_lite_cmdh *)acb->skb->data; |
| |
| cmdh->signature = cpu_to_le32(MUX_SIG_CMDH); |
| cmdh->command_type = cpu_to_le32(cmd); |
| cmdh->if_id = acb->if_id; |
| |
| acb->cmd = cmd; |
| |
| cmdh->cmd_len = cpu_to_le16(offsetof(struct mux_lite_cmdh, param) + |
| param_size); |
| cmdh->transaction_id = cpu_to_le32(ipc_mux->tx_transaction_id++); |
| |
| if (param) |
| memcpy(&cmdh->param, param, param_size); |
| |
| skb_put(acb->skb, le16_to_cpu(cmdh->cmd_len)); |
| |
| return cmdh; |
| } |
| |
| static int ipc_mux_acb_alloc(struct iosm_mux *ipc_mux) |
| { |
| struct mux_acb *acb = &ipc_mux->acb; |
| struct sk_buff *skb; |
| dma_addr_t mapping; |
| |
| /* Allocate skb memory for the uplink buffer. */ |
| skb = ipc_pcie_alloc_skb(ipc_mux->pcie, MUX_MAX_UL_ACB_BUF_SIZE, |
| GFP_ATOMIC, &mapping, DMA_TO_DEVICE, 0); |
| if (!skb) |
| return -ENOMEM; |
| |
| /* Save the skb address. */ |
| acb->skb = skb; |
| |
| memset(skb->data, 0, MUX_MAX_UL_ACB_BUF_SIZE); |
| |
| return 0; |
| } |
| |
| int ipc_mux_dl_acb_send_cmds(struct iosm_mux *ipc_mux, u32 cmd_type, u8 if_id, |
| u32 transaction_id, union mux_cmd_param *param, |
| size_t res_size, bool blocking, bool respond) |
| { |
| struct mux_acb *acb = &ipc_mux->acb; |
| union mux_type_cmdh cmdh; |
| int ret = 0; |
| |
| acb->if_id = if_id; |
| ret = ipc_mux_acb_alloc(ipc_mux); |
| if (ret) |
| return ret; |
| |
| if (ipc_mux->protocol == MUX_LITE) { |
| cmdh.ack_lite = ipc_mux_lite_add_cmd(ipc_mux, cmd_type, acb, |
| param, res_size); |
| |
| if (respond) |
| cmdh.ack_lite->transaction_id = |
| cpu_to_le32(transaction_id); |
| } else { |
| /* Initialize the ACB header. */ |
| ipc_mux_acb_init(ipc_mux); |
| cmdh.ack_aggr = ipc_mux_acb_add_cmd(ipc_mux, cmd_type, param, |
| res_size); |
| |
| if (respond) |
| cmdh.ack_aggr->transaction_id = |
| cpu_to_le32(transaction_id); |
| } |
| ret = ipc_mux_acb_send(ipc_mux, blocking); |
| |
| return ret; |
| } |
| |
| void ipc_mux_netif_tx_flowctrl(struct mux_session *session, int idx, bool on) |
| { |
| /* Inform the network interface to start/stop flow ctrl */ |
| ipc_wwan_tx_flowctrl(session->wwan, idx, on); |
| } |
| |
| static int ipc_mux_dl_cmdresps_decode_process(struct iosm_mux *ipc_mux, |
| union mux_cmd_param param, |
| __le32 command_type, u8 if_id, |
| __le32 transaction_id) |
| { |
| struct mux_acb *acb = &ipc_mux->acb; |
| |
| switch (le32_to_cpu(command_type)) { |
| case MUX_CMD_OPEN_SESSION_RESP: |
| case MUX_CMD_CLOSE_SESSION_RESP: |
| /* Resume the control application. */ |
| acb->got_param = param; |
| break; |
| |
| case MUX_LITE_CMD_FLOW_CTL_ACK: |
| /* This command type is not expected as response for |
| * Aggregation version of the protocol. So return non-zero. |
| */ |
| if (ipc_mux->protocol != MUX_LITE) |
| return -EINVAL; |
| |
| dev_dbg(ipc_mux->dev, "if_id %u FLOW_CTL_ACK %u received", |
| if_id, le32_to_cpu(transaction_id)); |
| break; |
| |
| case IOSM_AGGR_MUX_CMD_FLOW_CTL_ACK: |
| /* This command type is not expected as response for |
| * Lite version of the protocol. So return non-zero. |
| */ |
| if (ipc_mux->protocol == MUX_LITE) |
| return -EINVAL; |
| break; |
| |
| default: |
| return -EINVAL; |
| } |
| |
| acb->wanted_response = MUX_CMD_INVALID; |
| acb->got_response = le32_to_cpu(command_type); |
| complete(&ipc_mux->channel->ul_sem); |
| |
| return 0; |
| } |
| |
| static int ipc_mux_dl_cmds_decode_process(struct iosm_mux *ipc_mux, |
| union mux_cmd_param *param, |
| __le32 command_type, u8 if_id, |
| __le16 cmd_len, int size) |
| { |
| struct mux_session *session; |
| struct hrtimer *adb_timer; |
| |
| dev_dbg(ipc_mux->dev, "if_id[%d]: dlcmds decode process %d", |
| if_id, le32_to_cpu(command_type)); |
| |
| switch (le32_to_cpu(command_type)) { |
| case MUX_LITE_CMD_FLOW_CTL: |
| case IOSM_AGGR_MUX_CMD_FLOW_CTL_DISABLE: |
| |
| if (if_id >= IPC_MEM_MUX_IP_SESSION_ENTRIES) { |
| dev_err(ipc_mux->dev, "if_id [%d] not valid", |
| if_id); |
| return -EINVAL; /* No session interface id. */ |
| } |
| |
| session = &ipc_mux->session[if_id]; |
| adb_timer = &ipc_mux->imem->adb_timer; |
| |
| if (param->flow_ctl.mask == cpu_to_le32(0xFFFFFFFF)) { |
| /* Backward Compatibility */ |
| if (cmd_len == cpu_to_le16(size)) |
| session->flow_ctl_mask = |
| le32_to_cpu(param->flow_ctl.mask); |
| else |
| session->flow_ctl_mask = ~0; |
| /* if CP asks for FLOW CTRL Enable |
| * then set our internal flow control Tx flag |
| * to limit uplink session queueing |
| */ |
| session->net_tx_stop = true; |
| |
| /* We have to call Finish ADB here. |
| * Otherwise any already queued data |
| * will be sent to CP when ADB is full |
| * for some other sessions. |
| */ |
| if (ipc_mux->protocol == MUX_AGGREGATION) { |
| ipc_mux_ul_adb_finish(ipc_mux); |
| ipc_imem_hrtimer_stop(adb_timer); |
| } |
| /* Update the stats */ |
| session->flow_ctl_en_cnt++; |
| } else if (param->flow_ctl.mask == 0) { |
| /* Just reset the Flow control mask and let |
| * mux_flow_ctrl_low_thre_b take control on |
| * our internal Tx flag and enabling kernel |
| * flow control |
| */ |
| dev_dbg(ipc_mux->dev, "if_id[%u] flow_ctl mask 0x%08X", |
| if_id, le32_to_cpu(param->flow_ctl.mask)); |
| /* Backward Compatibility */ |
| if (cmd_len == cpu_to_le16(size)) |
| session->flow_ctl_mask = |
| le32_to_cpu(param->flow_ctl.mask); |
| else |
| session->flow_ctl_mask = 0; |
| /* Update the stats */ |
| session->flow_ctl_dis_cnt++; |
| } else { |
| break; |
| } |
| |
| ipc_mux->acc_adb_size = 0; |
| ipc_mux->acc_payload_size = 0; |
| |
| dev_dbg(ipc_mux->dev, "if_id[%u] FLOW CTRL 0x%08X", if_id, |
| le32_to_cpu(param->flow_ctl.mask)); |
| break; |
| |
| case MUX_LITE_CMD_LINK_STATUS_REPORT: |
| break; |
| |
| default: |
| return -EINVAL; |
| } |
| return 0; |
| } |
| |
| /* Decode and Send appropriate response to a command block. */ |
| static void ipc_mux_dl_cmd_decode(struct iosm_mux *ipc_mux, struct sk_buff *skb) |
| { |
| struct mux_lite_cmdh *cmdh = (struct mux_lite_cmdh *)skb->data; |
| __le32 trans_id = cmdh->transaction_id; |
| int size; |
| |
| if (ipc_mux_dl_cmdresps_decode_process(ipc_mux, cmdh->param, |
| cmdh->command_type, cmdh->if_id, |
| cmdh->transaction_id)) { |
| /* Unable to decode command response indicates the cmd_type |
| * may be a command instead of response. So try to decoding it. |
| */ |
| size = offsetof(struct mux_lite_cmdh, param) + |
| sizeof(cmdh->param.flow_ctl); |
| if (!ipc_mux_dl_cmds_decode_process(ipc_mux, &cmdh->param, |
| cmdh->command_type, |
| cmdh->if_id, |
| cmdh->cmd_len, size)) { |
| /* Decoded command may need a response. Give the |
| * response according to the command type. |
| */ |
| union mux_cmd_param *mux_cmd = NULL; |
| size_t size = 0; |
| u32 cmd = MUX_LITE_CMD_LINK_STATUS_REPORT_RESP; |
| |
| if (cmdh->command_type == |
| cpu_to_le32(MUX_LITE_CMD_LINK_STATUS_REPORT)) { |
| mux_cmd = &cmdh->param; |
| mux_cmd->link_status_resp.response = |
| cpu_to_le32(MUX_CMD_RESP_SUCCESS); |
| /* response field is u32 */ |
| size = sizeof(u32); |
| } else if (cmdh->command_type == |
| cpu_to_le32(MUX_LITE_CMD_FLOW_CTL)) { |
| cmd = MUX_LITE_CMD_FLOW_CTL_ACK; |
| } else { |
| return; |
| } |
| |
| if (ipc_mux_dl_acb_send_cmds(ipc_mux, cmd, cmdh->if_id, |
| le32_to_cpu(trans_id), |
| mux_cmd, size, false, |
| true)) |
| dev_err(ipc_mux->dev, |
| "if_id %d: cmd send failed", |
| cmdh->if_id); |
| } |
| } |
| } |
| |
| /* Pass the DL packet to the netif layer. */ |
| static int ipc_mux_net_receive(struct iosm_mux *ipc_mux, int if_id, |
| struct iosm_wwan *wwan, u32 offset, |
| u8 service_class, struct sk_buff *skb) |
| { |
| struct sk_buff *dest_skb = skb_clone(skb, GFP_ATOMIC); |
| |
| if (!dest_skb) |
| return -ENOMEM; |
| |
| skb_pull(dest_skb, offset); |
| skb_set_tail_pointer(dest_skb, dest_skb->len); |
| /* Pass the packet to the netif layer. */ |
| dest_skb->priority = service_class; |
| |
| return ipc_wwan_receive(wwan, dest_skb, false, if_id); |
| } |
| |
| /* Decode Flow Credit Table in the block */ |
| static void ipc_mux_dl_fcth_decode(struct iosm_mux *ipc_mux, |
| unsigned char *block) |
| { |
| struct ipc_mem_lite_gen_tbl *fct = (struct ipc_mem_lite_gen_tbl *)block; |
| struct iosm_wwan *wwan; |
| int ul_credits; |
| int if_id; |
| |
| if (fct->vfl_length != sizeof(fct->vfl.nr_of_bytes)) { |
| dev_err(ipc_mux->dev, "unexpected FCT length: %d", |
| fct->vfl_length); |
| return; |
| } |
| |
| if_id = fct->if_id; |
| if (if_id >= IPC_MEM_MUX_IP_SESSION_ENTRIES) { |
| dev_err(ipc_mux->dev, "not supported if_id: %d", if_id); |
| return; |
| } |
| |
| /* Is the session active ? */ |
| if_id = array_index_nospec(if_id, IPC_MEM_MUX_IP_SESSION_ENTRIES); |
| wwan = ipc_mux->session[if_id].wwan; |
| if (!wwan) { |
| dev_err(ipc_mux->dev, "session Net ID is NULL"); |
| return; |
| } |
| |
| ul_credits = le32_to_cpu(fct->vfl.nr_of_bytes); |
| |
| dev_dbg(ipc_mux->dev, "Flow_Credit:: if_id[%d] Old: %d Grants: %d", |
| if_id, ipc_mux->session[if_id].ul_flow_credits, ul_credits); |
| |
| /* Update the Flow Credit information from ADB */ |
| ipc_mux->session[if_id].ul_flow_credits += ul_credits; |
| |
| /* Check whether the TX can be started */ |
| if (ipc_mux->session[if_id].ul_flow_credits > 0) { |
| ipc_mux->session[if_id].net_tx_stop = false; |
| ipc_mux_netif_tx_flowctrl(&ipc_mux->session[if_id], |
| ipc_mux->session[if_id].if_id, false); |
| } |
| } |
| |
| /* Decode non-aggregated datagram */ |
| static void ipc_mux_dl_adgh_decode(struct iosm_mux *ipc_mux, |
| struct sk_buff *skb) |
| { |
| u32 pad_len, packet_offset; |
| struct iosm_wwan *wwan; |
| struct mux_adgh *adgh; |
| u8 *block = skb->data; |
| int rc = 0; |
| u8 if_id; |
| |
| adgh = (struct mux_adgh *)block; |
| |
| if (adgh->signature != cpu_to_le32(IOSM_AGGR_MUX_SIG_ADGH)) { |
| dev_err(ipc_mux->dev, "invalid ADGH signature received"); |
| return; |
| } |
| |
| if_id = adgh->if_id; |
| if (if_id >= IPC_MEM_MUX_IP_SESSION_ENTRIES) { |
| dev_err(ipc_mux->dev, "invalid if_id while decoding %d", if_id); |
| return; |
| } |
| |
| /* Is the session active ? */ |
| if_id = array_index_nospec(if_id, IPC_MEM_MUX_IP_SESSION_ENTRIES); |
| wwan = ipc_mux->session[if_id].wwan; |
| if (!wwan) { |
| dev_err(ipc_mux->dev, "session Net ID is NULL"); |
| return; |
| } |
| |
| /* Store the pad len for the corresponding session |
| * Pad bytes as negotiated in the open session less the header size |
| * (see session management chapter for details). |
| * If resulting padding is zero or less, the additional head padding is |
| * omitted. For e.g., if HEAD_PAD_LEN = 16 or less, this field is |
| * omitted if HEAD_PAD_LEN = 20, then this field will have 4 bytes |
| * set to zero |
| */ |
| pad_len = |
| ipc_mux->session[if_id].dl_head_pad_len - IPC_MEM_DL_ETH_OFFSET; |
| packet_offset = sizeof(*adgh) + pad_len; |
| |
| if_id += ipc_mux->wwan_q_offset; |
| |
| /* Pass the packet to the netif layer */ |
| rc = ipc_mux_net_receive(ipc_mux, if_id, wwan, packet_offset, |
| adgh->service_class, skb); |
| if (rc) { |
| dev_err(ipc_mux->dev, "mux adgh decoding error"); |
| return; |
| } |
| ipc_mux->session[if_id].flush = 1; |
| } |
| |
| static void ipc_mux_dl_acbcmd_decode(struct iosm_mux *ipc_mux, |
| struct mux_cmdh *cmdh, int size) |
| { |
| u32 link_st = IOSM_AGGR_MUX_CMD_LINK_STATUS_REPORT_RESP; |
| u32 fctl_dis = IOSM_AGGR_MUX_CMD_FLOW_CTL_DISABLE; |
| u32 fctl_ena = IOSM_AGGR_MUX_CMD_FLOW_CTL_ENABLE; |
| u32 fctl_ack = IOSM_AGGR_MUX_CMD_FLOW_CTL_ACK; |
| union mux_cmd_param *cmd_p = NULL; |
| u32 cmd = link_st; |
| u32 trans_id; |
| |
| if (!ipc_mux_dl_cmds_decode_process(ipc_mux, &cmdh->param, |
| cmdh->command_type, cmdh->if_id, |
| cmdh->cmd_len, size)) { |
| size = 0; |
| if (cmdh->command_type == cpu_to_le32(link_st)) { |
| cmd_p = &cmdh->param; |
| cmd_p->link_status_resp.response = MUX_CMD_RESP_SUCCESS; |
| } else if ((cmdh->command_type == cpu_to_le32(fctl_ena)) || |
| (cmdh->command_type == cpu_to_le32(fctl_dis))) { |
| cmd = fctl_ack; |
| } else { |
| return; |
| } |
| trans_id = le32_to_cpu(cmdh->transaction_id); |
| ipc_mux_dl_acb_send_cmds(ipc_mux, cmd, cmdh->if_id, |
| trans_id, cmd_p, size, false, true); |
| } |
| } |
| |
| /* Decode an aggregated command block. */ |
| static void ipc_mux_dl_acb_decode(struct iosm_mux *ipc_mux, struct sk_buff *skb) |
| { |
| struct mux_acbh *acbh; |
| struct mux_cmdh *cmdh; |
| u32 next_cmd_index; |
| u8 *block; |
| int size; |
| |
| acbh = (struct mux_acbh *)(skb->data); |
| block = (u8 *)(skb->data); |
| |
| next_cmd_index = le32_to_cpu(acbh->first_cmd_index); |
| next_cmd_index = array_index_nospec(next_cmd_index, |
| sizeof(struct mux_cmdh)); |
| |
| while (next_cmd_index != 0) { |
| cmdh = (struct mux_cmdh *)&block[next_cmd_index]; |
| next_cmd_index = le32_to_cpu(cmdh->next_cmd_index); |
| if (ipc_mux_dl_cmdresps_decode_process(ipc_mux, cmdh->param, |
| cmdh->command_type, |
| cmdh->if_id, |
| cmdh->transaction_id)) { |
| size = offsetof(struct mux_cmdh, param) + |
| sizeof(cmdh->param.flow_ctl); |
| ipc_mux_dl_acbcmd_decode(ipc_mux, cmdh, size); |
| } |
| } |
| } |
| |
| /* process datagram */ |
| static int mux_dl_process_dg(struct iosm_mux *ipc_mux, struct mux_adbh *adbh, |
| struct mux_adth_dg *dg, struct sk_buff *skb, |
| int if_id, int nr_of_dg) |
| { |
| u32 dl_head_pad_len = ipc_mux->session[if_id].dl_head_pad_len; |
| u32 packet_offset, i, rc; |
| |
| for (i = 0; i < nr_of_dg; i++, dg++) { |
| if (le32_to_cpu(dg->datagram_index) |
| < sizeof(struct mux_adbh)) |
| goto dg_error; |
| |
| /* Is the packet inside of the ADB */ |
| if (le32_to_cpu(dg->datagram_index) >= |
| le32_to_cpu(adbh->block_length)) { |
| goto dg_error; |
| } else { |
| packet_offset = |
| le32_to_cpu(dg->datagram_index) + |
| dl_head_pad_len; |
| /* Pass the packet to the netif layer. */ |
| rc = ipc_mux_net_receive(ipc_mux, if_id, ipc_mux->wwan, |
| packet_offset, |
| dg->service_class, |
| skb); |
| if (rc) |
| goto dg_error; |
| } |
| } |
| return 0; |
| dg_error: |
| return -1; |
| } |
| |
| /* Decode an aggregated data block. */ |
| static void mux_dl_adb_decode(struct iosm_mux *ipc_mux, |
| struct sk_buff *skb) |
| { |
| struct mux_adth_dg *dg; |
| struct iosm_wwan *wwan; |
| struct mux_adbh *adbh; |
| struct mux_adth *adth; |
| int nr_of_dg, if_id; |
| u32 adth_index; |
| u8 *block; |
| |
| block = skb->data; |
| adbh = (struct mux_adbh *)block; |
| |
| /* Process the aggregated datagram tables. */ |
| adth_index = le32_to_cpu(adbh->first_table_index); |
| |
| /* Has CP sent an empty ADB ? */ |
| if (adth_index < 1) { |
| dev_err(ipc_mux->dev, "unexpected empty ADB"); |
| goto adb_decode_err; |
| } |
| |
| /* Loop through mixed session tables. */ |
| while (adth_index) { |
| /* Get the reference to the table header. */ |
| adth = (struct mux_adth *)(block + adth_index); |
| |
| /* Get the interface id and map it to the netif id. */ |
| if_id = adth->if_id; |
| if (if_id >= IPC_MEM_MUX_IP_SESSION_ENTRIES) |
| goto adb_decode_err; |
| |
| if_id = array_index_nospec(if_id, |
| IPC_MEM_MUX_IP_SESSION_ENTRIES); |
| |
| /* Is the session active ? */ |
| wwan = ipc_mux->session[if_id].wwan; |
| if (!wwan) |
| goto adb_decode_err; |
| |
| /* Consistency checks for aggregated datagram table. */ |
| if (adth->signature != cpu_to_le32(IOSM_AGGR_MUX_SIG_ADTH)) |
| goto adb_decode_err; |
| |
| if (le16_to_cpu(adth->table_length) < (sizeof(struct mux_adth) - |
| sizeof(struct mux_adth_dg))) |
| goto adb_decode_err; |
| |
| /* Calculate the number of datagrams. */ |
| nr_of_dg = (le16_to_cpu(adth->table_length) - |
| sizeof(struct mux_adth) + |
| sizeof(struct mux_adth_dg)) / |
| sizeof(struct mux_adth_dg); |
| |
| /* Is the datagram table empty ? */ |
| if (nr_of_dg < 1) { |
| dev_err(ipc_mux->dev, |
| "adthidx=%u,nr_of_dg=%d,next_tblidx=%u", |
| adth_index, nr_of_dg, |
| le32_to_cpu(adth->next_table_index)); |
| |
| /* Move to the next aggregated datagram table. */ |
| adth_index = le32_to_cpu(adth->next_table_index); |
| continue; |
| } |
| |
| /* New aggregated datagram table. */ |
| dg = &adth->dg; |
| if (mux_dl_process_dg(ipc_mux, adbh, dg, skb, if_id, |
| nr_of_dg) < 0) |
| goto adb_decode_err; |
| |
| /* mark session for final flush */ |
| ipc_mux->session[if_id].flush = 1; |
| |
| /* Move to the next aggregated datagram table. */ |
| adth_index = le32_to_cpu(adth->next_table_index); |
| } |
| |
| adb_decode_err: |
| return; |
| } |
| |
| /** |
| * ipc_mux_dl_decode - Route the DL packet through the IP MUX layer |
| * depending on Header. |
| * @ipc_mux: Pointer to MUX data-struct |
| * @skb: Pointer to ipc_skb. |
| */ |
| void ipc_mux_dl_decode(struct iosm_mux *ipc_mux, struct sk_buff *skb) |
| { |
| u32 signature; |
| |
| if (!skb->data) |
| return; |
| |
| /* Decode the MUX header type. */ |
| signature = le32_to_cpup((__le32 *)skb->data); |
| |
| switch (signature) { |
| case IOSM_AGGR_MUX_SIG_ADBH: /* Aggregated Data Block Header */ |
| mux_dl_adb_decode(ipc_mux, skb); |
| break; |
| case IOSM_AGGR_MUX_SIG_ADGH: |
| ipc_mux_dl_adgh_decode(ipc_mux, skb); |
| break; |
| case MUX_SIG_FCTH: |
| ipc_mux_dl_fcth_decode(ipc_mux, skb->data); |
| break; |
| case IOSM_AGGR_MUX_SIG_ACBH: /* Aggregated Command Block Header */ |
| ipc_mux_dl_acb_decode(ipc_mux, skb); |
| break; |
| case MUX_SIG_CMDH: |
| ipc_mux_dl_cmd_decode(ipc_mux, skb); |
| break; |
| |
| default: |
| dev_err(ipc_mux->dev, "invalid ABH signature"); |
| } |
| |
| ipc_pcie_kfree_skb(ipc_mux->pcie, skb); |
| } |
| |
| static int ipc_mux_ul_skb_alloc(struct iosm_mux *ipc_mux, |
| struct mux_adb *ul_adb, u32 type) |
| { |
| /* Take the first element of the free list. */ |
| struct sk_buff *skb = skb_dequeue(&ul_adb->free_list); |
| u32 no_if = IPC_MEM_MUX_IP_SESSION_ENTRIES; |
| u32 *next_tb_id; |
| int qlt_size; |
| u32 if_id; |
| |
| if (!skb) |
| return -EBUSY; /* Wait for a free ADB skb. */ |
| |
| /* Mark it as UL ADB to select the right free operation. */ |
| IPC_CB(skb)->op_type = (u8)UL_MUX_OP_ADB; |
| |
| switch (type) { |
| case IOSM_AGGR_MUX_SIG_ADBH: |
| /* Save the ADB memory settings. */ |
| ul_adb->dest_skb = skb; |
| ul_adb->buf = skb->data; |
| ul_adb->size = IPC_MEM_MAX_ADB_BUF_SIZE; |
| |
| /* reset statistic counter */ |
| ul_adb->if_cnt = 0; |
| ul_adb->payload_size = 0; |
| ul_adb->dg_cnt_total = 0; |
| |
| /* Initialize the ADBH. */ |
| ul_adb->adbh = (struct mux_adbh *)ul_adb->buf; |
| memset(ul_adb->adbh, 0, sizeof(struct mux_adbh)); |
| ul_adb->adbh->signature = cpu_to_le32(IOSM_AGGR_MUX_SIG_ADBH); |
| ul_adb->adbh->block_length = |
| cpu_to_le32(sizeof(struct mux_adbh)); |
| next_tb_id = (unsigned int *)&ul_adb->adbh->first_table_index; |
| ul_adb->next_table_index = next_tb_id; |
| |
| /* Clear the local copy of DGs for new ADB */ |
| memset(ul_adb->dg, 0, sizeof(ul_adb->dg)); |
| |
| /* Clear the DG count and QLT updated status for new ADB */ |
| for (if_id = 0; if_id < no_if; if_id++) { |
| ul_adb->dg_count[if_id] = 0; |
| ul_adb->qlt_updated[if_id] = 0; |
| } |
| break; |
| |
| case IOSM_AGGR_MUX_SIG_ADGH: |
| /* Save the ADB memory settings. */ |
| ul_adb->dest_skb = skb; |
| ul_adb->buf = skb->data; |
| ul_adb->size = IPC_MEM_MAX_DL_MUX_LITE_BUF_SIZE; |
| /* reset statistic counter */ |
| ul_adb->if_cnt = 0; |
| ul_adb->payload_size = 0; |
| ul_adb->dg_cnt_total = 0; |
| |
| ul_adb->adgh = (struct mux_adgh *)skb->data; |
| memset(ul_adb->adgh, 0, sizeof(struct mux_adgh)); |
| break; |
| |
| case MUX_SIG_QLTH: |
| qlt_size = offsetof(struct ipc_mem_lite_gen_tbl, vfl) + |
| (MUX_QUEUE_LEVEL * sizeof(struct mux_lite_vfl)); |
| |
| if (qlt_size > IPC_MEM_MAX_DL_MUX_LITE_BUF_SIZE) { |
| dev_err(ipc_mux->dev, |
| "can't support. QLT size:%d SKB size: %d", |
| qlt_size, IPC_MEM_MAX_DL_MUX_LITE_BUF_SIZE); |
| return -ERANGE; |
| } |
| |
| ul_adb->qlth_skb = skb; |
| memset((ul_adb->qlth_skb)->data, 0, qlt_size); |
| skb_put(skb, qlt_size); |
| break; |
| } |
| |
| return 0; |
| } |
| |
| static void ipc_mux_ul_adgh_finish(struct iosm_mux *ipc_mux) |
| { |
| struct mux_adb *ul_adb = &ipc_mux->ul_adb; |
| u16 adgh_len; |
| long long bytes; |
| char *str; |
| |
| if (!ul_adb->dest_skb) { |
| dev_err(ipc_mux->dev, "no dest skb"); |
| return; |
| } |
| |
| adgh_len = le16_to_cpu(ul_adb->adgh->length); |
| skb_put(ul_adb->dest_skb, adgh_len); |
| skb_queue_tail(&ipc_mux->channel->ul_list, ul_adb->dest_skb); |
| ul_adb->dest_skb = NULL; |
| |
| if (ipc_mux->ul_flow == MUX_UL_ON_CREDITS) { |
| struct mux_session *session; |
| |
| session = &ipc_mux->session[ul_adb->adgh->if_id]; |
| str = "available_credits"; |
| bytes = (long long)session->ul_flow_credits; |
| |
| } else { |
| str = "pend_bytes"; |
| bytes = ipc_mux->ul_data_pend_bytes; |
| ipc_mux->ul_data_pend_bytes = ipc_mux->ul_data_pend_bytes + |
| adgh_len; |
| } |
| |
| dev_dbg(ipc_mux->dev, "UL ADGH: size=%u, if_id=%d, payload=%d, %s=%lld", |
| adgh_len, ul_adb->adgh->if_id, ul_adb->payload_size, |
| str, bytes); |
| } |
| |
| static void ipc_mux_ul_encode_adth(struct iosm_mux *ipc_mux, |
| struct mux_adb *ul_adb, int *out_offset) |
| { |
| int i, qlt_size, offset = *out_offset; |
| struct mux_qlth *p_adb_qlt; |
| struct mux_adth_dg *dg; |
| struct mux_adth *adth; |
| u16 adth_dg_size; |
| u32 *next_tb_id; |
| |
| qlt_size = offsetof(struct mux_qlth, ql) + |
| MUX_QUEUE_LEVEL * sizeof(struct mux_qlth_ql); |
| |
| for (i = 0; i < ipc_mux->nr_sessions; i++) { |
| if (ul_adb->dg_count[i] > 0) { |
| adth_dg_size = offsetof(struct mux_adth, dg) + |
| ul_adb->dg_count[i] * sizeof(*dg); |
| |
| *ul_adb->next_table_index = offset; |
| adth = (struct mux_adth *)&ul_adb->buf[offset]; |
| next_tb_id = (unsigned int *)&adth->next_table_index; |
| ul_adb->next_table_index = next_tb_id; |
| offset += adth_dg_size; |
| adth->signature = cpu_to_le32(IOSM_AGGR_MUX_SIG_ADTH); |
| adth->if_id = i; |
| adth->table_length = cpu_to_le16(adth_dg_size); |
| adth_dg_size -= offsetof(struct mux_adth, dg); |
| memcpy(&adth->dg, ul_adb->dg[i], adth_dg_size); |
| ul_adb->if_cnt++; |
| } |
| |
| if (ul_adb->qlt_updated[i]) { |
| *ul_adb->next_table_index = offset; |
| p_adb_qlt = (struct mux_qlth *)&ul_adb->buf[offset]; |
| ul_adb->next_table_index = |
| (u32 *)&p_adb_qlt->next_table_index; |
| memcpy(p_adb_qlt, ul_adb->pp_qlt[i], qlt_size); |
| offset += qlt_size; |
| } |
| } |
| *out_offset = offset; |
| } |
| |
| /** |
| * ipc_mux_ul_adb_finish - Add the TD of the aggregated session packets to TDR. |
| * @ipc_mux: Pointer to MUX data-struct. |
| */ |
| void ipc_mux_ul_adb_finish(struct iosm_mux *ipc_mux) |
| { |
| bool ul_data_pend = false; |
| struct mux_adb *ul_adb; |
| unsigned long flags; |
| int offset; |
| |
| ul_adb = &ipc_mux->ul_adb; |
| if (!ul_adb->dest_skb) |
| return; |
| |
| offset = *ul_adb->next_table_index; |
| ipc_mux_ul_encode_adth(ipc_mux, ul_adb, &offset); |
| ul_adb->adbh->block_length = cpu_to_le32(offset); |
| |
| if (le32_to_cpu(ul_adb->adbh->block_length) > ul_adb->size) { |
| ul_adb->dest_skb = NULL; |
| return; |
| } |
| |
| *ul_adb->next_table_index = 0; |
| ul_adb->adbh->sequence_nr = cpu_to_le16(ipc_mux->adb_tx_sequence_nr++); |
| skb_put(ul_adb->dest_skb, le32_to_cpu(ul_adb->adbh->block_length)); |
| |
| spin_lock_irqsave(&(&ipc_mux->channel->ul_list)->lock, flags); |
| __skb_queue_tail(&ipc_mux->channel->ul_list, ul_adb->dest_skb); |
| spin_unlock_irqrestore(&(&ipc_mux->channel->ul_list)->lock, flags); |
| |
| ul_adb->dest_skb = NULL; |
| /* Updates the TDs with ul_list */ |
| ul_data_pend = ipc_imem_ul_write_td(ipc_mux->imem); |
| |
| /* Delay the doorbell irq */ |
| if (ul_data_pend) |
| ipc_imem_td_update_timer_start(ipc_mux->imem); |
| |
| ipc_mux->acc_adb_size += le32_to_cpu(ul_adb->adbh->block_length); |
| ipc_mux->acc_payload_size += ul_adb->payload_size; |
| ipc_mux->ul_data_pend_bytes += ul_adb->payload_size; |
| } |
| |
| /* Allocates an ADB from the free list and initializes it with ADBH */ |
| static bool ipc_mux_ul_adb_allocate(struct iosm_mux *ipc_mux, |
| struct mux_adb *adb, int *size_needed, |
| u32 type) |
| { |
| bool ret_val = false; |
| int status; |
| |
| if (!adb->dest_skb) { |
| /* Allocate memory for the ADB including of the |
| * datagram table header. |
| */ |
| status = ipc_mux_ul_skb_alloc(ipc_mux, adb, type); |
| if (status) |
| /* Is a pending ADB available ? */ |
| ret_val = true; /* None. */ |
| |
| /* Update size need to zero only for new ADB memory */ |
| *size_needed = 0; |
| } |
| |
| return ret_val; |
| } |
| |
| /* Informs the network stack to stop sending further packets for all opened |
| * sessions |
| */ |
| static void ipc_mux_stop_tx_for_all_sessions(struct iosm_mux *ipc_mux) |
| { |
| struct mux_session *session; |
| int idx; |
| |
| for (idx = 0; idx < IPC_MEM_MUX_IP_SESSION_ENTRIES; idx++) { |
| session = &ipc_mux->session[idx]; |
| |
| if (!session->wwan) |
| continue; |
| |
| session->net_tx_stop = true; |
| } |
| } |
| |
| /* Sends Queue Level Table of all opened sessions */ |
| static bool ipc_mux_lite_send_qlt(struct iosm_mux *ipc_mux) |
| { |
| struct ipc_mem_lite_gen_tbl *qlt; |
| struct mux_session *session; |
| bool qlt_updated = false; |
| int i; |
| int qlt_size; |
| |
| if (!ipc_mux->initialized || ipc_mux->state != MUX_S_ACTIVE) |
| return qlt_updated; |
| |
| qlt_size = offsetof(struct ipc_mem_lite_gen_tbl, vfl) + |
| MUX_QUEUE_LEVEL * sizeof(struct mux_lite_vfl); |
| |
| for (i = 0; i < IPC_MEM_MUX_IP_SESSION_ENTRIES; i++) { |
| session = &ipc_mux->session[i]; |
| |
| if (!session->wwan || session->flow_ctl_mask) |
| continue; |
| |
| if (ipc_mux_ul_skb_alloc(ipc_mux, &ipc_mux->ul_adb, |
| MUX_SIG_QLTH)) { |
| dev_err(ipc_mux->dev, |
| "no reserved mem to send QLT of if_id: %d", i); |
| break; |
| } |
| |
| /* Prepare QLT */ |
| qlt = (struct ipc_mem_lite_gen_tbl *)(ipc_mux->ul_adb.qlth_skb) |
| ->data; |
| qlt->signature = cpu_to_le32(MUX_SIG_QLTH); |
| qlt->length = cpu_to_le16(qlt_size); |
| qlt->if_id = i; |
| qlt->vfl_length = MUX_QUEUE_LEVEL * sizeof(struct mux_lite_vfl); |
| qlt->reserved[0] = 0; |
| qlt->reserved[1] = 0; |
| |
| qlt->vfl.nr_of_bytes = cpu_to_le32(session->ul_list.qlen); |
| |
| /* Add QLT to the transfer list. */ |
| skb_queue_tail(&ipc_mux->channel->ul_list, |
| ipc_mux->ul_adb.qlth_skb); |
| |
| qlt_updated = true; |
| ipc_mux->ul_adb.qlth_skb = NULL; |
| } |
| |
| if (qlt_updated) |
| /* Updates the TDs with ul_list */ |
| (void)ipc_imem_ul_write_td(ipc_mux->imem); |
| |
| return qlt_updated; |
| } |
| |
| /* Checks the available credits for the specified session and returns |
| * number of packets for which credits are available. |
| */ |
| static int ipc_mux_ul_bytes_credits_check(struct iosm_mux *ipc_mux, |
| struct mux_session *session, |
| struct sk_buff_head *ul_list, |
| int max_nr_of_pkts) |
| { |
| int pkts_to_send = 0; |
| struct sk_buff *skb; |
| int credits = 0; |
| |
| if (ipc_mux->ul_flow == MUX_UL_ON_CREDITS) { |
| credits = session->ul_flow_credits; |
| if (credits <= 0) { |
| dev_dbg(ipc_mux->dev, |
| "FC::if_id[%d] Insuff.Credits/Qlen:%d/%u", |
| session->if_id, session->ul_flow_credits, |
| session->ul_list.qlen); /* nr_of_bytes */ |
| return 0; |
| } |
| } else { |
| credits = IPC_MEM_MUX_UL_FLOWCTRL_HIGH_B - |
| ipc_mux->ul_data_pend_bytes; |
| if (credits <= 0) { |
| ipc_mux_stop_tx_for_all_sessions(ipc_mux); |
| |
| dev_dbg(ipc_mux->dev, |
| "if_id[%d] encod. fail Bytes: %llu, thresh: %d", |
| session->if_id, ipc_mux->ul_data_pend_bytes, |
| IPC_MEM_MUX_UL_FLOWCTRL_HIGH_B); |
| return 0; |
| } |
| } |
| |
| /* Check if there are enough credits/bytes available to send the |
| * requested max_nr_of_pkts. Otherwise restrict the nr_of_pkts |
| * depending on available credits. |
| */ |
| skb_queue_walk(ul_list, skb) |
| { |
| if (!(credits >= skb->len && pkts_to_send < max_nr_of_pkts)) |
| break; |
| credits -= skb->len; |
| pkts_to_send++; |
| } |
| |
| return pkts_to_send; |
| } |
| |
| /* Encode the UL IP packet according to Lite spec. */ |
| static int ipc_mux_ul_adgh_encode(struct iosm_mux *ipc_mux, int session_id, |
| struct mux_session *session, |
| struct sk_buff_head *ul_list, |
| struct mux_adb *adb, int nr_of_pkts) |
| { |
| int offset = sizeof(struct mux_adgh); |
| int adb_updated = -EINVAL; |
| struct sk_buff *src_skb; |
| int aligned_size = 0; |
| int nr_of_skb = 0; |
| u32 pad_len = 0; |
| |
| /* Re-calculate the number of packets depending on number of bytes to be |
| * processed/available credits. |
| */ |
| nr_of_pkts = ipc_mux_ul_bytes_credits_check(ipc_mux, session, ul_list, |
| nr_of_pkts); |
| |
| /* If calculated nr_of_pkts from available credits is <= 0 |
| * then nothing to do. |
| */ |
| if (nr_of_pkts <= 0) |
| return 0; |
| |
| /* Read configured UL head_pad_length for session.*/ |
| if (session->ul_head_pad_len > IPC_MEM_DL_ETH_OFFSET) |
| pad_len = session->ul_head_pad_len - IPC_MEM_DL_ETH_OFFSET; |
| |
| /* Process all pending UL packets for this session |
| * depending on the allocated datagram table size. |
| */ |
| while (nr_of_pkts > 0) { |
| /* get destination skb allocated */ |
| if (ipc_mux_ul_adb_allocate(ipc_mux, adb, &ipc_mux->size_needed, |
| IOSM_AGGR_MUX_SIG_ADGH)) { |
| dev_err(ipc_mux->dev, "no reserved memory for ADGH"); |
| return -ENOMEM; |
| } |
| |
| /* Peek at the head of the list. */ |
| src_skb = skb_peek(ul_list); |
| if (!src_skb) { |
| dev_err(ipc_mux->dev, |
| "skb peek return NULL with count : %d", |
| nr_of_pkts); |
| break; |
| } |
| |
| /* Calculate the memory value. */ |
| aligned_size = ALIGN((pad_len + src_skb->len), 4); |
| |
| ipc_mux->size_needed = sizeof(struct mux_adgh) + aligned_size; |
| |
| if (ipc_mux->size_needed > adb->size) { |
| dev_dbg(ipc_mux->dev, "size needed %d, adgh size %d", |
| ipc_mux->size_needed, adb->size); |
| /* Return 1 if any IP packet is added to the transfer |
| * list. |
| */ |
| return nr_of_skb ? 1 : 0; |
| } |
| |
| /* Add buffer (without head padding to next pending transfer) */ |
| memcpy(adb->buf + offset + pad_len, src_skb->data, |
| src_skb->len); |
| |
| adb->adgh->signature = cpu_to_le32(IOSM_AGGR_MUX_SIG_ADGH); |
| adb->adgh->if_id = session_id; |
| adb->adgh->length = |
| cpu_to_le16(sizeof(struct mux_adgh) + pad_len + |
| src_skb->len); |
| adb->adgh->service_class = src_skb->priority; |
| adb->adgh->next_count = --nr_of_pkts; |
| adb->dg_cnt_total++; |
| adb->payload_size += src_skb->len; |
| |
| if (ipc_mux->ul_flow == MUX_UL_ON_CREDITS) |
| /* Decrement the credit value as we are processing the |
| * datagram from the UL list. |
| */ |
| session->ul_flow_credits -= src_skb->len; |
| |
| /* Remove the processed elements and free it. */ |
| src_skb = skb_dequeue(ul_list); |
| dev_kfree_skb(src_skb); |
| nr_of_skb++; |
| |
| ipc_mux_ul_adgh_finish(ipc_mux); |
| } |
| |
| if (nr_of_skb) { |
| /* Send QLT info to modem if pending bytes > high watermark |
| * in case of mux lite |
| */ |
| if (ipc_mux->ul_flow == MUX_UL_ON_CREDITS || |
| ipc_mux->ul_data_pend_bytes >= |
| IPC_MEM_MUX_UL_FLOWCTRL_LOW_B) |
| adb_updated = ipc_mux_lite_send_qlt(ipc_mux); |
| else |
| adb_updated = 1; |
| |
| /* Updates the TDs with ul_list */ |
| (void)ipc_imem_ul_write_td(ipc_mux->imem); |
| } |
| |
| return adb_updated; |
| } |
| |
| /** |
| * ipc_mux_ul_adb_update_ql - Adds Queue Level Table and Queue Level to ADB |
| * @ipc_mux: pointer to MUX instance data |
| * @p_adb: pointer to UL aggegated data block |
| * @session_id: session id |
| * @qlth_n_ql_size: Length (in bytes) of the datagram table |
| * @ul_list: pointer to skb buffer head |
| */ |
| void ipc_mux_ul_adb_update_ql(struct iosm_mux *ipc_mux, struct mux_adb *p_adb, |
| int session_id, int qlth_n_ql_size, |
| struct sk_buff_head *ul_list) |
| { |
| int qlevel = ul_list->qlen; |
| struct mux_qlth *p_qlt; |
| |
| p_qlt = (struct mux_qlth *)p_adb->pp_qlt[session_id]; |
| |
| /* Initialize QLTH if not been done */ |
| if (p_adb->qlt_updated[session_id] == 0) { |
| p_qlt->signature = cpu_to_le32(MUX_SIG_QLTH); |
| p_qlt->if_id = session_id; |
| p_qlt->table_length = cpu_to_le16(qlth_n_ql_size); |
| p_qlt->reserved = 0; |
| p_qlt->reserved2 = 0; |
| } |
| |
| /* Update Queue Level information always */ |
| p_qlt->ql.nr_of_bytes = cpu_to_le32(qlevel); |
| p_adb->qlt_updated[session_id] = 1; |
| } |
| |
| /* Update the next table index. */ |
| static int mux_ul_dg_update_tbl_index(struct iosm_mux *ipc_mux, |
| int session_id, |
| struct sk_buff_head *ul_list, |
| struct mux_adth_dg *dg, |
| int aligned_size, |
| u32 qlth_n_ql_size, |
| struct mux_adb *adb, |
| struct sk_buff *src_skb) |
| { |
| ipc_mux_ul_adb_update_ql(ipc_mux, adb, session_id, |
| qlth_n_ql_size, ul_list); |
| ipc_mux_ul_adb_finish(ipc_mux); |
| if (ipc_mux_ul_adb_allocate(ipc_mux, adb, &ipc_mux->size_needed, |
| IOSM_AGGR_MUX_SIG_ADBH)) { |
| dev_kfree_skb(src_skb); |
| return -ENOMEM; |
| } |
| ipc_mux->size_needed = le32_to_cpu(adb->adbh->block_length); |
| |
| ipc_mux->size_needed += offsetof(struct mux_adth, dg); |
| ipc_mux->size_needed += qlth_n_ql_size; |
| ipc_mux->size_needed += sizeof(*dg) + aligned_size; |
| return 0; |
| } |
| |
| /* Process encode session UL data. */ |
| static int mux_ul_dg_encode(struct iosm_mux *ipc_mux, struct mux_adb *adb, |
| struct mux_adth_dg *dg, |
| struct sk_buff_head *ul_list, |
| struct sk_buff *src_skb, int session_id, |
| int pkt_to_send, u32 qlth_n_ql_size, |
| int *out_offset, int head_pad_len) |
| { |
| int aligned_size; |
| int offset = *out_offset; |
| unsigned long flags; |
| int nr_of_skb = 0; |
| |
| while (pkt_to_send > 0) { |
| /* Peek at the head of the list. */ |
| src_skb = skb_peek(ul_list); |
| if (!src_skb) { |
| dev_err(ipc_mux->dev, |
| "skb peek return NULL with count : %d", |
| pkt_to_send); |
| return -1; |
| } |
| aligned_size = ALIGN((head_pad_len + src_skb->len), 4); |
| ipc_mux->size_needed += sizeof(*dg) + aligned_size; |
| |
| if (ipc_mux->size_needed > adb->size || |
| ((ipc_mux->size_needed + ipc_mux->ul_data_pend_bytes) >= |
| IPC_MEM_MUX_UL_FLOWCTRL_HIGH_B)) { |
| *adb->next_table_index = offset; |
| if (mux_ul_dg_update_tbl_index(ipc_mux, session_id, |
| ul_list, dg, |
| aligned_size, |
| qlth_n_ql_size, adb, |
| src_skb) < 0) |
| return -ENOMEM; |
| nr_of_skb = 0; |
| offset = le32_to_cpu(adb->adbh->block_length); |
| /* Load pointer to next available datagram entry */ |
| dg = adb->dg[session_id] + adb->dg_count[session_id]; |
| } |
| /* Add buffer without head padding to next pending transfer. */ |
| memcpy(adb->buf + offset + head_pad_len, |
| src_skb->data, src_skb->len); |
| /* Setup datagram entry. */ |
| dg->datagram_index = cpu_to_le32(offset); |
| dg->datagram_length = cpu_to_le16(src_skb->len + head_pad_len); |
| dg->service_class = (((struct sk_buff *)src_skb)->priority); |
| dg->reserved = 0; |
| adb->dg_cnt_total++; |
| adb->payload_size += le16_to_cpu(dg->datagram_length); |
| dg++; |
| adb->dg_count[session_id]++; |
| offset += aligned_size; |
| /* Remove the processed elements and free it. */ |
| spin_lock_irqsave(&ul_list->lock, flags); |
| src_skb = __skb_dequeue(ul_list); |
| spin_unlock_irqrestore(&ul_list->lock, flags); |
| |
| dev_kfree_skb(src_skb); |
| nr_of_skb++; |
| pkt_to_send--; |
| } |
| *out_offset = offset; |
| return nr_of_skb; |
| } |
| |
| /* Process encode session UL data to ADB. */ |
| static int mux_ul_adb_encode(struct iosm_mux *ipc_mux, int session_id, |
| struct mux_session *session, |
| struct sk_buff_head *ul_list, struct mux_adb *adb, |
| int pkt_to_send) |
| { |
| int adb_updated = -EINVAL; |
| int head_pad_len, offset; |
| struct sk_buff *src_skb = NULL; |
| struct mux_adth_dg *dg; |
| u32 qlth_n_ql_size; |
| |
| /* If any of the opened session has set Flow Control ON then limit the |
| * UL data to mux_flow_ctrl_high_thresh_b bytes |
| */ |
| if (ipc_mux->ul_data_pend_bytes >= |
| IPC_MEM_MUX_UL_FLOWCTRL_HIGH_B) { |
| ipc_mux_stop_tx_for_all_sessions(ipc_mux); |
| return adb_updated; |
| } |
| |
| qlth_n_ql_size = offsetof(struct mux_qlth, ql) + |
| MUX_QUEUE_LEVEL * sizeof(struct mux_qlth_ql); |
| head_pad_len = session->ul_head_pad_len; |
| |
| if (session->ul_head_pad_len > IPC_MEM_DL_ETH_OFFSET) |
| head_pad_len = session->ul_head_pad_len - IPC_MEM_DL_ETH_OFFSET; |
| |
| if (ipc_mux_ul_adb_allocate(ipc_mux, adb, &ipc_mux->size_needed, |
| IOSM_AGGR_MUX_SIG_ADBH)) |
| return -ENOMEM; |
| |
| offset = le32_to_cpu(adb->adbh->block_length); |
| |
| if (ipc_mux->size_needed == 0) |
| ipc_mux->size_needed = offset; |
| |
| /* Calculate the size needed for ADTH, QLTH and QL*/ |
| if (adb->dg_count[session_id] == 0) { |
| ipc_mux->size_needed += offsetof(struct mux_adth, dg); |
| ipc_mux->size_needed += qlth_n_ql_size; |
| } |
| |
| dg = adb->dg[session_id] + adb->dg_count[session_id]; |
| |
| if (mux_ul_dg_encode(ipc_mux, adb, dg, ul_list, src_skb, |
| session_id, pkt_to_send, qlth_n_ql_size, &offset, |
| head_pad_len) > 0) { |
| adb_updated = 1; |
| *adb->next_table_index = offset; |
| ipc_mux_ul_adb_update_ql(ipc_mux, adb, session_id, |
| qlth_n_ql_size, ul_list); |
| adb->adbh->block_length = cpu_to_le32(offset); |
| } |
| |
| return adb_updated; |
| } |
| |
| bool ipc_mux_ul_data_encode(struct iosm_mux *ipc_mux) |
| { |
| struct sk_buff_head *ul_list; |
| struct mux_session *session; |
| int updated = 0; |
| int session_id; |
| int dg_n; |
| int i; |
| |
| if (!ipc_mux || ipc_mux->state != MUX_S_ACTIVE || |
| ipc_mux->adb_prep_ongoing) |
| return false; |
| |
| ipc_mux->adb_prep_ongoing = true; |
| |
| for (i = 0; i < IPC_MEM_MUX_IP_SESSION_ENTRIES; i++) { |
| session_id = ipc_mux->rr_next_session; |
| session = &ipc_mux->session[session_id]; |
| |
| /* Go to next handle rr_next_session overflow */ |
| ipc_mux->rr_next_session++; |
| if (ipc_mux->rr_next_session >= IPC_MEM_MUX_IP_SESSION_ENTRIES) |
| ipc_mux->rr_next_session = 0; |
| |
| if (!session->wwan || session->flow_ctl_mask || |
| session->net_tx_stop) |
| continue; |
| |
| ul_list = &session->ul_list; |
| |
| /* Is something pending in UL and flow ctrl off */ |
| dg_n = skb_queue_len(ul_list); |
| if (dg_n > MUX_MAX_UL_DG_ENTRIES) |
| dg_n = MUX_MAX_UL_DG_ENTRIES; |
| |
| if (dg_n == 0) |
| /* Nothing to do for ipc_mux session |
| * -> try next session id. |
| */ |
| continue; |
| if (ipc_mux->protocol == MUX_LITE) |
| updated = ipc_mux_ul_adgh_encode(ipc_mux, session_id, |
| session, ul_list, |
| &ipc_mux->ul_adb, |
| dg_n); |
| else |
| updated = mux_ul_adb_encode(ipc_mux, session_id, |
| session, ul_list, |
| &ipc_mux->ul_adb, |
| dg_n); |
| } |
| |
| ipc_mux->adb_prep_ongoing = false; |
| return updated == 1; |
| } |
| |
| /* Calculates the Payload from any given ADB. */ |
| static int ipc_mux_get_payload_from_adb(struct iosm_mux *ipc_mux, |
| struct mux_adbh *p_adbh) |
| { |
| struct mux_adth_dg *dg; |
| struct mux_adth *adth; |
| u32 payload_size = 0; |
| u32 next_table_idx; |
| int nr_of_dg, i; |
| |
| /* Process the aggregated datagram tables. */ |
| next_table_idx = le32_to_cpu(p_adbh->first_table_index); |
| |
| if (next_table_idx < sizeof(struct mux_adbh)) { |
| dev_err(ipc_mux->dev, "unexpected empty ADB"); |
| return payload_size; |
| } |
| |
| while (next_table_idx != 0) { |
| /* Get the reference to the table header. */ |
| adth = (struct mux_adth *)((u8 *)p_adbh + next_table_idx); |
| |
| if (adth->signature == cpu_to_le32(IOSM_AGGR_MUX_SIG_ADTH)) { |
| nr_of_dg = (le16_to_cpu(adth->table_length) - |
| sizeof(struct mux_adth) + |
| sizeof(struct mux_adth_dg)) / |
| sizeof(struct mux_adth_dg); |
| |
| if (nr_of_dg <= 0) |
| return payload_size; |
| |
| dg = &adth->dg; |
| |
| for (i = 0; i < nr_of_dg; i++, dg++) { |
| if (le32_to_cpu(dg->datagram_index) < |
| sizeof(struct mux_adbh)) { |
| return payload_size; |
| } |
| payload_size += |
| le16_to_cpu(dg->datagram_length); |
| } |
| } |
| next_table_idx = le32_to_cpu(adth->next_table_index); |
| } |
| |
| return payload_size; |
| } |
| |
| void ipc_mux_ul_encoded_process(struct iosm_mux *ipc_mux, struct sk_buff *skb) |
| { |
| union mux_type_header hr; |
| u16 adgh_len; |
| int payload; |
| |
| if (ipc_mux->protocol == MUX_LITE) { |
| hr.adgh = (struct mux_adgh *)skb->data; |
| adgh_len = le16_to_cpu(hr.adgh->length); |
| if (hr.adgh->signature == cpu_to_le32(IOSM_AGGR_MUX_SIG_ADGH) && |
| ipc_mux->ul_flow == MUX_UL) |
| ipc_mux->ul_data_pend_bytes = |
| ipc_mux->ul_data_pend_bytes - adgh_len; |
| } else { |
| hr.adbh = (struct mux_adbh *)(skb->data); |
| payload = ipc_mux_get_payload_from_adb(ipc_mux, hr.adbh); |
| ipc_mux->ul_data_pend_bytes -= payload; |
| } |
| |
| if (ipc_mux->ul_flow == MUX_UL) |
| dev_dbg(ipc_mux->dev, "ul_data_pend_bytes: %lld", |
| ipc_mux->ul_data_pend_bytes); |
| |
| /* Reset the skb settings. */ |
| skb->tail = 0; |
| skb->len = 0; |
| |
| /* Add the consumed ADB to the free list. */ |
| skb_queue_tail((&ipc_mux->ul_adb.free_list), skb); |
| } |
| |
| /* Start the NETIF uplink send transfer in MUX mode. */ |
| static int ipc_mux_tq_ul_trigger_encode(struct iosm_imem *ipc_imem, int arg, |
| void *msg, size_t size) |
| { |
| struct iosm_mux *ipc_mux = ipc_imem->mux; |
| bool ul_data_pend = false; |
| |
| /* Add session UL data to a ADB and ADGH */ |
| ul_data_pend = ipc_mux_ul_data_encode(ipc_mux); |
| if (ul_data_pend) { |
| if (ipc_mux->protocol == MUX_AGGREGATION) |
| ipc_imem_adb_timer_start(ipc_mux->imem); |
| |
| /* Delay the doorbell irq */ |
| ipc_imem_td_update_timer_start(ipc_mux->imem); |
| } |
| /* reset the debounce flag */ |
| ipc_mux->ev_mux_net_transmit_pending = false; |
| |
| return 0; |
| } |
| |
| int ipc_mux_ul_trigger_encode(struct iosm_mux *ipc_mux, int if_id, |
| struct sk_buff *skb) |
| { |
| struct mux_session *session = &ipc_mux->session[if_id]; |
| int ret = -EINVAL; |
| |
| if (ipc_mux->channel && |
| ipc_mux->channel->state != IMEM_CHANNEL_ACTIVE) { |
| dev_err(ipc_mux->dev, |
| "channel state is not IMEM_CHANNEL_ACTIVE"); |
| goto out; |
| } |
| |
| if (!session->wwan) { |
| dev_err(ipc_mux->dev, "session net ID is NULL"); |
| ret = -EFAULT; |
| goto out; |
| } |
| |
| /* Session is under flow control. |
| * Check if packet can be queued in session list, if not |
| * suspend net tx |
| */ |
| if (skb_queue_len(&session->ul_list) >= |
| (session->net_tx_stop ? |
| IPC_MEM_MUX_UL_SESS_FCON_THRESHOLD : |
| (IPC_MEM_MUX_UL_SESS_FCON_THRESHOLD * |
| IPC_MEM_MUX_UL_SESS_FCOFF_THRESHOLD_FACTOR))) { |
| ipc_mux_netif_tx_flowctrl(session, session->if_id, true); |
| ret = -EBUSY; |
| goto out; |
| } |
| |
| /* Add skb to the uplink skb accumulator. */ |
| skb_queue_tail(&session->ul_list, skb); |
| |
| /* Inform the IPC kthread to pass uplink IP packets to CP. */ |
| if (!ipc_mux->ev_mux_net_transmit_pending) { |
| ipc_mux->ev_mux_net_transmit_pending = true; |
| ret = ipc_task_queue_send_task(ipc_mux->imem, |
| ipc_mux_tq_ul_trigger_encode, 0, |
| NULL, 0, false); |
| if (ret) |
| goto out; |
| } |
| dev_dbg(ipc_mux->dev, "mux ul if[%d] qlen=%d/%u, len=%d/%d, prio=%d", |
| if_id, skb_queue_len(&session->ul_list), session->ul_list.qlen, |
| skb->len, skb->truesize, skb->priority); |
| ret = 0; |
| out: |
| return ret; |
| } |