bus: mhi: host: Drop the auto_queue support
Now that the only user of the 'auto_queue' feature, (QRTR) has been converted to manage the buffers on its own, drop the code related to it. Signed-off-by: Manivannan Sadhasivam <manivannan.sadhasivam@oss.qualcomm.com> Reviewed-by: Loic Poulain <loic.poulain@oss.qualcomm.com> Reviewed-by: Jeff Hugo <jeff.hugo@oss.qualcomm.com> Link: https://patch.msgid.link/20251218-qrtr-fix-v2-2-c7499bfcfbe0@oss.qualcomm.com
This commit is contained in:
parent
51731792a2
commit
4a9ba211d0
|
|
@ -841,18 +841,8 @@ static int parse_ch_cfg(struct mhi_controller *mhi_cntrl,
|
|||
mhi_chan->lpm_notify = ch_cfg->lpm_notify;
|
||||
mhi_chan->offload_ch = ch_cfg->offload_channel;
|
||||
mhi_chan->db_cfg.reset_req = ch_cfg->doorbell_mode_switch;
|
||||
mhi_chan->pre_alloc = ch_cfg->auto_queue;
|
||||
mhi_chan->wake_capable = ch_cfg->wake_capable;
|
||||
|
||||
/*
|
||||
* If MHI host allocates buffers, then the channel direction
|
||||
* should be DMA_FROM_DEVICE
|
||||
*/
|
||||
if (mhi_chan->pre_alloc && mhi_chan->dir != DMA_FROM_DEVICE) {
|
||||
dev_err(dev, "Invalid channel configuration\n");
|
||||
goto error_chan_cfg;
|
||||
}
|
||||
|
||||
/*
|
||||
* Bi-directional and direction less channel must be an
|
||||
* offload channel
|
||||
|
|
|
|||
|
|
@ -286,7 +286,6 @@ struct mhi_chan {
|
|||
bool lpm_notify;
|
||||
bool configured;
|
||||
bool offload_ch;
|
||||
bool pre_alloc;
|
||||
bool wake_capable;
|
||||
};
|
||||
|
||||
|
|
@ -389,8 +388,6 @@ int mhi_rddm_prepare(struct mhi_controller *mhi_cntrl,
|
|||
struct image_info *img_info);
|
||||
void mhi_fw_load_handler(struct mhi_controller *mhi_cntrl);
|
||||
|
||||
/* Automatically allocate and queue inbound buffers */
|
||||
#define MHI_CH_INBOUND_ALLOC_BUFS BIT(0)
|
||||
int mhi_init_chan_ctxt(struct mhi_controller *mhi_cntrl,
|
||||
struct mhi_chan *mhi_chan);
|
||||
void mhi_deinit_chan_ctxt(struct mhi_controller *mhi_cntrl,
|
||||
|
|
|
|||
|
|
@ -664,23 +664,6 @@ static int parse_xfer_event(struct mhi_controller *mhi_cntrl,
|
|||
mhi_cntrl->runtime_put(mhi_cntrl);
|
||||
}
|
||||
|
||||
/*
|
||||
* Recycle the buffer if buffer is pre-allocated,
|
||||
* if there is an error, not much we can do apart
|
||||
* from dropping the packet
|
||||
*/
|
||||
if (mhi_chan->pre_alloc) {
|
||||
if (mhi_queue_buf(mhi_chan->mhi_dev,
|
||||
mhi_chan->dir,
|
||||
buf_info->cb_buf,
|
||||
buf_info->len, MHI_EOT)) {
|
||||
dev_err(dev,
|
||||
"Error recycling buffer for chan:%d\n",
|
||||
mhi_chan->chan);
|
||||
kfree(buf_info->cb_buf);
|
||||
}
|
||||
}
|
||||
|
||||
read_lock_bh(&mhi_chan->lock);
|
||||
}
|
||||
break;
|
||||
|
|
@ -1177,17 +1160,12 @@ static int mhi_queue(struct mhi_device *mhi_dev, struct mhi_buf_info *buf_info,
|
|||
int mhi_queue_skb(struct mhi_device *mhi_dev, enum dma_data_direction dir,
|
||||
struct sk_buff *skb, size_t len, enum mhi_flags mflags)
|
||||
{
|
||||
struct mhi_chan *mhi_chan = (dir == DMA_TO_DEVICE) ? mhi_dev->ul_chan :
|
||||
mhi_dev->dl_chan;
|
||||
struct mhi_buf_info buf_info = { };
|
||||
|
||||
buf_info.v_addr = skb->data;
|
||||
buf_info.cb_buf = skb;
|
||||
buf_info.len = len;
|
||||
|
||||
if (unlikely(mhi_chan->pre_alloc))
|
||||
return -EINVAL;
|
||||
|
||||
return mhi_queue(mhi_dev, &buf_info, dir, mflags);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(mhi_queue_skb);
|
||||
|
|
@ -1472,45 +1450,6 @@ static int mhi_prepare_channel(struct mhi_controller *mhi_cntrl,
|
|||
if (ret)
|
||||
goto error_pm_state;
|
||||
|
||||
if (mhi_chan->dir == DMA_FROM_DEVICE)
|
||||
mhi_chan->pre_alloc = !!(flags & MHI_CH_INBOUND_ALLOC_BUFS);
|
||||
|
||||
/* Pre-allocate buffer for xfer ring */
|
||||
if (mhi_chan->pre_alloc) {
|
||||
int nr_el = get_nr_avail_ring_elements(mhi_cntrl,
|
||||
&mhi_chan->tre_ring);
|
||||
size_t len = mhi_cntrl->buffer_len;
|
||||
|
||||
while (nr_el--) {
|
||||
void *buf;
|
||||
struct mhi_buf_info info = { };
|
||||
|
||||
buf = kmalloc(len, GFP_KERNEL);
|
||||
if (!buf) {
|
||||
ret = -ENOMEM;
|
||||
goto error_pre_alloc;
|
||||
}
|
||||
|
||||
/* Prepare transfer descriptors */
|
||||
info.v_addr = buf;
|
||||
info.cb_buf = buf;
|
||||
info.len = len;
|
||||
ret = mhi_gen_tre(mhi_cntrl, mhi_chan, &info, MHI_EOT);
|
||||
if (ret) {
|
||||
kfree(buf);
|
||||
goto error_pre_alloc;
|
||||
}
|
||||
}
|
||||
|
||||
read_lock_bh(&mhi_cntrl->pm_lock);
|
||||
if (MHI_DB_ACCESS_VALID(mhi_cntrl)) {
|
||||
read_lock_irq(&mhi_chan->lock);
|
||||
mhi_ring_chan_db(mhi_cntrl, mhi_chan);
|
||||
read_unlock_irq(&mhi_chan->lock);
|
||||
}
|
||||
read_unlock_bh(&mhi_cntrl->pm_lock);
|
||||
}
|
||||
|
||||
mutex_unlock(&mhi_chan->mutex);
|
||||
|
||||
return 0;
|
||||
|
|
@ -1522,12 +1461,6 @@ error_pm_state:
|
|||
error_init_chan:
|
||||
mutex_unlock(&mhi_chan->mutex);
|
||||
|
||||
return ret;
|
||||
|
||||
error_pre_alloc:
|
||||
mutex_unlock(&mhi_chan->mutex);
|
||||
mhi_unprepare_channel(mhi_cntrl, mhi_chan);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
|
@ -1600,12 +1533,8 @@ static void mhi_reset_data_chan(struct mhi_controller *mhi_cntrl,
|
|||
mhi_del_ring_element(mhi_cntrl, buf_ring);
|
||||
mhi_del_ring_element(mhi_cntrl, tre_ring);
|
||||
|
||||
if (mhi_chan->pre_alloc) {
|
||||
kfree(buf_info->cb_buf);
|
||||
} else {
|
||||
result.buf_addr = buf_info->cb_buf;
|
||||
mhi_chan->xfer_cb(mhi_chan->mhi_dev, &result);
|
||||
}
|
||||
result.buf_addr = buf_info->cb_buf;
|
||||
mhi_chan->xfer_cb(mhi_chan->mhi_dev, &result);
|
||||
}
|
||||
}
|
||||
|
||||
|
|
@ -1666,12 +1595,6 @@ int mhi_prepare_for_transfer(struct mhi_device *mhi_dev)
|
|||
}
|
||||
EXPORT_SYMBOL_GPL(mhi_prepare_for_transfer);
|
||||
|
||||
int mhi_prepare_for_transfer_autoqueue(struct mhi_device *mhi_dev)
|
||||
{
|
||||
return __mhi_prepare_for_transfer(mhi_dev, MHI_CH_INBOUND_ALLOC_BUFS);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(mhi_prepare_for_transfer_autoqueue);
|
||||
|
||||
void mhi_unprepare_from_transfer(struct mhi_device *mhi_dev)
|
||||
{
|
||||
struct mhi_controller *mhi_cntrl = mhi_dev->mhi_cntrl;
|
||||
|
|
|
|||
|
|
@ -215,7 +215,6 @@ enum mhi_db_brst_mode {
|
|||
* @lpm_notify: The channel master requires low power mode notifications
|
||||
* @offload_channel: The client manages the channel completely
|
||||
* @doorbell_mode_switch: Channel switches to doorbell mode on M0 transition
|
||||
* @auto_queue: Framework will automatically queue buffers for DL traffic
|
||||
* @wake-capable: Channel capable of waking up the system
|
||||
*/
|
||||
struct mhi_channel_config {
|
||||
|
|
@ -232,7 +231,6 @@ struct mhi_channel_config {
|
|||
bool lpm_notify;
|
||||
bool offload_channel;
|
||||
bool doorbell_mode_switch;
|
||||
bool auto_queue;
|
||||
bool wake_capable;
|
||||
};
|
||||
|
||||
|
|
@ -743,18 +741,6 @@ void mhi_device_put(struct mhi_device *mhi_dev);
|
|||
*/
|
||||
int mhi_prepare_for_transfer(struct mhi_device *mhi_dev);
|
||||
|
||||
/**
|
||||
* mhi_prepare_for_transfer_autoqueue - Setup UL and DL channels with auto queue
|
||||
* buffers for DL traffic
|
||||
* @mhi_dev: Device associated with the channels
|
||||
*
|
||||
* Allocate and initialize the channel context and also issue the START channel
|
||||
* command to both channels. Channels can be started only if both host and
|
||||
* device execution environments match and channels are in a DISABLED state.
|
||||
* The MHI core will automatically allocate and queue buffers for the DL traffic.
|
||||
*/
|
||||
int mhi_prepare_for_transfer_autoqueue(struct mhi_device *mhi_dev);
|
||||
|
||||
/**
|
||||
* mhi_unprepare_from_transfer - Reset UL and DL channels for data transfer.
|
||||
* Issue the RESET channel command and let the
|
||||
|
|
|
|||
Loading…
Reference in New Issue