| /* |
| * cmt_speech.c - HSI CMT speech driver |
| * |
| * Copyright (C) 2008,2009,2010 Nokia Corporation. All rights reserved. |
| * |
| * Contact: Kai Vehmanen <kai.vehmanen@nokia.com> |
| * Original author: Peter Ujfalusi <peter.ujfalusi@nokia.com> |
| * |
| * This program is free software; you can redistribute it and/or |
| * modify it under the terms of the GNU General Public License |
| * version 2 as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it will be useful, but |
| * WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * General Public License for more details. |
| * |
| * You should have received a copy of the GNU General Public License |
| * along with this program; if not, write to the Free Software |
| * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA |
| * 02110-1301 USA |
| */ |
| |
| #include <linux/errno.h> |
| #include <linux/module.h> |
| #include <linux/types.h> |
| #include <linux/init.h> |
| #include <linux/device.h> |
| #include <linux/miscdevice.h> |
| #include <linux/mm.h> |
| #include <linux/slab.h> |
| #include <linux/fs.h> |
| #include <linux/poll.h> |
| #include <linux/sched.h> |
| #include <linux/ioctl.h> |
| #include <linux/uaccess.h> |
| #include <linux/pm_qos.h> |
| #include <linux/hsi/hsi.h> |
| #include <linux/hsi/ssi_protocol.h> |
| #include <linux/hsi/cs-protocol.h> |
| |
| #define CS_MMAP_SIZE PAGE_SIZE |
| |
| struct char_queue { |
| struct list_head list; |
| u32 msg; |
| }; |
| |
| struct cs_char { |
| unsigned int opened; |
| struct hsi_client *cl; |
| struct cs_hsi_iface *hi; |
| struct list_head chardev_queue; |
| struct list_head dataind_queue; |
| int dataind_pending; |
| /* mmap things */ |
| unsigned long mmap_base; |
| unsigned long mmap_size; |
| spinlock_t lock; |
| struct fasync_struct *async_queue; |
| wait_queue_head_t wait; |
| /* hsi channel ids */ |
| int channel_id_cmd; |
| int channel_id_data; |
| }; |
| |
| #define SSI_CHANNEL_STATE_READING 1 |
| #define SSI_CHANNEL_STATE_WRITING (1 << 1) |
| #define SSI_CHANNEL_STATE_POLL (1 << 2) |
| #define SSI_CHANNEL_STATE_ERROR (1 << 3) |
| |
| #define TARGET_MASK 0xf000000 |
| #define TARGET_REMOTE (1 << CS_DOMAIN_SHIFT) |
| #define TARGET_LOCAL 0 |
| |
| /* Number of pre-allocated commands buffers */ |
| #define CS_MAX_CMDS 4 |
| |
| /* |
| * During data transfers, transactions must be handled |
| * within 20ms (fixed value in cmtspeech HSI protocol) |
| */ |
| #define CS_QOS_LATENCY_FOR_DATA_USEC 20000 |
| |
| /* Timeout to wait for pending HSI transfers to complete */ |
| #define CS_HSI_TRANSFER_TIMEOUT_MS 500 |
| |
| |
| #define RX_PTR_BOUNDARY_SHIFT 8 |
| #define RX_PTR_MAX_SHIFT (RX_PTR_BOUNDARY_SHIFT + \ |
| CS_MAX_BUFFERS_SHIFT) |
| struct cs_hsi_iface { |
| struct hsi_client *cl; |
| struct hsi_client *master; |
| |
| unsigned int iface_state; |
| unsigned int wakeline_state; |
| unsigned int control_state; |
| unsigned int data_state; |
| |
| /* state exposed to application */ |
| struct cs_mmap_config_block *mmap_cfg; |
| |
| unsigned long mmap_base; |
| unsigned long mmap_size; |
| |
| unsigned int rx_slot; |
| unsigned int tx_slot; |
| |
| /* note: for security reasons, we do not trust the contents of |
| * mmap_cfg, but instead duplicate the variables here */ |
| unsigned int buf_size; |
| unsigned int rx_bufs; |
| unsigned int tx_bufs; |
| unsigned int rx_ptr_boundary; |
| unsigned int rx_offsets[CS_MAX_BUFFERS]; |
| unsigned int tx_offsets[CS_MAX_BUFFERS]; |
| |
| /* size of aligned memory blocks */ |
| unsigned int slot_size; |
| unsigned int flags; |
| |
| struct list_head cmdqueue; |
| |
| struct hsi_msg *data_rx_msg; |
| struct hsi_msg *data_tx_msg; |
| wait_queue_head_t datawait; |
| |
| struct pm_qos_request pm_qos_req; |
| |
| spinlock_t lock; |
| }; |
| |
| static struct cs_char cs_char_data; |
| |
| static void cs_hsi_read_on_control(struct cs_hsi_iface *hi); |
| static void cs_hsi_read_on_data(struct cs_hsi_iface *hi); |
| |
| static inline void rx_ptr_shift_too_big(void) |
| { |
| BUILD_BUG_ON((1LLU << RX_PTR_MAX_SHIFT) > UINT_MAX); |
| } |
| |
| static void cs_notify(u32 message, struct list_head *head) |
| { |
| struct char_queue *entry; |
| |
| spin_lock(&cs_char_data.lock); |
| |
| if (!cs_char_data.opened) { |
| spin_unlock(&cs_char_data.lock); |
| goto out; |
| } |
| |
| entry = kmalloc(sizeof(*entry), GFP_ATOMIC); |
| if (!entry) { |
| dev_err(&cs_char_data.cl->device, |
| "Can't allocate new entry for the queue.\n"); |
| spin_unlock(&cs_char_data.lock); |
| goto out; |
| } |
| |
| entry->msg = message; |
| list_add_tail(&entry->list, head); |
| |
| spin_unlock(&cs_char_data.lock); |
| |
| wake_up_interruptible(&cs_char_data.wait); |
| kill_fasync(&cs_char_data.async_queue, SIGIO, POLL_IN); |
| |
| out: |
| return; |
| } |
| |
| static u32 cs_pop_entry(struct list_head *head) |
| { |
| struct char_queue *entry; |
| u32 data; |
| |
| entry = list_entry(head->next, struct char_queue, list); |
| data = entry->msg; |
| list_del(&entry->list); |
| kfree(entry); |
| |
| return data; |
| } |
| |
| static void cs_notify_control(u32 message) |
| { |
| cs_notify(message, &cs_char_data.chardev_queue); |
| } |
| |
| static void cs_notify_data(u32 message, int maxlength) |
| { |
| cs_notify(message, &cs_char_data.dataind_queue); |
| |
| spin_lock(&cs_char_data.lock); |
| cs_char_data.dataind_pending++; |
| while (cs_char_data.dataind_pending > maxlength && |
| !list_empty(&cs_char_data.dataind_queue)) { |
| dev_dbg(&cs_char_data.cl->device, "data notification " |
| "queue overrun (%u entries)\n", cs_char_data.dataind_pending); |
| |
| cs_pop_entry(&cs_char_data.dataind_queue); |
| cs_char_data.dataind_pending--; |
| } |
| spin_unlock(&cs_char_data.lock); |
| } |
| |
| static inline void cs_set_cmd(struct hsi_msg *msg, u32 cmd) |
| { |
| u32 *data = sg_virt(msg->sgt.sgl); |
| *data = cmd; |
| } |
| |
| static inline u32 cs_get_cmd(struct hsi_msg *msg) |
| { |
| u32 *data = sg_virt(msg->sgt.sgl); |
| return *data; |
| } |
| |
| static void cs_release_cmd(struct hsi_msg *msg) |
| { |
| struct cs_hsi_iface *hi = msg->context; |
| |
| list_add_tail(&msg->link, &hi->cmdqueue); |
| } |
| |
| static void cs_cmd_destructor(struct hsi_msg *msg) |
| { |
| struct cs_hsi_iface *hi = msg->context; |
| |
| spin_lock(&hi->lock); |
| |
| dev_dbg(&cs_char_data.cl->device, "control cmd destructor\n"); |
| |
| if (hi->iface_state != CS_STATE_CLOSED) |
| dev_err(&hi->cl->device, "Cmd flushed while driver active\n"); |
| |
| if (msg->ttype == HSI_MSG_READ) |
| hi->control_state &= |
| ~(SSI_CHANNEL_STATE_POLL | SSI_CHANNEL_STATE_READING); |
| else if (msg->ttype == HSI_MSG_WRITE && |
| hi->control_state & SSI_CHANNEL_STATE_WRITING) |
| hi->control_state &= ~SSI_CHANNEL_STATE_WRITING; |
| |
| cs_release_cmd(msg); |
| |
| spin_unlock(&hi->lock); |
| } |
| |
| static struct hsi_msg *cs_claim_cmd(struct cs_hsi_iface* ssi) |
| { |
| struct hsi_msg *msg; |
| |
| BUG_ON(list_empty(&ssi->cmdqueue)); |
| |
| msg = list_first_entry(&ssi->cmdqueue, struct hsi_msg, link); |
| list_del(&msg->link); |
| msg->destructor = cs_cmd_destructor; |
| |
| return msg; |
| } |
| |
| static void cs_free_cmds(struct cs_hsi_iface *ssi) |
| { |
| struct hsi_msg *msg, *tmp; |
| |
| list_for_each_entry_safe(msg, tmp, &ssi->cmdqueue, link) { |
| list_del(&msg->link); |
| msg->destructor = NULL; |
| kfree(sg_virt(msg->sgt.sgl)); |
| hsi_free_msg(msg); |
| } |
| } |
| |
| static int cs_alloc_cmds(struct cs_hsi_iface *hi) |
| { |
| struct hsi_msg *msg; |
| u32 *buf; |
| unsigned int i; |
| |
| INIT_LIST_HEAD(&hi->cmdqueue); |
| |
| for (i = 0; i < CS_MAX_CMDS; i++) { |
| msg = hsi_alloc_msg(1, GFP_KERNEL); |
| if (!msg) |
| goto out; |
| buf = kmalloc(sizeof(*buf), GFP_KERNEL); |
| if (!buf) { |
| hsi_free_msg(msg); |
| goto out; |
| } |
| sg_init_one(msg->sgt.sgl, buf, sizeof(*buf)); |
| msg->channel = cs_char_data.channel_id_cmd; |
| msg->context = hi; |
| list_add_tail(&msg->link, &hi->cmdqueue); |
| } |
| |
| return 0; |
| |
| out: |
| cs_free_cmds(hi); |
| return -ENOMEM; |
| } |
| |
| static void cs_hsi_data_destructor(struct hsi_msg *msg) |
| { |
| struct cs_hsi_iface *hi = msg->context; |
| const char *dir = (msg->ttype == HSI_MSG_READ) ? "TX" : "RX"; |
| |
| dev_dbg(&cs_char_data.cl->device, "Freeing data %s message\n", dir); |
| |
| spin_lock(&hi->lock); |
| if (hi->iface_state != CS_STATE_CLOSED) |
| dev_err(&cs_char_data.cl->device, |
| "Data %s flush while device active\n", dir); |
| if (msg->ttype == HSI_MSG_READ) |
| hi->data_state &= |
| ~(SSI_CHANNEL_STATE_POLL | SSI_CHANNEL_STATE_READING); |
| else |
| hi->data_state &= ~SSI_CHANNEL_STATE_WRITING; |
| |
| msg->status = HSI_STATUS_COMPLETED; |
| if (unlikely(waitqueue_active(&hi->datawait))) |
| wake_up_interruptible(&hi->datawait); |
| |
| spin_unlock(&hi->lock); |
| } |
| |
| static int cs_hsi_alloc_data(struct cs_hsi_iface *hi) |
| { |
| struct hsi_msg *txmsg, *rxmsg; |
| int res = 0; |
| |
| rxmsg = hsi_alloc_msg(1, GFP_KERNEL); |
| if (!rxmsg) { |
| res = -ENOMEM; |
| goto out1; |
| } |
| rxmsg->channel = cs_char_data.channel_id_data; |
| rxmsg->destructor = cs_hsi_data_destructor; |
| rxmsg->context = hi; |
| |
| txmsg = hsi_alloc_msg(1, GFP_KERNEL); |
| if (!txmsg) { |
| res = -ENOMEM; |
| goto out2; |
| } |
| txmsg->channel = cs_char_data.channel_id_data; |
| txmsg->destructor = cs_hsi_data_destructor; |
| txmsg->context = hi; |
| |
| hi->data_rx_msg = rxmsg; |
| hi->data_tx_msg = txmsg; |
| |
| return 0; |
| |
| out2: |
| hsi_free_msg(rxmsg); |
| out1: |
| return res; |
| } |
| |
| static void cs_hsi_free_data_msg(struct hsi_msg *msg) |
| { |
| WARN_ON(msg->status != HSI_STATUS_COMPLETED && |
| msg->status != HSI_STATUS_ERROR); |
| hsi_free_msg(msg); |
| } |
| |
| static void cs_hsi_free_data(struct cs_hsi_iface *hi) |
| { |
| cs_hsi_free_data_msg(hi->data_rx_msg); |
| cs_hsi_free_data_msg(hi->data_tx_msg); |
| } |
| |
| static inline void __cs_hsi_error_pre(struct cs_hsi_iface *hi, |
| struct hsi_msg *msg, const char *info, |
| unsigned int *state) |
| { |
| spin_lock(&hi->lock); |
| dev_err(&hi->cl->device, "HSI %s error, msg %d, state %u\n", |
| info, msg->status, *state); |
| } |
| |
| static inline void __cs_hsi_error_post(struct cs_hsi_iface *hi) |
| { |
| spin_unlock(&hi->lock); |
| } |
| |
| static inline void __cs_hsi_error_read_bits(unsigned int *state) |
| { |
| *state |= SSI_CHANNEL_STATE_ERROR; |
| *state &= ~(SSI_CHANNEL_STATE_READING | SSI_CHANNEL_STATE_POLL); |
| } |
| |
| static inline void __cs_hsi_error_write_bits(unsigned int *state) |
| { |
| *state |= SSI_CHANNEL_STATE_ERROR; |
| *state &= ~SSI_CHANNEL_STATE_WRITING; |
| } |
| |
| static void cs_hsi_control_read_error(struct cs_hsi_iface *hi, |
| struct hsi_msg *msg) |
| { |
| __cs_hsi_error_pre(hi, msg, "control read", &hi->control_state); |
| cs_release_cmd(msg); |
| __cs_hsi_error_read_bits(&hi->control_state); |
| __cs_hsi_error_post(hi); |
| } |
| |
| static void cs_hsi_control_write_error(struct cs_hsi_iface *hi, |
| struct hsi_msg *msg) |
| { |
| __cs_hsi_error_pre(hi, msg, "control write", &hi->control_state); |
| cs_release_cmd(msg); |
| __cs_hsi_error_write_bits(&hi->control_state); |
| __cs_hsi_error_post(hi); |
| |
| } |
| |
| static void cs_hsi_data_read_error(struct cs_hsi_iface *hi, struct hsi_msg *msg) |
| { |
| __cs_hsi_error_pre(hi, msg, "data read", &hi->data_state); |
| __cs_hsi_error_read_bits(&hi->data_state); |
| __cs_hsi_error_post(hi); |
| } |
| |
| static void cs_hsi_data_write_error(struct cs_hsi_iface *hi, |
| struct hsi_msg *msg) |
| { |
| __cs_hsi_error_pre(hi, msg, "data write", &hi->data_state); |
| __cs_hsi_error_write_bits(&hi->data_state); |
| __cs_hsi_error_post(hi); |
| } |
| |
| static void cs_hsi_read_on_control_complete(struct hsi_msg *msg) |
| { |
| u32 cmd = cs_get_cmd(msg); |
| struct cs_hsi_iface *hi = msg->context; |
| |
| spin_lock(&hi->lock); |
| hi->control_state &= ~SSI_CHANNEL_STATE_READING; |
| if (msg->status == HSI_STATUS_ERROR) { |
| dev_err(&hi->cl->device, "Control RX error detected\n"); |
| cs_hsi_control_read_error(hi, msg); |
| spin_unlock(&hi->lock); |
| goto out; |
| } |
| dev_dbg(&hi->cl->device, "Read on control: %08X\n", cmd); |
| cs_release_cmd(msg); |
| if (hi->flags & CS_FEAT_TSTAMP_RX_CTRL) { |
| struct timespec tspec; |
| struct cs_timestamp *tstamp = |
| &hi->mmap_cfg->tstamp_rx_ctrl; |
| |
| ktime_get_ts(&tspec); |
| |
| tstamp->tv_sec = (__u32) tspec.tv_sec; |
| tstamp->tv_nsec = (__u32) tspec.tv_nsec; |
| } |
| spin_unlock(&hi->lock); |
| |
| cs_notify_control(cmd); |
| |
| out: |
| cs_hsi_read_on_control(hi); |
| } |
| |
| static void cs_hsi_peek_on_control_complete(struct hsi_msg *msg) |
| { |
| struct cs_hsi_iface *hi = msg->context; |
| int ret; |
| |
| if (msg->status == HSI_STATUS_ERROR) { |
| dev_err(&hi->cl->device, "Control peek RX error detected\n"); |
| cs_hsi_control_read_error(hi, msg); |
| return; |
| } |
| |
| WARN_ON(!(hi->control_state & SSI_CHANNEL_STATE_READING)); |
| |
| dev_dbg(&hi->cl->device, "Peek on control complete, reading\n"); |
| msg->sgt.nents = 1; |
| msg->complete = cs_hsi_read_on_control_complete; |
| ret = hsi_async_read(hi->cl, msg); |
| if (ret) |
| cs_hsi_control_read_error(hi, msg); |
| } |
| |
| static void cs_hsi_read_on_control(struct cs_hsi_iface *hi) |
| { |
| struct hsi_msg *msg; |
| int ret; |
| |
| spin_lock(&hi->lock); |
| if (hi->control_state & SSI_CHANNEL_STATE_READING) { |
| dev_err(&hi->cl->device, "Control read already pending (%d)\n", |
| hi->control_state); |
| spin_unlock(&hi->lock); |
| return; |
| } |
| if (hi->control_state & SSI_CHANNEL_STATE_ERROR) { |
| dev_err(&hi->cl->device, "Control read error (%d)\n", |
| hi->control_state); |
| spin_unlock(&hi->lock); |
| return; |
| } |
| hi->control_state |= SSI_CHANNEL_STATE_READING; |
| dev_dbg(&hi->cl->device, "Issuing RX on control\n"); |
| msg = cs_claim_cmd(hi); |
| spin_unlock(&hi->lock); |
| |
| msg->sgt.nents = 0; |
| msg->complete = cs_hsi_peek_on_control_complete; |
| ret = hsi_async_read(hi->cl, msg); |
| if (ret) |
| cs_hsi_control_read_error(hi, msg); |
| } |
| |
| static void cs_hsi_write_on_control_complete(struct hsi_msg *msg) |
| { |
| struct cs_hsi_iface *hi = msg->context; |
| if (msg->status == HSI_STATUS_COMPLETED) { |
| spin_lock(&hi->lock); |
| hi->control_state &= ~SSI_CHANNEL_STATE_WRITING; |
| cs_release_cmd(msg); |
| spin_unlock(&hi->lock); |
| } else if (msg->status == HSI_STATUS_ERROR) { |
| cs_hsi_control_write_error(hi, msg); |
| } else { |
| dev_err(&hi->cl->device, |
| "unexpected status in control write callback %d\n", |
| msg->status); |
| } |
| } |
| |
| static int cs_hsi_write_on_control(struct cs_hsi_iface *hi, u32 message) |
| { |
| struct hsi_msg *msg; |
| int ret; |
| |
| spin_lock(&hi->lock); |
| if (hi->control_state & SSI_CHANNEL_STATE_ERROR) { |
| spin_unlock(&hi->lock); |
| return -EIO; |
| } |
| if (hi->control_state & SSI_CHANNEL_STATE_WRITING) { |
| dev_err(&hi->cl->device, |
| "Write still pending on control channel.\n"); |
| spin_unlock(&hi->lock); |
| return -EBUSY; |
| } |
| hi->control_state |= SSI_CHANNEL_STATE_WRITING; |
| msg = cs_claim_cmd(hi); |
| spin_unlock(&hi->lock); |
| |
| cs_set_cmd(msg, message); |
| msg->sgt.nents = 1; |
| msg->complete = cs_hsi_write_on_control_complete; |
| dev_dbg(&hi->cl->device, |
| "Sending control message %08X\n", message); |
| ret = hsi_async_write(hi->cl, msg); |
| if (ret) { |
| dev_err(&hi->cl->device, |
| "async_write failed with %d\n", ret); |
| cs_hsi_control_write_error(hi, msg); |
| } |
| |
| /* |
| * Make sure control read is always pending when issuing |
| * new control writes. This is needed as the controller |
| * may flush our messages if e.g. the peer device reboots |
| * unexpectedly (and we cannot directly resubmit a new read from |
| * the message destructor; see cs_cmd_destructor()). |
| */ |
| if (!(hi->control_state & SSI_CHANNEL_STATE_READING)) { |
| dev_err(&hi->cl->device, "Restarting control reads\n"); |
| cs_hsi_read_on_control(hi); |
| } |
| |
| return 0; |
| } |
| |
| static void cs_hsi_read_on_data_complete(struct hsi_msg *msg) |
| { |
| struct cs_hsi_iface *hi = msg->context; |
| u32 payload; |
| |
| if (unlikely(msg->status == HSI_STATUS_ERROR)) { |
| cs_hsi_data_read_error(hi, msg); |
| return; |
| } |
| |
| spin_lock(&hi->lock); |
| WARN_ON(!(hi->data_state & SSI_CHANNEL_STATE_READING)); |
| hi->data_state &= ~SSI_CHANNEL_STATE_READING; |
| payload = CS_RX_DATA_RECEIVED; |
| payload |= hi->rx_slot; |
| hi->rx_slot++; |
| hi->rx_slot %= hi->rx_ptr_boundary; |
| /* expose current rx ptr in mmap area */ |
| hi->mmap_cfg->rx_ptr = hi->rx_slot; |
| if (unlikely(waitqueue_active(&hi->datawait))) |
| wake_up_interruptible(&hi->datawait); |
| spin_unlock(&hi->lock); |
| |
| cs_notify_data(payload, hi->rx_bufs); |
| cs_hsi_read_on_data(hi); |
| } |
| |
| static void cs_hsi_peek_on_data_complete(struct hsi_msg *msg) |
| { |
| struct cs_hsi_iface *hi = msg->context; |
| u32 *address; |
| int ret; |
| |
| if (unlikely(msg->status == HSI_STATUS_ERROR)) { |
| cs_hsi_data_read_error(hi, msg); |
| return; |
| } |
| if (unlikely(hi->iface_state != CS_STATE_CONFIGURED)) { |
| dev_err(&hi->cl->device, "Data received in invalid state\n"); |
| cs_hsi_data_read_error(hi, msg); |
| return; |
| } |
| |
| spin_lock(&hi->lock); |
| WARN_ON(!(hi->data_state & SSI_CHANNEL_STATE_POLL)); |
| hi->data_state &= ~SSI_CHANNEL_STATE_POLL; |
| hi->data_state |= SSI_CHANNEL_STATE_READING; |
| spin_unlock(&hi->lock); |
| |
| address = (u32 *)(hi->mmap_base + |
| hi->rx_offsets[hi->rx_slot % hi->rx_bufs]); |
| sg_init_one(msg->sgt.sgl, address, hi->buf_size); |
| msg->sgt.nents = 1; |
| msg->complete = cs_hsi_read_on_data_complete; |
| ret = hsi_async_read(hi->cl, msg); |
| if (ret) |
| cs_hsi_data_read_error(hi, msg); |
| } |
| |
| /* |
| * Read/write transaction is ongoing. Returns false if in |
| * SSI_CHANNEL_STATE_POLL state. |
| */ |
| static inline int cs_state_xfer_active(unsigned int state) |
| { |
| return (state & SSI_CHANNEL_STATE_WRITING) || |
| (state & SSI_CHANNEL_STATE_READING); |
| } |
| |
| /* |
| * No pending read/writes |
| */ |
| static inline int cs_state_idle(unsigned int state) |
| { |
| return !(state & ~SSI_CHANNEL_STATE_ERROR); |
| } |
| |
| static void cs_hsi_read_on_data(struct cs_hsi_iface *hi) |
| { |
| struct hsi_msg *rxmsg; |
| int ret; |
| |
| spin_lock(&hi->lock); |
| if (hi->data_state & |
| (SSI_CHANNEL_STATE_READING | SSI_CHANNEL_STATE_POLL)) { |
| dev_dbg(&hi->cl->device, "Data read already pending (%u)\n", |
| hi->data_state); |
| spin_unlock(&hi->lock); |
| return; |
| } |
| hi->data_state |= SSI_CHANNEL_STATE_POLL; |
| spin_unlock(&hi->lock); |
| |
| rxmsg = hi->data_rx_msg; |
| sg_init_one(rxmsg->sgt.sgl, (void *)hi->mmap_base, 0); |
| rxmsg->sgt.nents = 0; |
| rxmsg->complete = cs_hsi_peek_on_data_complete; |
| |
| ret = hsi_async_read(hi->cl, rxmsg); |
| if (ret) |
| cs_hsi_data_read_error(hi, rxmsg); |
| } |
| |
| static void cs_hsi_write_on_data_complete(struct hsi_msg *msg) |
| { |
| struct cs_hsi_iface *hi = msg->context; |
| |
| if (msg->status == HSI_STATUS_COMPLETED) { |
| spin_lock(&hi->lock); |
| hi->data_state &= ~SSI_CHANNEL_STATE_WRITING; |
| if (unlikely(waitqueue_active(&hi->datawait))) |
| wake_up_interruptible(&hi->datawait); |
| spin_unlock(&hi->lock); |
| } else { |
| cs_hsi_data_write_error(hi, msg); |
| } |
| } |
| |
| static int cs_hsi_write_on_data(struct cs_hsi_iface *hi, unsigned int slot) |
| { |
| u32 *address; |
| struct hsi_msg *txmsg; |
| int ret; |
| |
| spin_lock(&hi->lock); |
| if (hi->iface_state != CS_STATE_CONFIGURED) { |
| dev_err(&hi->cl->device, "Not configured, aborting\n"); |
| ret = -EINVAL; |
| goto error; |
| } |
| if (hi->data_state & SSI_CHANNEL_STATE_ERROR) { |
| dev_err(&hi->cl->device, "HSI error, aborting\n"); |
| ret = -EIO; |
| goto error; |
| } |
| if (hi->data_state & SSI_CHANNEL_STATE_WRITING) { |
| dev_err(&hi->cl->device, "Write pending on data channel.\n"); |
| ret = -EBUSY; |
| goto error; |
| } |
| hi->data_state |= SSI_CHANNEL_STATE_WRITING; |
| spin_unlock(&hi->lock); |
| |
| hi->tx_slot = slot; |
| address = (u32 *)(hi->mmap_base + hi->tx_offsets[hi->tx_slot]); |
| txmsg = hi->data_tx_msg; |
| sg_init_one(txmsg->sgt.sgl, address, hi->buf_size); |
| txmsg->complete = cs_hsi_write_on_data_complete; |
| ret = hsi_async_write(hi->cl, txmsg); |
| if (ret) |
| cs_hsi_data_write_error(hi, txmsg); |
| |
| return ret; |
| |
| error: |
| spin_unlock(&hi->lock); |
| if (ret == -EIO) |
| cs_hsi_data_write_error(hi, hi->data_tx_msg); |
| |
| return ret; |
| } |
| |
| static unsigned int cs_hsi_get_state(struct cs_hsi_iface *hi) |
| { |
| return hi->iface_state; |
| } |
| |
| static int cs_hsi_command(struct cs_hsi_iface *hi, u32 cmd) |
| { |
| int ret = 0; |
| |
| local_bh_disable(); |
| switch (cmd & TARGET_MASK) { |
| case TARGET_REMOTE: |
| ret = cs_hsi_write_on_control(hi, cmd); |
| break; |
| case TARGET_LOCAL: |
| if ((cmd & CS_CMD_MASK) == CS_TX_DATA_READY) |
| ret = cs_hsi_write_on_data(hi, cmd & CS_PARAM_MASK); |
| else |
| ret = -EINVAL; |
| break; |
| default: |
| ret = -EINVAL; |
| break; |
| } |
| local_bh_enable(); |
| |
| return ret; |
| } |
| |
| static void cs_hsi_set_wakeline(struct cs_hsi_iface *hi, bool new_state) |
| { |
| int change = 0; |
| |
| spin_lock_bh(&hi->lock); |
| if (hi->wakeline_state != new_state) { |
| hi->wakeline_state = new_state; |
| change = 1; |
| dev_dbg(&hi->cl->device, "setting wake line to %d (%p)\n", |
| new_state, hi->cl); |
| } |
| spin_unlock_bh(&hi->lock); |
| |
| if (change) { |
| if (new_state) |
| ssip_slave_start_tx(hi->master); |
| else |
| ssip_slave_stop_tx(hi->master); |
| } |
| |
| dev_dbg(&hi->cl->device, "wake line set to %d (%p)\n", |
| new_state, hi->cl); |
| } |
| |
| static void set_buffer_sizes(struct cs_hsi_iface *hi, int rx_bufs, int tx_bufs) |
| { |
| hi->rx_bufs = rx_bufs; |
| hi->tx_bufs = tx_bufs; |
| hi->mmap_cfg->rx_bufs = rx_bufs; |
| hi->mmap_cfg->tx_bufs = tx_bufs; |
| |
| if (hi->flags & CS_FEAT_ROLLING_RX_COUNTER) { |
| /* |
| * For more robust overrun detection, let the rx |
| * pointer run in range 0..'boundary-1'. Boundary |
| * is a multiple of rx_bufs, and limited in max size |
| * by RX_PTR_MAX_SHIFT to allow for fast ptr-diff |
| * calculation. |
| */ |
| hi->rx_ptr_boundary = (rx_bufs << RX_PTR_BOUNDARY_SHIFT); |
| hi->mmap_cfg->rx_ptr_boundary = hi->rx_ptr_boundary; |
| } else { |
| hi->rx_ptr_boundary = hi->rx_bufs; |
| } |
| } |
| |
| static int check_buf_params(struct cs_hsi_iface *hi, |
| const struct cs_buffer_config *buf_cfg) |
| { |
| size_t buf_size_aligned = L1_CACHE_ALIGN(buf_cfg->buf_size) * |
| (buf_cfg->rx_bufs + buf_cfg->tx_bufs); |
| size_t ctrl_size_aligned = L1_CACHE_ALIGN(sizeof(*hi->mmap_cfg)); |
| int r = 0; |
| |
| if (buf_cfg->rx_bufs > CS_MAX_BUFFERS || |
| buf_cfg->tx_bufs > CS_MAX_BUFFERS) { |
| r = -EINVAL; |
| } else if ((buf_size_aligned + ctrl_size_aligned) >= hi->mmap_size) { |
| dev_err(&hi->cl->device, "No space for the requested buffer " |
| "configuration\n"); |
| r = -ENOBUFS; |
| } |
| |
| return r; |
| } |
| |
| /** |
| * Block until pending data transfers have completed. |
| */ |
| static int cs_hsi_data_sync(struct cs_hsi_iface *hi) |
| { |
| int r = 0; |
| |
| spin_lock_bh(&hi->lock); |
| |
| if (!cs_state_xfer_active(hi->data_state)) { |
| dev_dbg(&hi->cl->device, "hsi_data_sync break, idle\n"); |
| goto out; |
| } |
| |
| for (;;) { |
| int s; |
| DEFINE_WAIT(wait); |
| if (!cs_state_xfer_active(hi->data_state)) |
| goto out; |
| if (signal_pending(current)) { |
| r = -ERESTARTSYS; |
| goto out; |
| } |
| /** |
| * prepare_to_wait must be called with hi->lock held |
| * so that callbacks can check for waitqueue_active() |
| */ |
| prepare_to_wait(&hi->datawait, &wait, TASK_INTERRUPTIBLE); |
| spin_unlock_bh(&hi->lock); |
| s = schedule_timeout( |
| msecs_to_jiffies(CS_HSI_TRANSFER_TIMEOUT_MS)); |
| spin_lock_bh(&hi->lock); |
| finish_wait(&hi->datawait, &wait); |
| if (!s) { |
| dev_dbg(&hi->cl->device, |
| "hsi_data_sync timeout after %d ms\n", |
| CS_HSI_TRANSFER_TIMEOUT_MS); |
| r = -EIO; |
| goto out; |
| } |
| } |
| |
| out: |
| spin_unlock_bh(&hi->lock); |
| dev_dbg(&hi->cl->device, "hsi_data_sync done with res %d\n", r); |
| |
| return r; |
| } |
| |
| static void cs_hsi_data_enable(struct cs_hsi_iface *hi, |
| struct cs_buffer_config *buf_cfg) |
| { |
| unsigned int data_start, i; |
| |
| BUG_ON(hi->buf_size == 0); |
| |
| set_buffer_sizes(hi, buf_cfg->rx_bufs, buf_cfg->tx_bufs); |
| |
| hi->slot_size = L1_CACHE_ALIGN(hi->buf_size); |
| dev_dbg(&hi->cl->device, |
| "setting slot size to %u, buf size %u, align %u\n", |
| hi->slot_size, hi->buf_size, L1_CACHE_BYTES); |
| |
| data_start = L1_CACHE_ALIGN(sizeof(*hi->mmap_cfg)); |
| dev_dbg(&hi->cl->device, |
| "setting data start at %u, cfg block %u, align %u\n", |
| data_start, sizeof(*hi->mmap_cfg), L1_CACHE_BYTES); |
| |
| for (i = 0; i < hi->mmap_cfg->rx_bufs; i++) { |
| hi->rx_offsets[i] = data_start + i * hi->slot_size; |
| hi->mmap_cfg->rx_offsets[i] = hi->rx_offsets[i]; |
| dev_dbg(&hi->cl->device, "DL buf #%u at %u\n", |
| i, hi->rx_offsets[i]); |
| } |
| for (i = 0; i < hi->mmap_cfg->tx_bufs; i++) { |
| hi->tx_offsets[i] = data_start + |
| (i + hi->mmap_cfg->rx_bufs) * hi->slot_size; |
| hi->mmap_cfg->tx_offsets[i] = hi->tx_offsets[i]; |
| dev_dbg(&hi->cl->device, "UL buf #%u at %u\n", |
| i, hi->rx_offsets[i]); |
| } |
| |
| hi->iface_state = CS_STATE_CONFIGURED; |
| } |
| |
| static void cs_hsi_data_disable(struct cs_hsi_iface *hi, int old_state) |
| { |
| if (old_state == CS_STATE_CONFIGURED) { |
| dev_dbg(&hi->cl->device, |
| "closing data channel with slot size 0\n"); |
| hi->iface_state = CS_STATE_OPENED; |
| } |
| } |
| |
| static int cs_hsi_buf_config(struct cs_hsi_iface *hi, |
| struct cs_buffer_config *buf_cfg) |
| { |
| int r = 0; |
| unsigned int old_state = hi->iface_state; |
| |
| spin_lock_bh(&hi->lock); |
| /* Prevent new transactions during buffer reconfig */ |
| if (old_state == CS_STATE_CONFIGURED) |
| hi->iface_state = CS_STATE_OPENED; |
| spin_unlock_bh(&hi->lock); |
| |
| /* |
| * make sure that no non-zero data reads are ongoing before |
| * proceeding to change the buffer layout |
| */ |
| r = cs_hsi_data_sync(hi); |
| if (r < 0) |
| return r; |
| |
| WARN_ON(cs_state_xfer_active(hi->data_state)); |
| |
| spin_lock_bh(&hi->lock); |
| r = check_buf_params(hi, buf_cfg); |
| if (r < 0) |
| goto error; |
| |
| hi->buf_size = buf_cfg->buf_size; |
| hi->mmap_cfg->buf_size = hi->buf_size; |
| hi->flags = buf_cfg->flags; |
| |
| hi->rx_slot = 0; |
| hi->tx_slot = 0; |
| hi->slot_size = 0; |
| |
| if (hi->buf_size) |
| cs_hsi_data_enable(hi, buf_cfg); |
| else |
| cs_hsi_data_disable(hi, old_state); |
| |
| spin_unlock_bh(&hi->lock); |
| |
| if (old_state != hi->iface_state) { |
| if (hi->iface_state == CS_STATE_CONFIGURED) { |
| pm_qos_add_request(&hi->pm_qos_req, |
| PM_QOS_CPU_DMA_LATENCY, |
| CS_QOS_LATENCY_FOR_DATA_USEC); |
| local_bh_disable(); |
| cs_hsi_read_on_data(hi); |
| local_bh_enable(); |
| } else if (old_state == CS_STATE_CONFIGURED) { |
| pm_qos_remove_request(&hi->pm_qos_req); |
| } |
| } |
| return r; |
| |
| error: |
| spin_unlock_bh(&hi->lock); |
| return r; |
| } |
| |
| static int cs_hsi_start(struct cs_hsi_iface **hi, struct hsi_client *cl, |
| unsigned long mmap_base, unsigned long mmap_size) |
| { |
| int err = 0; |
| struct cs_hsi_iface *hsi_if = kzalloc(sizeof(*hsi_if), GFP_KERNEL); |
| |
| dev_dbg(&cl->device, "cs_hsi_start\n"); |
| |
| if (!hsi_if) { |
| err = -ENOMEM; |
| goto leave0; |
| } |
| spin_lock_init(&hsi_if->lock); |
| hsi_if->cl = cl; |
| hsi_if->iface_state = CS_STATE_CLOSED; |
| hsi_if->mmap_cfg = (struct cs_mmap_config_block *)mmap_base; |
| hsi_if->mmap_base = mmap_base; |
| hsi_if->mmap_size = mmap_size; |
| memset(hsi_if->mmap_cfg, 0, sizeof(*hsi_if->mmap_cfg)); |
| init_waitqueue_head(&hsi_if->datawait); |
| err = cs_alloc_cmds(hsi_if); |
| if (err < 0) { |
| dev_err(&cl->device, "Unable to alloc HSI messages\n"); |
| goto leave1; |
| } |
| err = cs_hsi_alloc_data(hsi_if); |
| if (err < 0) { |
| dev_err(&cl->device, "Unable to alloc HSI messages for data\n"); |
| goto leave2; |
| } |
| err = hsi_claim_port(cl, 1); |
| if (err < 0) { |
| dev_err(&cl->device, |
| "Could not open, HSI port already claimed\n"); |
| goto leave3; |
| } |
| hsi_if->master = ssip_slave_get_master(cl); |
| if (IS_ERR(hsi_if->master)) { |
| err = PTR_ERR(hsi_if->master); |
| dev_err(&cl->device, "Could not get HSI master client\n"); |
| goto leave4; |
| } |
| if (!ssip_slave_running(hsi_if->master)) { |
| err = -ENODEV; |
| dev_err(&cl->device, |
| "HSI port not initialized\n"); |
| goto leave4; |
| } |
| |
| hsi_if->iface_state = CS_STATE_OPENED; |
| local_bh_disable(); |
| cs_hsi_read_on_control(hsi_if); |
| local_bh_enable(); |
| |
| dev_dbg(&cl->device, "cs_hsi_start...done\n"); |
| |
| BUG_ON(!hi); |
| *hi = hsi_if; |
| |
| return 0; |
| |
| leave4: |
| hsi_release_port(cl); |
| leave3: |
| cs_hsi_free_data(hsi_if); |
| leave2: |
| cs_free_cmds(hsi_if); |
| leave1: |
| kfree(hsi_if); |
| leave0: |
| dev_dbg(&cl->device, "cs_hsi_start...done/error\n\n"); |
| |
| return err; |
| } |
| |
| static void cs_hsi_stop(struct cs_hsi_iface *hi) |
| { |
| dev_dbg(&hi->cl->device, "cs_hsi_stop\n"); |
| cs_hsi_set_wakeline(hi, 0); |
| ssip_slave_put_master(hi->master); |
| |
| /* hsi_release_port() needs to be called with CS_STATE_CLOSED */ |
| hi->iface_state = CS_STATE_CLOSED; |
| hsi_release_port(hi->cl); |
| |
| /* |
| * hsi_release_port() should flush out all the pending |
| * messages, so cs_state_idle() should be true for both |
| * control and data channels. |
| */ |
| WARN_ON(!cs_state_idle(hi->control_state)); |
| WARN_ON(!cs_state_idle(hi->data_state)); |
| |
| if (pm_qos_request_active(&hi->pm_qos_req)) |
| pm_qos_remove_request(&hi->pm_qos_req); |
| |
| spin_lock_bh(&hi->lock); |
| cs_hsi_free_data(hi); |
| cs_free_cmds(hi); |
| spin_unlock_bh(&hi->lock); |
| kfree(hi); |
| } |
| |
| static int cs_char_vma_fault(struct vm_area_struct *vma, struct vm_fault *vmf) |
| { |
| struct cs_char *csdata = vma->vm_private_data; |
| struct page *page; |
| |
| page = virt_to_page(csdata->mmap_base); |
| get_page(page); |
| vmf->page = page; |
| |
| return 0; |
| } |
| |
| static const struct vm_operations_struct cs_char_vm_ops = { |
| .fault = cs_char_vma_fault, |
| }; |
| |
| static int cs_char_fasync(int fd, struct file *file, int on) |
| { |
| struct cs_char *csdata = file->private_data; |
| |
| if (fasync_helper(fd, file, on, &csdata->async_queue) < 0) |
| return -EIO; |
| |
| return 0; |
| } |
| |
| static unsigned int cs_char_poll(struct file *file, poll_table *wait) |
| { |
| struct cs_char *csdata = file->private_data; |
| unsigned int ret = 0; |
| |
| poll_wait(file, &cs_char_data.wait, wait); |
| spin_lock_bh(&csdata->lock); |
| if (!list_empty(&csdata->chardev_queue)) |
| ret = POLLIN | POLLRDNORM; |
| else if (!list_empty(&csdata->dataind_queue)) |
| ret = POLLIN | POLLRDNORM; |
| spin_unlock_bh(&csdata->lock); |
| |
| return ret; |
| } |
| |
| static ssize_t cs_char_read(struct file *file, char __user *buf, size_t count, |
| loff_t *unused) |
| { |
| struct cs_char *csdata = file->private_data; |
| u32 data; |
| ssize_t retval; |
| |
| if (count < sizeof(data)) |
| return -EINVAL; |
| |
| for (;;) { |
| DEFINE_WAIT(wait); |
| |
| spin_lock_bh(&csdata->lock); |
| if (!list_empty(&csdata->chardev_queue)) { |
| data = cs_pop_entry(&csdata->chardev_queue); |
| } else if (!list_empty(&csdata->dataind_queue)) { |
| data = cs_pop_entry(&csdata->dataind_queue); |
| csdata->dataind_pending--; |
| } else { |
| data = 0; |
| } |
| spin_unlock_bh(&csdata->lock); |
| |
| if (data) |
| break; |
| if (file->f_flags & O_NONBLOCK) { |
| retval = -EAGAIN; |
| goto out; |
| } else if (signal_pending(current)) { |
| retval = -ERESTARTSYS; |
| goto out; |
| } |
| prepare_to_wait_exclusive(&csdata->wait, &wait, |
| TASK_INTERRUPTIBLE); |
| schedule(); |
| finish_wait(&csdata->wait, &wait); |
| } |
| |
| retval = put_user(data, (u32 __user *)buf); |
| if (!retval) |
| retval = sizeof(data); |
| |
| out: |
| return retval; |
| } |
| |
| static ssize_t cs_char_write(struct file *file, const char __user *buf, |
| size_t count, loff_t *unused) |
| { |
| struct cs_char *csdata = file->private_data; |
| u32 data; |
| int err; |
| ssize_t retval; |
| |
| if (count < sizeof(data)) |
| return -EINVAL; |
| |
| if (get_user(data, (u32 __user *)buf)) |
| retval = -EFAULT; |
| else |
| retval = count; |
| |
| err = cs_hsi_command(csdata->hi, data); |
| if (err < 0) |
| retval = err; |
| |
| return retval; |
| } |
| |
| static long cs_char_ioctl(struct file *file, unsigned int cmd, |
| unsigned long arg) |
| { |
| struct cs_char *csdata = file->private_data; |
| int r = 0; |
| |
| switch (cmd) { |
| case CS_GET_STATE: { |
| unsigned int state; |
| |
| state = cs_hsi_get_state(csdata->hi); |
| if (copy_to_user((void __user *)arg, &state, sizeof(state))) |
| r = -EFAULT; |
| |
| break; |
| } |
| case CS_SET_WAKELINE: { |
| unsigned int state; |
| |
| if (copy_from_user(&state, (void __user *)arg, sizeof(state))) { |
| r = -EFAULT; |
| break; |
| } |
| |
| if (state > 1) { |
| r = -EINVAL; |
| break; |
| } |
| |
| cs_hsi_set_wakeline(csdata->hi, !!state); |
| |
| break; |
| } |
| case CS_GET_IF_VERSION: { |
| unsigned int ifver = CS_IF_VERSION; |
| |
| if (copy_to_user((void __user *)arg, &ifver, sizeof(ifver))) |
| r = -EFAULT; |
| |
| break; |
| } |
| case CS_CONFIG_BUFS: { |
| struct cs_buffer_config buf_cfg; |
| |
| if (copy_from_user(&buf_cfg, (void __user *)arg, |
| sizeof(buf_cfg))) |
| r = -EFAULT; |
| else |
| r = cs_hsi_buf_config(csdata->hi, &buf_cfg); |
| |
| break; |
| } |
| default: |
| r = -ENOTTY; |
| break; |
| } |
| |
| return r; |
| } |
| |
| static int cs_char_mmap(struct file *file, struct vm_area_struct *vma) |
| { |
| if (vma->vm_end < vma->vm_start) |
| return -EINVAL; |
| |
| if (((vma->vm_end - vma->vm_start) >> PAGE_SHIFT) != 1) |
| return -EINVAL; |
| |
| vma->vm_flags |= VM_IO | VM_DONTDUMP | VM_DONTEXPAND; |
| vma->vm_ops = &cs_char_vm_ops; |
| vma->vm_private_data = file->private_data; |
| |
| return 0; |
| } |
| |
| static int cs_char_open(struct inode *unused, struct file *file) |
| { |
| int ret = 0; |
| unsigned long p; |
| |
| spin_lock_bh(&cs_char_data.lock); |
| if (cs_char_data.opened) { |
| ret = -EBUSY; |
| spin_unlock_bh(&cs_char_data.lock); |
| goto out1; |
| } |
| cs_char_data.opened = 1; |
| cs_char_data.dataind_pending = 0; |
| spin_unlock_bh(&cs_char_data.lock); |
| |
| p = get_zeroed_page(GFP_KERNEL); |
| if (!p) { |
| ret = -ENOMEM; |
| goto out2; |
| } |
| |
| ret = cs_hsi_start(&cs_char_data.hi, cs_char_data.cl, p, CS_MMAP_SIZE); |
| if (ret) { |
| dev_err(&cs_char_data.cl->device, "Unable to initialize HSI\n"); |
| goto out3; |
| } |
| |
| /* these are only used in release so lock not needed */ |
| cs_char_data.mmap_base = p; |
| cs_char_data.mmap_size = CS_MMAP_SIZE; |
| |
| file->private_data = &cs_char_data; |
| |
| return 0; |
| |
| out3: |
| free_page(p); |
| out2: |
| spin_lock_bh(&cs_char_data.lock); |
| cs_char_data.opened = 0; |
| spin_unlock_bh(&cs_char_data.lock); |
| out1: |
| return ret; |
| } |
| |
| static void cs_free_char_queue(struct list_head *head) |
| { |
| struct char_queue *entry; |
| struct list_head *cursor, *next; |
| |
| if (!list_empty(head)) { |
| list_for_each_safe(cursor, next, head) { |
| entry = list_entry(cursor, struct char_queue, list); |
| list_del(&entry->list); |
| kfree(entry); |
| } |
| } |
| |
| } |
| |
| static int cs_char_release(struct inode *unused, struct file *file) |
| { |
| struct cs_char *csdata = file->private_data; |
| |
| cs_hsi_stop(csdata->hi); |
| spin_lock_bh(&csdata->lock); |
| csdata->hi = NULL; |
| free_page(csdata->mmap_base); |
| cs_free_char_queue(&csdata->chardev_queue); |
| cs_free_char_queue(&csdata->dataind_queue); |
| csdata->opened = 0; |
| spin_unlock_bh(&csdata->lock); |
| |
| return 0; |
| } |
| |
| static const struct file_operations cs_char_fops = { |
| .owner = THIS_MODULE, |
| .read = cs_char_read, |
| .write = cs_char_write, |
| .poll = cs_char_poll, |
| .unlocked_ioctl = cs_char_ioctl, |
| .mmap = cs_char_mmap, |
| .open = cs_char_open, |
| .release = cs_char_release, |
| .fasync = cs_char_fasync, |
| }; |
| |
| static struct miscdevice cs_char_miscdev = { |
| .minor = MISC_DYNAMIC_MINOR, |
| .name = "cmt_speech", |
| .fops = &cs_char_fops |
| }; |
| |
| static int cs_hsi_client_probe(struct device *dev) |
| { |
| int err = 0; |
| struct hsi_client *cl = to_hsi_client(dev); |
| |
| dev_dbg(dev, "hsi_client_probe\n"); |
| init_waitqueue_head(&cs_char_data.wait); |
| spin_lock_init(&cs_char_data.lock); |
| cs_char_data.opened = 0; |
| cs_char_data.cl = cl; |
| cs_char_data.hi = NULL; |
| INIT_LIST_HEAD(&cs_char_data.chardev_queue); |
| INIT_LIST_HEAD(&cs_char_data.dataind_queue); |
| |
| cs_char_data.channel_id_cmd = hsi_get_channel_id_by_name(cl, |
| "speech-control"); |
| if (cs_char_data.channel_id_cmd < 0) { |
| err = cs_char_data.channel_id_cmd; |
| dev_err(dev, "Could not get cmd channel (%d)\n", err); |
| return err; |
| } |
| |
| cs_char_data.channel_id_data = hsi_get_channel_id_by_name(cl, |
| "speech-data"); |
| if (cs_char_data.channel_id_data < 0) { |
| err = cs_char_data.channel_id_data; |
| dev_err(dev, "Could not get data channel (%d)\n", err); |
| return err; |
| } |
| |
| err = misc_register(&cs_char_miscdev); |
| if (err) |
| dev_err(dev, "Failed to register: %d\n", err); |
| |
| return err; |
| } |
| |
| static int cs_hsi_client_remove(struct device *dev) |
| { |
| struct cs_hsi_iface *hi; |
| |
| dev_dbg(dev, "hsi_client_remove\n"); |
| misc_deregister(&cs_char_miscdev); |
| spin_lock_bh(&cs_char_data.lock); |
| hi = cs_char_data.hi; |
| cs_char_data.hi = NULL; |
| spin_unlock_bh(&cs_char_data.lock); |
| if (hi) |
| cs_hsi_stop(hi); |
| |
| return 0; |
| } |
| |
| static struct hsi_client_driver cs_hsi_driver = { |
| .driver = { |
| .name = "cmt-speech", |
| .owner = THIS_MODULE, |
| .probe = cs_hsi_client_probe, |
| .remove = cs_hsi_client_remove, |
| }, |
| }; |
| |
| static int __init cs_char_init(void) |
| { |
| pr_info("CMT speech driver added\n"); |
| return hsi_register_client_driver(&cs_hsi_driver); |
| } |
| module_init(cs_char_init); |
| |
| static void __exit cs_char_exit(void) |
| { |
| hsi_unregister_client_driver(&cs_hsi_driver); |
| pr_info("CMT speech driver removed\n"); |
| } |
| module_exit(cs_char_exit); |
| |
| MODULE_ALIAS("hsi:cmt-speech"); |
| MODULE_AUTHOR("Kai Vehmanen <kai.vehmanen@nokia.com>"); |
| MODULE_AUTHOR("Peter Ujfalusi <peter.ujfalusi@nokia.com>"); |
| MODULE_DESCRIPTION("CMT speech driver"); |
| MODULE_LICENSE("GPL v2"); |