diff options
Diffstat (limited to 'gxp-mailbox-driver.c')
-rw-r--r-- | gxp-mailbox-driver.c | 511 |
1 files changed, 511 insertions, 0 deletions
diff --git a/gxp-mailbox-driver.c b/gxp-mailbox-driver.c new file mode 100644 index 0000000..40fdba1 --- /dev/null +++ b/gxp-mailbox-driver.c @@ -0,0 +1,511 @@ +// SPDX-License-Identifier: GPL-2.0 +/* + * GXP hardware-based mailbox driver implementation. + * + * Copyright (C) 2021 Google LLC + */ + +#include <asm/barrier.h> +#include <linux/bitops.h> +#include <linux/interrupt.h> +#include <linux/kthread.h> +#include <linux/of_irq.h> +#include <linux/spinlock.h> + +#include "gxp-config.h" /* GXP_USE_LEGACY_MAILBOX */ +#include "gxp-mailbox-driver.h" +#include "gxp-mailbox-regs.h" +#include "gxp-mailbox.h" + +static u32 data_read(struct gxp_mailbox *mailbox, uint reg_offset) +{ + return readl(mailbox->data_reg_base + reg_offset); +} + +static void data_write(struct gxp_mailbox *mailbox, uint reg_offset, u32 value) +{ + writel(value, mailbox->data_reg_base + reg_offset); +} + +/* IRQ Handling */ + +/* Interrupt to signal a response from the device to host */ +#define MBOX_DEVICE_TO_HOST_RESPONSE_IRQ_MASK BIT(0) + +static irqreturn_t mailbox_irq_handler(int irq, void *arg) +{ + u32 masked_status; + struct gxp_mailbox *mailbox = (struct gxp_mailbox *)arg; + struct work_struct **handlers = mailbox->interrupt_handlers; + u32 next_int; + + /* Contains only the non-masked, pending interrupt bits */ + masked_status = gxp_mailbox_get_host_mask_status(mailbox); + + /* Clear all pending IRQ bits */ + gxp_mailbox_clear_host_interrupt(mailbox, masked_status); + + if (masked_status & MBOX_DEVICE_TO_HOST_RESPONSE_IRQ_MASK) { + mailbox->handle_irq(mailbox); + masked_status &= ~MBOX_DEVICE_TO_HOST_RESPONSE_IRQ_MASK; + } + + while ((next_int = ffs(masked_status))) { + next_int--; /* ffs returns 1-based indices */ + masked_status &= ~BIT(next_int); + + if (handlers[next_int]) + schedule_work(handlers[next_int]); + else + pr_err_ratelimited( + "mailbox%d: received unknown interrupt bit 0x%X\n", + mailbox->core_id, next_int); + } + + return IRQ_HANDLED; +} + +static void register_irq(struct gxp_mailbox *mailbox) +{ + int err; + unsigned int virq; + + virq = irq_of_parse_and_map(mailbox->gxp->dev->of_node, + mailbox->core_id); + if (!virq) { + pr_err("Unable to parse interrupt for core %d from the DT\n", + mailbox->core_id); + return; + } + + err = request_irq(virq, mailbox_irq_handler, /*flags=*/0, + "aurora_mbx_irq", (void *)mailbox); + if (err) { + pr_err("Unable to register IRQ num=%d; error=%d\n", virq, err); + return; + } + + mailbox->interrupt_virq = virq; + pr_debug("Core %d's mailbox interrupt registered as IRQ %u.\n", + mailbox->core_id, virq); +} + +static void unregister_irq(struct gxp_mailbox *mailbox) +{ + if (mailbox->interrupt_virq) { + pr_debug("Freeing IRQ %d\n", mailbox->interrupt_virq); + free_irq(mailbox->interrupt_virq, mailbox); + mailbox->interrupt_virq = 0; + } +} + +/* gxp-mailbox-driver.h interface */ + +u32 gxp_circ_queue_cnt(u32 head, u32 tail, u32 queue_size, u32 wrap_bit) +{ + if (CIRCULAR_QUEUE_WRAPPED(tail, wrap_bit) != + CIRCULAR_QUEUE_WRAPPED(head, wrap_bit)) + return queue_size - CIRCULAR_QUEUE_REAL_INDEX(head, wrap_bit) + + CIRCULAR_QUEUE_REAL_INDEX(tail, wrap_bit); + else + return tail - head; +} + +u32 gxp_circ_queue_inc(u32 index, u32 inc, u32 queue_size, u32 wrap_bit) +{ + u32 new_index = CIRCULAR_QUEUE_REAL_INDEX(index, wrap_bit) + inc; + + if (new_index >= queue_size) + return (index + inc - queue_size) ^ wrap_bit; + else + return index + inc; +} + +void gxp_mailbox_driver_init(struct gxp_mailbox *mailbox) +{ + spin_lock_init(&mailbox->cmd_tail_resp_head_lock); + spin_lock_init(&mailbox->cmd_head_resp_tail_lock); +} + +void gxp_mailbox_driver_exit(struct gxp_mailbox *mailbox) +{ + /* Nothing to cleanup */ +} + +void gxp_mailbox_driver_enable_interrupts(struct gxp_mailbox *mailbox) +{ + register_irq(mailbox); +} + +void gxp_mailbox_driver_disable_interrupts(struct gxp_mailbox *mailbox) +{ + unregister_irq(mailbox); +} + +void __iomem *gxp_mailbox_get_csr_base(struct gxp_dev *gxp, uint index) +{ + return gxp->mbx[index].vaddr; +} + +void __iomem *gxp_mailbox_get_data_base(struct gxp_dev *gxp, uint index) +{ + return gxp->mbx[index].vaddr + MBOX_DATA_REG_BASE; +} + +/* gxp-mailbox-driver.h: Data register-based calls */ + +void gxp_mailbox_write_status(struct gxp_mailbox *mailbox, u32 status) +{ + data_write(mailbox, MBOX_DATA_STATUS_OFFSET, status); +} + +void gxp_mailbox_write_descriptor(struct gxp_mailbox *mailbox, + dma_addr_t descriptor_addr) +{ + data_write(mailbox, MBOX_DATA_DESCRIPTOR_ADDR_OFFSET, (u32)descriptor_addr); +} + +void gxp_mailbox_write_cmd_queue_tail(struct gxp_mailbox *mailbox, u16 val) +{ + u32 current_resp_head; + u32 new_cmd_tail; + unsigned long flags; + + spin_lock_irqsave(&mailbox->cmd_tail_resp_head_lock, flags); + + current_resp_head = data_read(mailbox, MBOX_DATA_CMD_TAIL_RESP_HEAD_OFFSET) & + RESP_HEAD_MASK; + new_cmd_tail = (u32)val << CMD_TAIL_SHIFT; + data_write(mailbox, MBOX_DATA_CMD_TAIL_RESP_HEAD_OFFSET, + new_cmd_tail | current_resp_head); + + spin_unlock_irqrestore(&mailbox->cmd_tail_resp_head_lock, flags); +} + +void gxp_mailbox_write_resp_queue_head(struct gxp_mailbox *mailbox, u16 val) +{ + u32 current_cmd_tail; + u32 new_resp_head; + unsigned long flags; + + spin_lock_irqsave(&mailbox->cmd_tail_resp_head_lock, flags); + + current_cmd_tail = data_read(mailbox, MBOX_DATA_CMD_TAIL_RESP_HEAD_OFFSET) & + CMD_TAIL_MASK; + new_resp_head = (u32)val << RESP_HEAD_SHIFT; + data_write(mailbox, MBOX_DATA_CMD_TAIL_RESP_HEAD_OFFSET, + current_cmd_tail | new_resp_head); + + spin_unlock_irqrestore(&mailbox->cmd_tail_resp_head_lock, flags); +} + +u16 gxp_mailbox_read_cmd_queue_head(struct gxp_mailbox *mailbox) +{ + u32 reg_val; + unsigned long flags; + + spin_lock_irqsave(&mailbox->cmd_head_resp_tail_lock, flags); + + reg_val = data_read(mailbox, MBOX_DATA_CMD_HEAD_RESP_TAIL_OFFSET); + + spin_unlock_irqrestore(&mailbox->cmd_head_resp_tail_lock, flags); + + return (u16)((reg_val & CMD_HEAD_MASK) >> CMD_HEAD_SHIFT); +} + +u16 gxp_mailbox_read_resp_queue_tail(struct gxp_mailbox *mailbox) +{ + u32 reg_val; + unsigned long flags; + + spin_lock_irqsave(&mailbox->cmd_head_resp_tail_lock, flags); + + reg_val = data_read(mailbox, MBOX_DATA_CMD_HEAD_RESP_TAIL_OFFSET); + + spin_unlock_irqrestore(&mailbox->cmd_head_resp_tail_lock, flags); + + return (u16)((reg_val & RESP_TAIL_MASK) >> RESP_TAIL_SHIFT); +} + +void gxp_mailbox_write_cmd_queue_head(struct gxp_mailbox *mailbox, u16 val) +{ + u32 current_resp_tail; + u32 new_cmd_head; + unsigned long flags; + + spin_lock_irqsave(&mailbox->cmd_head_resp_tail_lock, flags); + + current_resp_tail = data_read(mailbox, MBOX_DATA_CMD_HEAD_RESP_TAIL_OFFSET) & + RESP_TAIL_MASK; + new_cmd_head = (u32)val << CMD_HEAD_SHIFT; + data_write(mailbox, MBOX_DATA_CMD_HEAD_RESP_TAIL_OFFSET, + new_cmd_head | current_resp_tail); + + spin_unlock_irqrestore(&mailbox->cmd_head_resp_tail_lock, flags); +} + +void gxp_mailbox_write_resp_queue_tail(struct gxp_mailbox *mailbox, u16 val) +{ + u32 current_cmd_head; + u32 new_resp_tail; + unsigned long flags; + + spin_lock_irqsave(&mailbox->cmd_head_resp_tail_lock, flags); + + current_cmd_head = data_read(mailbox, MBOX_DATA_CMD_HEAD_RESP_TAIL_OFFSET) & + CMD_HEAD_MASK; + new_resp_tail = (u32)val << RESP_TAIL_SHIFT; + data_write(mailbox, MBOX_DATA_CMD_HEAD_RESP_TAIL_OFFSET, + current_cmd_head | new_resp_tail); + + spin_unlock_irqrestore(&mailbox->cmd_head_resp_tail_lock, flags); +} + +u16 gxp_mailbox_read_cmd_queue_tail(struct gxp_mailbox *mailbox) +{ + u32 reg_val; + unsigned long flags; + + spin_lock_irqsave(&mailbox->cmd_tail_resp_head_lock, flags); + + reg_val = data_read(mailbox, MBOX_DATA_CMD_TAIL_RESP_HEAD_OFFSET); + + spin_unlock_irqrestore(&mailbox->cmd_tail_resp_head_lock, flags); + + return (u16)((reg_val & CMD_TAIL_MASK) >> CMD_TAIL_SHIFT); +} + +u16 gxp_mailbox_read_resp_queue_head(struct gxp_mailbox *mailbox) +{ + u32 reg_val; + unsigned long flags; + + spin_lock_irqsave(&mailbox->cmd_tail_resp_head_lock, flags); + + reg_val = data_read(mailbox, MBOX_DATA_CMD_TAIL_RESP_HEAD_OFFSET); + + spin_unlock_irqrestore(&mailbox->cmd_tail_resp_head_lock, flags); + + return (u16)((reg_val & RESP_HEAD_MASK) >> RESP_HEAD_SHIFT); +} + +void gxp_mailbox_set_cmd_queue_tail(struct gxp_mailbox *mailbox, u32 value) +{ + mailbox->cmd_queue_tail = value; + gxp_mailbox_write_cmd_queue_tail(mailbox, value); +} + +void gxp_mailbox_set_resp_queue_head(struct gxp_mailbox *mailbox, u32 value) +{ + mailbox->resp_queue_head = value; + gxp_mailbox_write_resp_queue_head(mailbox, value); +} + +int gxp_mailbox_inc_cmd_queue_tail_nolock(struct gxp_mailbox *mailbox, u32 inc, + u32 wrap_bit) +{ + u32 head; + u32 remain_size; + u32 new_tail; + + if (inc > mailbox->cmd_queue_size) + return -EINVAL; + + head = gxp_mailbox_read_cmd_queue_head(mailbox); + remain_size = mailbox->cmd_queue_size - + gxp_circ_queue_cnt(head, mailbox->cmd_queue_tail, + mailbox->cmd_queue_size, wrap_bit); + /* no enough space left */ + if (inc > remain_size) + return -EBUSY; + + new_tail = gxp_circ_queue_inc(mailbox->cmd_queue_tail, inc, + mailbox->cmd_queue_size, wrap_bit); + gxp_mailbox_set_cmd_queue_tail(mailbox, new_tail); + return 0; +} + +int gxp_mailbox_inc_cmd_queue_tail_locked(struct gxp_mailbox *mailbox, u32 inc, + u32 wrap_bit) +{ + lockdep_assert_held(&mailbox->cmd_queue_lock); + return gxp_mailbox_inc_cmd_queue_tail_nolock(mailbox, inc, wrap_bit); +} + +int gxp_mailbox_inc_resp_queue_head_nolock(struct gxp_mailbox *mailbox, u32 inc, + u32 wrap_bit) +{ + u32 tail; + u32 size; + u32 new_head; + + if (inc > mailbox->resp_queue_size) + return -EINVAL; + + tail = gxp_mailbox_read_resp_queue_tail(mailbox); + size = gxp_circ_queue_cnt(mailbox->resp_queue_head, tail, + mailbox->resp_queue_size, wrap_bit); + if (inc > size) + return -EINVAL; + new_head = gxp_circ_queue_inc(mailbox->resp_queue_head, inc, + mailbox->resp_queue_size, wrap_bit); + gxp_mailbox_set_resp_queue_head(mailbox, new_head); + + return 0; +} + +int gxp_mailbox_inc_resp_queue_head_locked(struct gxp_mailbox *mailbox, u32 inc, + u32 wrap_bit) +{ + lockdep_assert_held(&mailbox->resp_queue_lock); + return gxp_mailbox_inc_resp_queue_head_nolock(mailbox, inc, wrap_bit); +} + +#if !GXP_USE_LEGACY_MAILBOX +u32 gxp_mailbox_gcip_ops_get_cmd_queue_head(struct gcip_mailbox *mailbox) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + return gxp_mailbox_read_cmd_queue_head(gxp_mbx); +} + +u32 gxp_mailbox_gcip_ops_get_cmd_queue_tail(struct gcip_mailbox *mailbox) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + return gxp_mbx->cmd_queue_tail; +} + +void gxp_mailbox_gcip_ops_inc_cmd_queue_tail(struct gcip_mailbox *mailbox, + u32 inc) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + lockdep_assert_held(&gxp_mbx->cmd_queue_lock); + gxp_mailbox_inc_cmd_queue_tail_nolock(gxp_mbx, inc, + mailbox->queue_wrap_bit); +} + +int gxp_mailbox_gcip_ops_acquire_cmd_queue_lock(struct gcip_mailbox *mailbox, + bool try) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + mutex_lock(&gxp_mbx->cmd_queue_lock); + return 1; +} + +void gxp_mailbox_gcip_ops_release_cmd_queue_lock(struct gcip_mailbox *mailbox) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + mutex_unlock(&gxp_mbx->cmd_queue_lock); +} + +u32 gxp_mailbox_gcip_ops_get_resp_queue_size(struct gcip_mailbox *mailbox) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + return gxp_mbx->resp_queue_size; +} + +u32 gxp_mailbox_gcip_ops_get_resp_queue_head(struct gcip_mailbox *mailbox) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + return gxp_mbx->resp_queue_head; +} + +u32 gxp_mailbox_gcip_ops_get_resp_queue_tail(struct gcip_mailbox *mailbox) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + return gxp_mailbox_read_resp_queue_tail(gxp_mbx); +} + +void gxp_mailbox_gcip_ops_inc_resp_queue_head(struct gcip_mailbox *mailbox, + u32 inc) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + lockdep_assert_held(&gxp_mbx->resp_queue_lock); + gxp_mailbox_inc_resp_queue_head_nolock(gxp_mbx, inc, + mailbox->queue_wrap_bit); +} + +int gxp_mailbox_gcip_ops_acquire_resp_queue_lock(struct gcip_mailbox *mailbox, + bool try) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + mutex_lock(&gxp_mbx->resp_queue_lock); + return 1; +} + +void gxp_mailbox_gcip_ops_release_resp_queue_lock(struct gcip_mailbox *mailbox) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + mutex_unlock(&gxp_mbx->resp_queue_lock); +} + +void gxp_mailbox_gcip_ops_acquire_wait_list_lock(struct gcip_mailbox *mailbox, + bool irqsave, + unsigned long *flags) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + mutex_lock(&gxp_mbx->wait_list_lock); +} + +void gxp_mailbox_gcip_ops_release_wait_list_lock(struct gcip_mailbox *mailbox, + bool irqrestore, + unsigned long flags) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + mutex_unlock(&gxp_mbx->wait_list_lock); +} + +int gxp_mailbox_gcip_ops_wait_for_cmd_queue_not_full( + struct gcip_mailbox *mailbox) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + u32 tail = gxp_mbx->cmd_queue_tail; + + /* + * If the cmd queue is full, it's up to the caller to retry. + */ + if (gxp_mailbox_read_cmd_queue_head(gxp_mbx) == + (tail ^ mailbox->queue_wrap_bit)) { + return -EAGAIN; + } + + return 0; +} + +int gxp_mailbox_gcip_ops_after_enqueue_cmd(struct gcip_mailbox *mailbox, + void *cmd) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + + /* triggers doorbell */ + gxp_mailbox_generate_device_interrupt(gxp_mbx, BIT(0)); + return 1; +} + +void gxp_mailbox_gcip_ops_after_fetch_resps(struct gcip_mailbox *mailbox, + u32 num_resps) +{ + struct gxp_mailbox *gxp_mbx = mailbox->data; + u32 size = gxp_mbx->resp_queue_size; + + /* + * Now that the response queue has been drained, send an interrupt + * to the device in case firmware was waiting for us to consume + * responses. + */ + if (num_resps == size) + gxp_mailbox_generate_device_interrupt(gxp_mbx, BIT(0)); +} +#endif /* !GXP_USE_LEGACY_MAILBOX */ |