|  | // SPDX-License-Identifier: GPL-2.0-or-later | 
|  | /* | 
|  | * linux/drivers/mmc/core/sdio_irq.c | 
|  | * | 
|  | * Author:      Nicolas Pitre | 
|  | * Created:     June 18, 2007 | 
|  | * Copyright:   MontaVista Software Inc. | 
|  | * | 
|  | * Copyright 2008 Pierre Ossman | 
|  | */ | 
|  |  | 
|  | #include <linux/kernel.h> | 
|  | #include <linux/sched.h> | 
|  | #include <uapi/linux/sched/types.h> | 
|  | #include <linux/kthread.h> | 
|  | #include <linux/export.h> | 
|  | #include <linux/wait.h> | 
|  | #include <linux/delay.h> | 
|  |  | 
|  | #include <linux/mmc/core.h> | 
|  | #include <linux/mmc/host.h> | 
|  | #include <linux/mmc/card.h> | 
|  | #include <linux/mmc/sdio.h> | 
|  | #include <linux/mmc/sdio_func.h> | 
|  |  | 
|  | #include "sdio_ops.h" | 
|  | #include "core.h" | 
|  | #include "card.h" | 
|  |  | 
|  | static int sdio_get_pending_irqs(struct mmc_host *host, u8 *pending) | 
|  | { | 
|  | struct mmc_card *card = host->card; | 
|  | int ret; | 
|  |  | 
|  | WARN_ON(!host->claimed); | 
|  |  | 
|  | ret = mmc_io_rw_direct(card, 0, 0, SDIO_CCCR_INTx, 0, pending); | 
|  | if (ret) { | 
|  | pr_debug("%s: error %d reading SDIO_CCCR_INTx\n", | 
|  | mmc_card_id(card), ret); | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | if (*pending && mmc_card_broken_irq_polling(card) && | 
|  | !(host->caps & MMC_CAP_SDIO_IRQ)) { | 
|  | unsigned char dummy; | 
|  |  | 
|  | /* A fake interrupt could be created when we poll SDIO_CCCR_INTx | 
|  | * register with a Marvell SD8797 card. A dummy CMD52 read to | 
|  | * function 0 register 0xff can avoid this. | 
|  | */ | 
|  | mmc_io_rw_direct(card, 0, 0, 0xff, 0, &dummy); | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int process_sdio_pending_irqs(struct mmc_host *host) | 
|  | { | 
|  | struct mmc_card *card = host->card; | 
|  | int i, ret, count; | 
|  | bool sdio_irq_pending = host->sdio_irq_pending; | 
|  | unsigned char pending; | 
|  | struct sdio_func *func; | 
|  |  | 
|  | /* Don't process SDIO IRQs if the card is suspended. */ | 
|  | if (mmc_card_suspended(card)) | 
|  | return 0; | 
|  |  | 
|  | /* Clear the flag to indicate that we have processed the IRQ. */ | 
|  | host->sdio_irq_pending = false; | 
|  |  | 
|  | /* | 
|  | * Optimization, if there is only 1 function interrupt registered | 
|  | * and we know an IRQ was signaled then call irq handler directly. | 
|  | * Otherwise do the full probe. | 
|  | */ | 
|  | func = card->sdio_single_irq; | 
|  | if (func && sdio_irq_pending) { | 
|  | func->irq_handler(func); | 
|  | return 1; | 
|  | } | 
|  |  | 
|  | ret = sdio_get_pending_irqs(host, &pending); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | count = 0; | 
|  | for (i = 1; i <= 7; i++) { | 
|  | if (pending & (1 << i)) { | 
|  | func = card->sdio_func[i - 1]; | 
|  | if (!func) { | 
|  | pr_warn("%s: pending IRQ for non-existent function\n", | 
|  | mmc_card_id(card)); | 
|  | ret = -EINVAL; | 
|  | } else if (func->irq_handler) { | 
|  | func->irq_handler(func); | 
|  | count++; | 
|  | } else { | 
|  | pr_warn("%s: pending IRQ with no handler\n", | 
|  | sdio_func_id(func)); | 
|  | ret = -EINVAL; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | if (count) | 
|  | return count; | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static void sdio_run_irqs(struct mmc_host *host) | 
|  | { | 
|  | mmc_claim_host(host); | 
|  | if (host->sdio_irqs) { | 
|  | process_sdio_pending_irqs(host); | 
|  | if (!host->sdio_irq_pending) | 
|  | host->ops->ack_sdio_irq(host); | 
|  | } | 
|  | mmc_release_host(host); | 
|  | } | 
|  |  | 
|  | void sdio_irq_work(struct work_struct *work) | 
|  | { | 
|  | struct mmc_host *host = | 
|  | container_of(work, struct mmc_host, sdio_irq_work.work); | 
|  |  | 
|  | sdio_run_irqs(host); | 
|  | } | 
|  |  | 
|  | void sdio_signal_irq(struct mmc_host *host) | 
|  | { | 
|  | host->sdio_irq_pending = true; | 
|  | queue_delayed_work(system_wq, &host->sdio_irq_work, 0); | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(sdio_signal_irq); | 
|  |  | 
|  | static int sdio_irq_thread(void *_host) | 
|  | { | 
|  | struct mmc_host *host = _host; | 
|  | struct sched_param param = { .sched_priority = 1 }; | 
|  | unsigned long period, idle_period; | 
|  | int ret; | 
|  |  | 
|  | sched_setscheduler(current, SCHED_FIFO, ¶m); | 
|  |  | 
|  | /* | 
|  | * We want to allow for SDIO cards to work even on non SDIO | 
|  | * aware hosts.  One thing that non SDIO host cannot do is | 
|  | * asynchronous notification of pending SDIO card interrupts | 
|  | * hence we poll for them in that case. | 
|  | */ | 
|  | idle_period = msecs_to_jiffies(10); | 
|  | period = (host->caps & MMC_CAP_SDIO_IRQ) ? | 
|  | MAX_SCHEDULE_TIMEOUT : idle_period; | 
|  |  | 
|  | pr_debug("%s: IRQ thread started (poll period = %lu jiffies)\n", | 
|  | mmc_hostname(host), period); | 
|  |  | 
|  | do { | 
|  | /* | 
|  | * We claim the host here on drivers behalf for a couple | 
|  | * reasons: | 
|  | * | 
|  | * 1) it is already needed to retrieve the CCCR_INTx; | 
|  | * 2) we want the driver(s) to clear the IRQ condition ASAP; | 
|  | * 3) we need to control the abort condition locally. | 
|  | * | 
|  | * Just like traditional hard IRQ handlers, we expect SDIO | 
|  | * IRQ handlers to be quick and to the point, so that the | 
|  | * holding of the host lock does not cover too much work | 
|  | * that doesn't require that lock to be held. | 
|  | */ | 
|  | ret = __mmc_claim_host(host, NULL, | 
|  | &host->sdio_irq_thread_abort); | 
|  | if (ret) | 
|  | break; | 
|  | ret = process_sdio_pending_irqs(host); | 
|  | mmc_release_host(host); | 
|  |  | 
|  | /* | 
|  | * Give other threads a chance to run in the presence of | 
|  | * errors. | 
|  | */ | 
|  | if (ret < 0) { | 
|  | set_current_state(TASK_INTERRUPTIBLE); | 
|  | if (!kthread_should_stop()) | 
|  | schedule_timeout(HZ); | 
|  | set_current_state(TASK_RUNNING); | 
|  | } | 
|  |  | 
|  | /* | 
|  | * Adaptive polling frequency based on the assumption | 
|  | * that an interrupt will be closely followed by more. | 
|  | * This has a substantial benefit for network devices. | 
|  | */ | 
|  | if (!(host->caps & MMC_CAP_SDIO_IRQ)) { | 
|  | if (ret > 0) | 
|  | period /= 2; | 
|  | else { | 
|  | period++; | 
|  | if (period > idle_period) | 
|  | period = idle_period; | 
|  | } | 
|  | } | 
|  |  | 
|  | set_current_state(TASK_INTERRUPTIBLE); | 
|  | if (host->caps & MMC_CAP_SDIO_IRQ) | 
|  | host->ops->enable_sdio_irq(host, 1); | 
|  | if (!kthread_should_stop()) | 
|  | schedule_timeout(period); | 
|  | set_current_state(TASK_RUNNING); | 
|  | } while (!kthread_should_stop()); | 
|  |  | 
|  | if (host->caps & MMC_CAP_SDIO_IRQ) | 
|  | host->ops->enable_sdio_irq(host, 0); | 
|  |  | 
|  | pr_debug("%s: IRQ thread exiting with code %d\n", | 
|  | mmc_hostname(host), ret); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  |  | 
|  | static int sdio_card_irq_get(struct mmc_card *card) | 
|  | { | 
|  | struct mmc_host *host = card->host; | 
|  |  | 
|  | WARN_ON(!host->claimed); | 
|  |  | 
|  | if (!host->sdio_irqs++) { | 
|  | if (!(host->caps2 & MMC_CAP2_SDIO_IRQ_NOTHREAD)) { | 
|  | atomic_set(&host->sdio_irq_thread_abort, 0); | 
|  | host->sdio_irq_thread = | 
|  | kthread_run(sdio_irq_thread, host, | 
|  | "ksdioirqd/%s", mmc_hostname(host)); | 
|  | if (IS_ERR(host->sdio_irq_thread)) { | 
|  | int err = PTR_ERR(host->sdio_irq_thread); | 
|  | host->sdio_irqs--; | 
|  | return err; | 
|  | } | 
|  | } else if (host->caps & MMC_CAP_SDIO_IRQ) { | 
|  | host->ops->enable_sdio_irq(host, 1); | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | static int sdio_card_irq_put(struct mmc_card *card) | 
|  | { | 
|  | struct mmc_host *host = card->host; | 
|  |  | 
|  | WARN_ON(!host->claimed); | 
|  |  | 
|  | if (host->sdio_irqs < 1) | 
|  | return -EINVAL; | 
|  |  | 
|  | if (!--host->sdio_irqs) { | 
|  | if (!(host->caps2 & MMC_CAP2_SDIO_IRQ_NOTHREAD)) { | 
|  | atomic_set(&host->sdio_irq_thread_abort, 1); | 
|  | kthread_stop(host->sdio_irq_thread); | 
|  | } else if (host->caps & MMC_CAP_SDIO_IRQ) { | 
|  | host->ops->enable_sdio_irq(host, 0); | 
|  | } | 
|  | } | 
|  |  | 
|  | return 0; | 
|  | } | 
|  |  | 
|  | /* If there is only 1 function registered set sdio_single_irq */ | 
|  | static void sdio_single_irq_set(struct mmc_card *card) | 
|  | { | 
|  | struct sdio_func *func; | 
|  | int i; | 
|  |  | 
|  | card->sdio_single_irq = NULL; | 
|  | if ((card->host->caps & MMC_CAP_SDIO_IRQ) && | 
|  | card->host->sdio_irqs == 1) | 
|  | for (i = 0; i < card->sdio_funcs; i++) { | 
|  | func = card->sdio_func[i]; | 
|  | if (func && func->irq_handler) { | 
|  | card->sdio_single_irq = func; | 
|  | break; | 
|  | } | 
|  | } | 
|  | } | 
|  |  | 
|  | /** | 
|  | *	sdio_claim_irq - claim the IRQ for a SDIO function | 
|  | *	@func: SDIO function | 
|  | *	@handler: IRQ handler callback | 
|  | * | 
|  | *	Claim and activate the IRQ for the given SDIO function. The provided | 
|  | *	handler will be called when that IRQ is asserted.  The host is always | 
|  | *	claimed already when the handler is called so the handler should not | 
|  | *	call sdio_claim_host() or sdio_release_host(). | 
|  | */ | 
|  | int sdio_claim_irq(struct sdio_func *func, sdio_irq_handler_t *handler) | 
|  | { | 
|  | int ret; | 
|  | unsigned char reg; | 
|  |  | 
|  | if (!func) | 
|  | return -EINVAL; | 
|  |  | 
|  | pr_debug("SDIO: Enabling IRQ for %s...\n", sdio_func_id(func)); | 
|  |  | 
|  | if (func->irq_handler) { | 
|  | pr_debug("SDIO: IRQ for %s already in use.\n", sdio_func_id(func)); | 
|  | return -EBUSY; | 
|  | } | 
|  |  | 
|  | ret = mmc_io_rw_direct(func->card, 0, 0, SDIO_CCCR_IENx, 0, ®); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | reg |= 1 << func->num; | 
|  |  | 
|  | reg |= 1; /* Master interrupt enable */ | 
|  |  | 
|  | ret = mmc_io_rw_direct(func->card, 1, 0, SDIO_CCCR_IENx, reg, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | func->irq_handler = handler; | 
|  | ret = sdio_card_irq_get(func->card); | 
|  | if (ret) | 
|  | func->irq_handler = NULL; | 
|  | sdio_single_irq_set(func->card); | 
|  |  | 
|  | return ret; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(sdio_claim_irq); | 
|  |  | 
|  | /** | 
|  | *	sdio_release_irq - release the IRQ for a SDIO function | 
|  | *	@func: SDIO function | 
|  | * | 
|  | *	Disable and release the IRQ for the given SDIO function. | 
|  | */ | 
|  | int sdio_release_irq(struct sdio_func *func) | 
|  | { | 
|  | int ret; | 
|  | unsigned char reg; | 
|  |  | 
|  | if (!func) | 
|  | return -EINVAL; | 
|  |  | 
|  | pr_debug("SDIO: Disabling IRQ for %s...\n", sdio_func_id(func)); | 
|  |  | 
|  | if (func->irq_handler) { | 
|  | func->irq_handler = NULL; | 
|  | sdio_card_irq_put(func->card); | 
|  | sdio_single_irq_set(func->card); | 
|  | } | 
|  |  | 
|  | ret = mmc_io_rw_direct(func->card, 0, 0, SDIO_CCCR_IENx, 0, ®); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | reg &= ~(1 << func->num); | 
|  |  | 
|  | /* Disable master interrupt with the last function interrupt */ | 
|  | if (!(reg & 0xFE)) | 
|  | reg = 0; | 
|  |  | 
|  | ret = mmc_io_rw_direct(func->card, 1, 0, SDIO_CCCR_IENx, reg, NULL); | 
|  | if (ret) | 
|  | return ret; | 
|  |  | 
|  | return 0; | 
|  | } | 
|  | EXPORT_SYMBOL_GPL(sdio_release_irq); | 
|  |  |