2021-06-03 13:41:55 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
//
|
|
|
|
// Copyright (C) 2021 ROHM Semiconductors
|
|
|
|
// regulator IRQ based event notification helpers
|
|
|
|
//
|
|
|
|
// Logic has been partially adapted from qcom-labibb driver.
|
|
|
|
//
|
|
|
|
// Author: Matti Vaittinen <matti.vaittinen@fi.rohmeurope.com>
|
|
|
|
|
|
|
|
#include <linux/device.h>
|
|
|
|
#include <linux/err.h>
|
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/reboot.h>
|
|
|
|
#include <linux/regmap.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/spinlock.h>
|
|
|
|
#include <linux/regulator/driver.h>
|
|
|
|
|
|
|
|
#include "internal.h"
|
|
|
|
|
|
|
|
#define REGULATOR_FORCED_SAFETY_SHUTDOWN_WAIT_MS 10000
|
|
|
|
|
|
|
|
struct regulator_irq {
|
|
|
|
struct regulator_irq_data rdata;
|
|
|
|
struct regulator_irq_desc desc;
|
|
|
|
int irq;
|
|
|
|
int retry_cnt;
|
|
|
|
struct delayed_work isr_work;
|
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Should only be called from threaded handler to prevent potential deadlock
|
|
|
|
*/
|
|
|
|
static void rdev_flag_err(struct regulator_dev *rdev, int err)
|
|
|
|
{
|
|
|
|
spin_lock(&rdev->err_lock);
|
|
|
|
rdev->cached_err |= err;
|
|
|
|
spin_unlock(&rdev->err_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void rdev_clear_err(struct regulator_dev *rdev, int err)
|
|
|
|
{
|
|
|
|
spin_lock(&rdev->err_lock);
|
|
|
|
rdev->cached_err &= ~err;
|
|
|
|
spin_unlock(&rdev->err_lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void regulator_notifier_isr_work(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct regulator_irq *h;
|
|
|
|
struct regulator_irq_desc *d;
|
|
|
|
struct regulator_irq_data *rid;
|
|
|
|
int ret = 0;
|
|
|
|
int tmo, i;
|
|
|
|
int num_rdevs;
|
|
|
|
|
|
|
|
h = container_of(work, struct regulator_irq,
|
|
|
|
isr_work.work);
|
|
|
|
d = &h->desc;
|
|
|
|
rid = &h->rdata;
|
|
|
|
num_rdevs = rid->num_states;
|
|
|
|
|
|
|
|
reread:
|
|
|
|
if (d->fatal_cnt && h->retry_cnt > d->fatal_cnt) {
|
|
|
|
if (!d->die)
|
|
|
|
return hw_protection_shutdown("Regulator HW failure? - no IC recovery",
|
|
|
|
REGULATOR_FORCED_SAFETY_SHUTDOWN_WAIT_MS);
|
|
|
|
ret = d->die(rid);
|
|
|
|
/*
|
|
|
|
* If the 'last resort' IC recovery failed we will have
|
|
|
|
* nothing else left to do...
|
|
|
|
*/
|
|
|
|
if (ret)
|
|
|
|
return hw_protection_shutdown("Regulator HW failure. IC recovery failed",
|
|
|
|
REGULATOR_FORCED_SAFETY_SHUTDOWN_WAIT_MS);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If h->die() was implemented we assume recovery has been
|
|
|
|
* attempted (probably regulator was shut down) and we
|
|
|
|
* just enable IRQ and bail-out.
|
|
|
|
*/
|
|
|
|
goto enable_out;
|
|
|
|
}
|
|
|
|
if (d->renable) {
|
|
|
|
ret = d->renable(rid);
|
|
|
|
|
|
|
|
if (ret == REGULATOR_FAILED_RETRY) {
|
|
|
|
/* Driver could not get current status */
|
|
|
|
h->retry_cnt++;
|
|
|
|
if (!d->reread_ms)
|
|
|
|
goto reread;
|
|
|
|
|
|
|
|
tmo = d->reread_ms;
|
|
|
|
goto reschedule;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (ret) {
|
|
|
|
/*
|
|
|
|
* IC status reading succeeded. update error info
|
|
|
|
* just in case the renable changed it.
|
|
|
|
*/
|
|
|
|
for (i = 0; i < num_rdevs; i++) {
|
|
|
|
struct regulator_err_state *stat;
|
|
|
|
struct regulator_dev *rdev;
|
|
|
|
|
|
|
|
stat = &rid->states[i];
|
|
|
|
rdev = stat->rdev;
|
|
|
|
rdev_clear_err(rdev, (~stat->errors) &
|
|
|
|
stat->possible_errs);
|
|
|
|
}
|
|
|
|
h->retry_cnt++;
|
|
|
|
/*
|
|
|
|
* The IC indicated problem is still ON - no point in
|
|
|
|
* re-enabling the IRQ. Retry later.
|
|
|
|
*/
|
|
|
|
tmo = d->irq_off_ms;
|
|
|
|
goto reschedule;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Either IC reported problem cleared or no status checker was provided.
|
|
|
|
* If problems are gone - good. If not - then the IRQ will fire again
|
|
|
|
* and we'll have a new nice loop. In any case we should clear error
|
|
|
|
* flags here and re-enable IRQs.
|
|
|
|
*/
|
|
|
|
for (i = 0; i < num_rdevs; i++) {
|
|
|
|
struct regulator_err_state *stat;
|
|
|
|
struct regulator_dev *rdev;
|
|
|
|
|
|
|
|
stat = &rid->states[i];
|
|
|
|
rdev = stat->rdev;
|
|
|
|
rdev_clear_err(rdev, stat->possible_errs);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Things have been seemingly successful => zero retry-counter.
|
|
|
|
*/
|
|
|
|
h->retry_cnt = 0;
|
|
|
|
|
|
|
|
enable_out:
|
|
|
|
enable_irq(h->irq);
|
|
|
|
|
|
|
|
return;
|
|
|
|
|
|
|
|
reschedule:
|
|
|
|
if (!d->high_prio)
|
|
|
|
mod_delayed_work(system_wq, &h->isr_work,
|
|
|
|
msecs_to_jiffies(tmo));
|
|
|
|
else
|
|
|
|
mod_delayed_work(system_highpri_wq, &h->isr_work,
|
|
|
|
msecs_to_jiffies(tmo));
|
|
|
|
}
|
|
|
|
|
|
|
|
static irqreturn_t regulator_notifier_isr(int irq, void *data)
|
|
|
|
{
|
|
|
|
struct regulator_irq *h = data;
|
|
|
|
struct regulator_irq_desc *d;
|
|
|
|
struct regulator_irq_data *rid;
|
|
|
|
unsigned long rdev_map = 0;
|
|
|
|
int num_rdevs;
|
|
|
|
int ret, i;
|
|
|
|
|
|
|
|
d = &h->desc;
|
|
|
|
rid = &h->rdata;
|
|
|
|
num_rdevs = rid->num_states;
|
|
|
|
|
|
|
|
if (d->fatal_cnt)
|
|
|
|
h->retry_cnt++;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* we spare a few cycles by not clearing statuses prior to this call.
|
|
|
|
* The IC driver must initialize the status buffers for rdevs
|
|
|
|
* which it indicates having active events via rdev_map.
|
|
|
|
*
|
|
|
|
* Maybe we should just to be on a safer side(?)
|
|
|
|
*/
|
|
|
|
ret = d->map_event(irq, rid, &rdev_map);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If status reading fails (which is unlikely) we don't ack/disable
|
|
|
|
* IRQ but just increase fail count and retry when IRQ fires again.
|
|
|
|
* If retry_count exceeds the given safety limit we call IC specific die
|
|
|
|
* handler which can try disabling regulator(s).
|
|
|
|
*
|
2021-08-23 15:56:51 +08:00
|
|
|
* If no die handler is given we will just power-off as a last resort.
|
2021-06-03 13:41:55 +08:00
|
|
|
*
|
|
|
|
* We could try disabling all associated rdevs - but we might shoot
|
|
|
|
* ourselves in the head and leave the problematic regulator enabled. So
|
|
|
|
* if IC has no die-handler populated we just assume the regulator
|
|
|
|
* can't be disabled.
|
|
|
|
*/
|
|
|
|
if (unlikely(ret == REGULATOR_FAILED_RETRY))
|
|
|
|
goto fail_out;
|
|
|
|
|
|
|
|
h->retry_cnt = 0;
|
|
|
|
/*
|
|
|
|
* Let's not disable IRQ if there were no status bits for us. We'd
|
|
|
|
* better leave spurious IRQ handling to genirq
|
|
|
|
*/
|
|
|
|
if (ret || !rdev_map)
|
|
|
|
return IRQ_NONE;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Some events are bogus if the regulator is disabled. Skip such events
|
|
|
|
* if all relevant regulators are disabled
|
|
|
|
*/
|
|
|
|
if (d->skip_off) {
|
|
|
|
for_each_set_bit(i, &rdev_map, num_rdevs) {
|
|
|
|
struct regulator_dev *rdev;
|
|
|
|
const struct regulator_ops *ops;
|
|
|
|
|
|
|
|
rdev = rid->states[i].rdev;
|
|
|
|
ops = rdev->desc->ops;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If any of the flagged regulators is enabled we do
|
|
|
|
* handle this
|
|
|
|
*/
|
|
|
|
if (ops->is_enabled(rdev))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (i == num_rdevs)
|
|
|
|
return IRQ_NONE;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Disable IRQ if HW keeps line asserted */
|
|
|
|
if (d->irq_off_ms)
|
|
|
|
disable_irq_nosync(irq);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* IRQ seems to be for us. Let's fire correct notifiers / store error
|
|
|
|
* flags
|
|
|
|
*/
|
|
|
|
for_each_set_bit(i, &rdev_map, num_rdevs) {
|
|
|
|
struct regulator_err_state *stat;
|
|
|
|
struct regulator_dev *rdev;
|
|
|
|
|
|
|
|
stat = &rid->states[i];
|
|
|
|
rdev = stat->rdev;
|
|
|
|
|
|
|
|
rdev_dbg(rdev, "Sending regulator notification EVT 0x%lx\n",
|
|
|
|
stat->notifs);
|
|
|
|
|
|
|
|
regulator_notifier_call_chain(rdev, stat->notifs, NULL);
|
|
|
|
rdev_flag_err(rdev, stat->errors);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (d->irq_off_ms) {
|
|
|
|
if (!d->high_prio)
|
|
|
|
schedule_delayed_work(&h->isr_work,
|
|
|
|
msecs_to_jiffies(d->irq_off_ms));
|
|
|
|
else
|
|
|
|
mod_delayed_work(system_highpri_wq,
|
|
|
|
&h->isr_work,
|
|
|
|
msecs_to_jiffies(d->irq_off_ms));
|
|
|
|
}
|
|
|
|
|
|
|
|
return IRQ_HANDLED;
|
|
|
|
|
|
|
|
fail_out:
|
|
|
|
if (d->fatal_cnt && h->retry_cnt > d->fatal_cnt) {
|
|
|
|
/* If we have no recovery, just try shut down straight away */
|
|
|
|
if (!d->die) {
|
|
|
|
hw_protection_shutdown("Regulator failure. Retry count exceeded",
|
|
|
|
REGULATOR_FORCED_SAFETY_SHUTDOWN_WAIT_MS);
|
|
|
|
} else {
|
|
|
|
ret = d->die(rid);
|
|
|
|
/* If die() failed shut down as a last attempt to save the HW */
|
|
|
|
if (ret)
|
|
|
|
hw_protection_shutdown("Regulator failure. Recovery failed",
|
|
|
|
REGULATOR_FORCED_SAFETY_SHUTDOWN_WAIT_MS);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return IRQ_NONE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int init_rdev_state(struct device *dev, struct regulator_irq *h,
|
|
|
|
struct regulator_dev **rdev, int common_err,
|
|
|
|
int *rdev_err, int rdev_amount)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
h->rdata.states = devm_kzalloc(dev, sizeof(*h->rdata.states) *
|
|
|
|
rdev_amount, GFP_KERNEL);
|
|
|
|
if (!h->rdata.states)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
h->rdata.num_states = rdev_amount;
|
|
|
|
h->rdata.data = h->desc.data;
|
|
|
|
|
|
|
|
for (i = 0; i < rdev_amount; i++) {
|
|
|
|
h->rdata.states[i].possible_errs = common_err;
|
|
|
|
if (rdev_err)
|
|
|
|
h->rdata.states[i].possible_errs |= *rdev_err++;
|
|
|
|
h->rdata.states[i].rdev = *rdev++;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void init_rdev_errors(struct regulator_irq *h)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < h->rdata.num_states; i++)
|
|
|
|
if (h->rdata.states[i].possible_errs)
|
|
|
|
h->rdata.states[i].rdev->use_cached_err = true;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* regulator_irq_helper - register IRQ based regulator event/error notifier
|
|
|
|
*
|
|
|
|
* @dev: device providing the IRQs
|
|
|
|
* @d: IRQ helper descriptor.
|
|
|
|
* @irq: IRQ used to inform events/errors to be notified.
|
|
|
|
* @irq_flags: Extra IRQ flags to be OR'ed with the default
|
|
|
|
* IRQF_ONESHOT when requesting the (threaded) irq.
|
|
|
|
* @common_errs: Errors which can be flagged by this IRQ for all rdevs.
|
|
|
|
* When IRQ is re-enabled these errors will be cleared
|
|
|
|
* from all associated regulators
|
|
|
|
* @per_rdev_errs: Optional error flag array describing errors specific
|
|
|
|
* for only some of the regulators. These errors will be
|
|
|
|
* or'ed with common errors. If this is given the array
|
|
|
|
* should contain rdev_amount flags. Can be set to NULL
|
|
|
|
* if there is no regulator specific error flags for this
|
|
|
|
* IRQ.
|
|
|
|
* @rdev: Array of pointers to regulators associated with this
|
|
|
|
* IRQ.
|
|
|
|
* @rdev_amount: Amount of regulators associated with this IRQ.
|
|
|
|
*
|
|
|
|
* Return: handle to irq_helper or an ERR_PTR() encoded error code.
|
|
|
|
*/
|
|
|
|
void *regulator_irq_helper(struct device *dev,
|
|
|
|
const struct regulator_irq_desc *d, int irq,
|
|
|
|
int irq_flags, int common_errs, int *per_rdev_errs,
|
|
|
|
struct regulator_dev **rdev, int rdev_amount)
|
|
|
|
{
|
|
|
|
struct regulator_irq *h;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (!rdev_amount || !d || !d->map_event || !d->name)
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
|
|
h = devm_kzalloc(dev, sizeof(*h), GFP_KERNEL);
|
|
|
|
if (!h)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
|
|
|
h->irq = irq;
|
|
|
|
h->desc = *d;
|
|
|
|
|
|
|
|
ret = init_rdev_state(dev, h, rdev, common_errs, per_rdev_errs,
|
|
|
|
rdev_amount);
|
|
|
|
if (ret)
|
|
|
|
return ERR_PTR(ret);
|
|
|
|
|
|
|
|
init_rdev_errors(h);
|
|
|
|
|
|
|
|
if (h->desc.irq_off_ms)
|
|
|
|
INIT_DELAYED_WORK(&h->isr_work, regulator_notifier_isr_work);
|
|
|
|
|
|
|
|
ret = request_threaded_irq(h->irq, NULL, regulator_notifier_isr,
|
|
|
|
IRQF_ONESHOT | irq_flags, h->desc.name, h);
|
|
|
|
if (ret) {
|
|
|
|
dev_err(dev, "Failed to request IRQ %d\n", irq);
|
|
|
|
|
|
|
|
return ERR_PTR(ret);
|
|
|
|
}
|
|
|
|
|
|
|
|
return h;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(regulator_irq_helper);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* regulator_irq_helper_cancel - drop IRQ based regulator event/error notifier
|
|
|
|
*
|
|
|
|
* @handle: Pointer to handle returned by a successful call to
|
|
|
|
* regulator_irq_helper(). Will be NULLed upon return.
|
|
|
|
*
|
|
|
|
* The associated IRQ is released and work is cancelled when the function
|
|
|
|
* returns.
|
|
|
|
*/
|
|
|
|
void regulator_irq_helper_cancel(void **handle)
|
|
|
|
{
|
|
|
|
if (handle && *handle) {
|
|
|
|
struct regulator_irq *h = *handle;
|
|
|
|
|
|
|
|
free_irq(h->irq, h);
|
|
|
|
if (h->desc.irq_off_ms)
|
|
|
|
cancel_delayed_work_sync(&h->isr_work);
|
|
|
|
|
|
|
|
h = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(regulator_irq_helper_cancel);
|