forked from Mirrors/opensbi

Since commit 50d4fde1c5
("lib: Remove redundant sbi_platform_ipi_clear()
calls"), the IPI sent from the boot hart in wake_coldboot_harts() is not
cleared in the secondary harts until they reach sbi_ipi_init(). However,
sbi_hsm_init() and sbi_hsm_hart_wait() are called earlier, so a secondary
hart might enter sbi_hsm_hart_wait() with an already pending IPI.
sbi_hsm_hart_wait() makes sure the hart leaves the loop only when it is
actually ready, so a pending unrelated IPI should not cause safety issues.
However, it might be inefficient on certain hardware, because it prevents
"wfi" from stalling the hart even if the hardware supports this, making the
hart needlessly spin in a "busy-wait" loop.
This behaviour can be observed, for example, in a QEMU VM (QEMU 7.2.0) with
"-machine virt" running a Linux guest. Inserting delays in
sbi_hsm_hart_start() allows reproducing the issue more reliably.
The comment in wait_for_coldboot() suggests that the initial IPI is needed
in the warm resume path, so let us clear it before init_warm_startup()
only.
To do this, sbi_ipi_raw_clear() was created similar to sbi_ipi_raw_send().
Signed-off-by: Evgenii Shatokhin <e.shatokhin@yadro.com>
Reviewed-by: Anup Patel <anup@brainfault.org>
292 lines
6.1 KiB
C
292 lines
6.1 KiB
C
/*
|
|
* SPDX-License-Identifier: BSD-2-Clause
|
|
*
|
|
* Copyright (c) 2019 Western Digital Corporation or its affiliates.
|
|
*
|
|
* Authors:
|
|
* Anup Patel <anup.patel@wdc.com>
|
|
* Nick Kossifidis <mick@ics.forth.gr>
|
|
*/
|
|
|
|
#include <sbi/riscv_asm.h>
|
|
#include <sbi/riscv_atomic.h>
|
|
#include <sbi/riscv_barrier.h>
|
|
#include <sbi/sbi_bitops.h>
|
|
#include <sbi/sbi_domain.h>
|
|
#include <sbi/sbi_error.h>
|
|
#include <sbi/sbi_hart.h>
|
|
#include <sbi/sbi_hsm.h>
|
|
#include <sbi/sbi_init.h>
|
|
#include <sbi/sbi_ipi.h>
|
|
#include <sbi/sbi_platform.h>
|
|
#include <sbi/sbi_pmu.h>
|
|
#include <sbi/sbi_string.h>
|
|
#include <sbi/sbi_tlb.h>
|
|
|
|
struct sbi_ipi_data {
|
|
unsigned long ipi_type;
|
|
};
|
|
|
|
static unsigned long ipi_data_off;
|
|
static const struct sbi_ipi_device *ipi_dev = NULL;
|
|
static const struct sbi_ipi_event_ops *ipi_ops_array[SBI_IPI_EVENT_MAX];
|
|
|
|
static int sbi_ipi_send(struct sbi_scratch *scratch, u32 remote_hartid,
|
|
u32 event, void *data)
|
|
{
|
|
int ret;
|
|
struct sbi_scratch *remote_scratch = NULL;
|
|
struct sbi_ipi_data *ipi_data;
|
|
const struct sbi_ipi_event_ops *ipi_ops;
|
|
|
|
if ((SBI_IPI_EVENT_MAX <= event) ||
|
|
!ipi_ops_array[event])
|
|
return SBI_EINVAL;
|
|
ipi_ops = ipi_ops_array[event];
|
|
|
|
remote_scratch = sbi_hartid_to_scratch(remote_hartid);
|
|
if (!remote_scratch)
|
|
return SBI_EINVAL;
|
|
|
|
ipi_data = sbi_scratch_offset_ptr(remote_scratch, ipi_data_off);
|
|
|
|
if (ipi_ops->update) {
|
|
ret = ipi_ops->update(scratch, remote_scratch,
|
|
remote_hartid, data);
|
|
if (ret < 0)
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Set IPI type on remote hart's scratch area and
|
|
* trigger the interrupt
|
|
*/
|
|
atomic_raw_set_bit(event, &ipi_data->ipi_type);
|
|
smp_wmb();
|
|
|
|
if (ipi_dev && ipi_dev->ipi_send)
|
|
ipi_dev->ipi_send(remote_hartid);
|
|
|
|
sbi_pmu_ctr_incr_fw(SBI_PMU_FW_IPI_SENT);
|
|
|
|
if (ipi_ops->sync)
|
|
ipi_ops->sync(scratch);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* As this this function only handlers scalar values of hart mask, it must be
|
|
* set to all online harts if the intention is to send IPIs to all the harts.
|
|
* If hmask is zero, no IPIs will be sent.
|
|
*/
|
|
int sbi_ipi_send_many(ulong hmask, ulong hbase, u32 event, void *data)
|
|
{
|
|
int rc;
|
|
ulong i, m;
|
|
struct sbi_domain *dom = sbi_domain_thishart_ptr();
|
|
struct sbi_scratch *scratch = sbi_scratch_thishart_ptr();
|
|
|
|
if (hbase != -1UL) {
|
|
rc = sbi_hsm_hart_interruptible_mask(dom, hbase, &m);
|
|
if (rc)
|
|
return rc;
|
|
m &= hmask;
|
|
|
|
/* Send IPIs */
|
|
for (i = hbase; m; i++, m >>= 1) {
|
|
if (m & 1UL)
|
|
sbi_ipi_send(scratch, i, event, data);
|
|
}
|
|
} else {
|
|
hbase = 0;
|
|
while (!sbi_hsm_hart_interruptible_mask(dom, hbase, &m)) {
|
|
/* Send IPIs */
|
|
for (i = hbase; m; i++, m >>= 1) {
|
|
if (m & 1UL)
|
|
sbi_ipi_send(scratch, i, event, data);
|
|
}
|
|
hbase += BITS_PER_LONG;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int sbi_ipi_event_create(const struct sbi_ipi_event_ops *ops)
|
|
{
|
|
int i, ret = SBI_ENOSPC;
|
|
|
|
if (!ops || !ops->process)
|
|
return SBI_EINVAL;
|
|
|
|
for (i = 0; i < SBI_IPI_EVENT_MAX; i++) {
|
|
if (!ipi_ops_array[i]) {
|
|
ret = i;
|
|
ipi_ops_array[i] = ops;
|
|
break;
|
|
}
|
|
}
|
|
|
|
return ret;
|
|
}
|
|
|
|
void sbi_ipi_event_destroy(u32 event)
|
|
{
|
|
if (SBI_IPI_EVENT_MAX <= event)
|
|
return;
|
|
|
|
ipi_ops_array[event] = NULL;
|
|
}
|
|
|
|
static void sbi_ipi_process_smode(struct sbi_scratch *scratch)
|
|
{
|
|
csr_set(CSR_MIP, MIP_SSIP);
|
|
}
|
|
|
|
static struct sbi_ipi_event_ops ipi_smode_ops = {
|
|
.name = "IPI_SMODE",
|
|
.process = sbi_ipi_process_smode,
|
|
};
|
|
|
|
static u32 ipi_smode_event = SBI_IPI_EVENT_MAX;
|
|
|
|
int sbi_ipi_send_smode(ulong hmask, ulong hbase)
|
|
{
|
|
return sbi_ipi_send_many(hmask, hbase, ipi_smode_event, NULL);
|
|
}
|
|
|
|
void sbi_ipi_clear_smode(void)
|
|
{
|
|
csr_clear(CSR_MIP, MIP_SSIP);
|
|
}
|
|
|
|
static void sbi_ipi_process_halt(struct sbi_scratch *scratch)
|
|
{
|
|
sbi_hsm_hart_stop(scratch, true);
|
|
}
|
|
|
|
static struct sbi_ipi_event_ops ipi_halt_ops = {
|
|
.name = "IPI_HALT",
|
|
.process = sbi_ipi_process_halt,
|
|
};
|
|
|
|
static u32 ipi_halt_event = SBI_IPI_EVENT_MAX;
|
|
|
|
int sbi_ipi_send_halt(ulong hmask, ulong hbase)
|
|
{
|
|
return sbi_ipi_send_many(hmask, hbase, ipi_halt_event, NULL);
|
|
}
|
|
|
|
void sbi_ipi_process(void)
|
|
{
|
|
unsigned long ipi_type;
|
|
unsigned int ipi_event;
|
|
const struct sbi_ipi_event_ops *ipi_ops;
|
|
struct sbi_scratch *scratch = sbi_scratch_thishart_ptr();
|
|
struct sbi_ipi_data *ipi_data =
|
|
sbi_scratch_offset_ptr(scratch, ipi_data_off);
|
|
u32 hartid = current_hartid();
|
|
|
|
sbi_pmu_ctr_incr_fw(SBI_PMU_FW_IPI_RECVD);
|
|
if (ipi_dev && ipi_dev->ipi_clear)
|
|
ipi_dev->ipi_clear(hartid);
|
|
|
|
ipi_type = atomic_raw_xchg_ulong(&ipi_data->ipi_type, 0);
|
|
ipi_event = 0;
|
|
while (ipi_type) {
|
|
if (!(ipi_type & 1UL))
|
|
goto skip;
|
|
|
|
ipi_ops = ipi_ops_array[ipi_event];
|
|
if (ipi_ops && ipi_ops->process)
|
|
ipi_ops->process(scratch);
|
|
|
|
skip:
|
|
ipi_type = ipi_type >> 1;
|
|
ipi_event++;
|
|
};
|
|
}
|
|
|
|
int sbi_ipi_raw_send(u32 target_hart)
|
|
{
|
|
if (!ipi_dev || !ipi_dev->ipi_send)
|
|
return SBI_EINVAL;
|
|
|
|
ipi_dev->ipi_send(target_hart);
|
|
return 0;
|
|
}
|
|
|
|
void sbi_ipi_raw_clear(u32 target_hart)
|
|
{
|
|
if (ipi_dev && ipi_dev->ipi_clear)
|
|
ipi_dev->ipi_clear(target_hart);
|
|
}
|
|
|
|
const struct sbi_ipi_device *sbi_ipi_get_device(void)
|
|
{
|
|
return ipi_dev;
|
|
}
|
|
|
|
void sbi_ipi_set_device(const struct sbi_ipi_device *dev)
|
|
{
|
|
if (!dev || ipi_dev)
|
|
return;
|
|
|
|
ipi_dev = dev;
|
|
}
|
|
|
|
int sbi_ipi_init(struct sbi_scratch *scratch, bool cold_boot)
|
|
{
|
|
int ret;
|
|
struct sbi_ipi_data *ipi_data;
|
|
|
|
if (cold_boot) {
|
|
ipi_data_off = sbi_scratch_alloc_offset(sizeof(*ipi_data));
|
|
if (!ipi_data_off)
|
|
return SBI_ENOMEM;
|
|
ret = sbi_ipi_event_create(&ipi_smode_ops);
|
|
if (ret < 0)
|
|
return ret;
|
|
ipi_smode_event = ret;
|
|
ret = sbi_ipi_event_create(&ipi_halt_ops);
|
|
if (ret < 0)
|
|
return ret;
|
|
ipi_halt_event = ret;
|
|
} else {
|
|
if (!ipi_data_off)
|
|
return SBI_ENOMEM;
|
|
if (SBI_IPI_EVENT_MAX <= ipi_smode_event ||
|
|
SBI_IPI_EVENT_MAX <= ipi_halt_event)
|
|
return SBI_ENOSPC;
|
|
}
|
|
|
|
ipi_data = sbi_scratch_offset_ptr(scratch, ipi_data_off);
|
|
ipi_data->ipi_type = 0x00;
|
|
|
|
/*
|
|
* Initialize platform IPI support. This will also clear any
|
|
* pending IPIs for current/calling HART.
|
|
*/
|
|
ret = sbi_platform_ipi_init(sbi_platform_ptr(scratch), cold_boot);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* Enable software interrupts */
|
|
csr_set(CSR_MIE, MIP_MSIP);
|
|
|
|
return 0;
|
|
}
|
|
|
|
void sbi_ipi_exit(struct sbi_scratch *scratch)
|
|
{
|
|
/* Disable software interrupts */
|
|
csr_clear(CSR_MIE, MIP_MSIP);
|
|
|
|
/* Process pending IPIs */
|
|
sbi_ipi_process();
|
|
|
|
/* Platform exit */
|
|
sbi_platform_ipi_exit(sbi_platform_ptr(scratch));
|
|
}
|