Page Menu
Home
FreeBSD
Search
Configure Global Search
Log In
Files
F109355691
D16861.id.diff
No One
Temporary
Actions
View File
Edit File
Delete File
View Transforms
Subscribe
Mute Notifications
Flag For Later
Award Token
Size
39 KB
Referenced Files
None
Subscribers
None
D16861.id.diff
View Options
Index: head/sys/sys/interrupt.h
===================================================================
--- head/sys/sys/interrupt.h
+++ head/sys/sys/interrupt.h
@@ -154,8 +154,13 @@
extern void *vm_ih;
/* Counts and names for statistics (defined in MD code). */
+#if defined(__amd64__) || defined(__i386__)
+extern u_long *intrcnt; /* counts for for each device and stray */
+extern char *intrnames; /* string table containing device names */
+#else
extern u_long intrcnt[]; /* counts for for each device and stray */
extern char intrnames[]; /* string table containing device names */
+#endif
extern size_t sintrcnt; /* size of intrcnt table */
extern size_t sintrnames; /* size of intrnames table */
Index: head/sys/x86/acpica/madt.c
===================================================================
--- head/sys/x86/acpica/madt.c
+++ head/sys/x86/acpica/madt.c
@@ -428,10 +428,6 @@
apic->Id);
if (ioapics[apic->Id].io_apic != NULL)
panic("%s: Double APIC ID %u", __func__, apic->Id);
- if (apic->GlobalIrqBase >= FIRST_MSI_INT) {
- printf("MADT: Ignoring bogus I/O APIC ID %u", apic->Id);
- break;
- }
ioapics[apic->Id].io_apic = ioapic_create(apic->Address,
apic->Id, apic->GlobalIrqBase);
ioapics[apic->Id].io_vector = apic->GlobalIrqBase;
Index: head/sys/x86/include/apicvar.h
===================================================================
--- head/sys/x86/include/apicvar.h
+++ head/sys/x86/include/apicvar.h
@@ -158,10 +158,10 @@
#define APIC_BUS_PCI 2
#define APIC_BUS_MAX APIC_BUS_PCI
-#define IRQ_EXTINT (NUM_IO_INTS + 1)
-#define IRQ_NMI (NUM_IO_INTS + 2)
-#define IRQ_SMI (NUM_IO_INTS + 3)
-#define IRQ_DISABLED (NUM_IO_INTS + 4)
+#define IRQ_EXTINT -1
+#define IRQ_NMI -2
+#define IRQ_SMI -3
+#define IRQ_DISABLED -4
/*
* An APIC enumerator is a pseudo bus driver that enumerates APIC's including
Index: head/sys/x86/include/intr_machdep.h
===================================================================
--- head/sys/x86/include/intr_machdep.h
+++ head/sys/x86/include/intr_machdep.h
@@ -34,55 +34,41 @@
#ifdef _KERNEL
/*
- * The maximum number of I/O interrupts we allow. This number is rather
- * arbitrary as it is just the maximum IRQ resource value. The interrupt
- * source for a given IRQ maps that I/O interrupt to device interrupt
- * source whether it be a pin on an interrupt controller or an MSI interrupt.
- * The 16 ISA IRQs are assigned fixed IDT vectors, but all other device
- * interrupts allocate IDT vectors on demand. Currently we have 191 IDT
- * vectors available for device interrupts. On many systems with I/O APICs,
- * a lot of the IRQs are not used, so this number can be much larger than
- * 191 and still be safe since only interrupt sources in actual use will
- * allocate IDT vectors.
+ * Values used in determining the allocation of IRQ values among
+ * different types of I/O interrupts. These values are used as
+ * indices into a interrupt source array to map I/O interrupts to a
+ * device interrupt source whether it be a pin on an interrupt
+ * controller or an MSI interrupt. The 16 ISA IRQs are assigned fixed
+ * IDT vectors, but all other device interrupts allocate IDT vectors
+ * on demand. Currently we have 191 IDT vectors available for device
+ * interrupts on each CPU. On many systems with I/O APICs, a lot of
+ * the IRQs are not used, so the total number of IRQ values reserved
+ * can exceed the number of available IDT slots.
*
- * The first 255 IRQs (0 - 254) are reserved for ISA IRQs and PCI intline IRQs.
- * IRQ values from 256 to 767 are used by MSI. When running under the Xen
- * Hypervisor, IRQ values from 768 to 4863 are available for binding to
- * event channel events. We leave 255 unused to avoid confusion since 255 is
- * used in PCI to indicate an invalid IRQ.
+ * The first 16 IRQs (0 - 15) are reserved for ISA IRQs. Interrupt
+ * pins on I/O APICs for non-ISA interrupts use IRQ values starting at
+ * IRQ 17. This layout matches the GSI numbering used by ACPI so that
+ * IRQ values returned by ACPI methods such as _CRS can be used
+ * directly by the ACPI bus driver.
+ *
+ * MSI interrupts allocate a block of interrupts starting at either
+ * the end of the I/O APIC range or 256, whichever is higher. When
+ * running under the Xen Hypervisor, an additional range of IRQ values
+ * are available for binding to event channel events. We use 256 as
+ * the minimum IRQ value for MSI interrupts to attempt to leave 255
+ * unused since 255 is used in PCI to indicate an invalid INTx IRQ.
*/
#define NUM_MSI_INTS 512
-#define FIRST_MSI_INT 256
-#ifdef XENHVM
-#include <xen/xen-os.h>
-#include <xen/interface/event_channel.h>
-#define NUM_EVTCHN_INTS NR_EVENT_CHANNELS
-#define FIRST_EVTCHN_INT \
- (FIRST_MSI_INT + NUM_MSI_INTS)
-#define LAST_EVTCHN_INT \
- (FIRST_EVTCHN_INT + NUM_EVTCHN_INTS - 1)
-#else
-#define NUM_EVTCHN_INTS 0
-#endif
-#define NUM_IO_INTS (FIRST_MSI_INT + NUM_MSI_INTS + NUM_EVTCHN_INTS)
+#define MINIMUM_MSI_INT 256
+extern u_int first_msi_irq;
+extern u_int num_io_irqs;
+
/*
* Default base address for MSI messages on x86 platforms.
*/
#define MSI_INTEL_ADDR_BASE 0xfee00000
-/*
- * - 1 ??? dummy counter.
- * - 2 counters for each I/O interrupt.
- * - 1 counter for each CPU for lapic timer.
- * - 8 counters for each CPU for IPI counters for SMP.
- */
-#ifdef SMP
-#define INTRCNT_COUNT (1 + NUM_IO_INTS * 2 + (1 + 8) * MAXCPU)
-#else
-#define INTRCNT_COUNT (1 + NUM_IO_INTS * 2 + 1)
-#endif
-
#ifndef LOCORE
typedef void inthand_t(void);
@@ -97,6 +83,7 @@
* return the vector associated with this source.
*/
struct pic {
+ void (*pic_register_sources)(struct pic *);
void (*pic_enable_source)(struct intsrc *);
void (*pic_disable_source)(struct intsrc *, int);
void (*pic_eoi_source)(struct intsrc *);
@@ -184,6 +171,9 @@
int msi_release(int *irqs, int count);
int msix_alloc(device_t dev, int *irq);
int msix_release(int irq);
+#ifdef XENHVM
+void xen_intr_alloc_irqs(void);
+#endif
#endif /* !LOCORE */
#endif /* _KERNEL */
Index: head/sys/x86/iommu/intel_intrmap.c
===================================================================
--- head/sys/x86/iommu/intel_intrmap.c
+++ head/sys/x86/iommu/intel_intrmap.c
@@ -337,7 +337,7 @@
"QI disabled, disabling interrupt remapping\n");
return (0);
}
- unit->irte_cnt = clp2(NUM_IO_INTS);
+ unit->irte_cnt = clp2(num_io_irqs);
unit->irt = (dmar_irte_t *)(uintptr_t)kmem_alloc_contig(
unit->irte_cnt * sizeof(dmar_irte_t), M_ZERO | M_WAITOK, 0,
dmar_high, PAGE_SIZE, 0, DMAR_IS_COHERENT(unit) ?
Index: head/sys/x86/isa/atpic.c
===================================================================
--- head/sys/x86/isa/atpic.c
+++ head/sys/x86/isa/atpic.c
@@ -95,6 +95,7 @@
#define ATPIC(io, base, eoi) { \
.at_pic = { \
+ .pic_register_sources = atpic_register_sources, \
.pic_enable_source = atpic_enable_source, \
.pic_disable_source = atpic_disable_source, \
.pic_eoi_source = (eoi), \
@@ -133,6 +134,7 @@
u_long at_straycount;
};
+static void atpic_register_sources(struct pic *pic);
static void atpic_enable_source(struct intsrc *isrc);
static void atpic_disable_source(struct intsrc *isrc, int eoi);
static void atpic_eoi_master(struct intsrc *isrc);
@@ -203,6 +205,36 @@
}
static void
+atpic_register_sources(struct pic *pic)
+{
+ struct atpic *ap = (struct atpic *)pic;
+ struct atpic_intsrc *ai;
+ int i;
+
+ /*
+ * If any of the ISA IRQs have an interrupt source already, then
+ * assume that the I/O APICs are being used and don't register any
+ * of our interrupt sources. This makes sure we don't accidentally
+ * use mixed mode. The "accidental" use could otherwise occur on
+ * machines that route the ACPI SCI interrupt to a different ISA
+ * IRQ (at least one machine routes it to IRQ 13) thus disabling
+ * that APIC ISA routing and allowing the ATPIC source for that IRQ
+ * to leak through. We used to depend on this feature for routing
+ * IRQ0 via mixed mode, but now we don't use mixed mode at all.
+ */
+ for (i = 0; i < NUM_ISA_IRQS; i++)
+ if (intr_lookup_source(i) != NULL)
+ return;
+
+ /* Loop through all interrupt sources and add them. */
+ for (i = 0, ai = atintrs + ap->at_irqbase; i < 8; i++, ai++) {
+ if (ap->at_irqbase + i == ICU_SLAVEID)
+ continue;
+ intr_register_source(&ai->at_intsrc);
+ }
+}
+
+static void
atpic_enable_source(struct intsrc *isrc)
{
struct atpic_intsrc *ai = (struct atpic_intsrc *)isrc;
@@ -467,8 +499,6 @@
static void
atpic_init(void *dummy __unused)
{
- struct atpic_intsrc *ai;
- int i;
/*
* Register our PICs, even if we aren't going to use any of their
@@ -478,27 +508,8 @@
intr_register_pic(&atpics[1].at_pic) != 0)
panic("Unable to register ATPICs");
- /*
- * If any of the ISA IRQs have an interrupt source already, then
- * assume that the APICs are being used and don't register any
- * of our interrupt sources. This makes sure we don't accidentally
- * use mixed mode. The "accidental" use could otherwise occur on
- * machines that route the ACPI SCI interrupt to a different ISA
- * IRQ (at least one machines routes it to IRQ 13) thus disabling
- * that APIC ISA routing and allowing the ATPIC source for that IRQ
- * to leak through. We used to depend on this feature for routing
- * IRQ0 via mixed mode, but now we don't use mixed mode at all.
- */
- for (i = 0; i < NUM_ISA_IRQS; i++)
- if (intr_lookup_source(i) != NULL)
- return;
-
- /* Loop through all interrupt sources and add them. */
- for (i = 0, ai = atintrs; i < NUM_ISA_IRQS; i++, ai++) {
- if (i == ICU_SLAVEID)
- continue;
- intr_register_source(&ai->at_intsrc);
- }
+ if (num_io_irqs == 0)
+ num_io_irqs = NUM_ISA_IRQS;
}
SYSINIT(atpic_init, SI_SUB_INTR, SI_ORDER_FOURTH, atpic_init, NULL);
Index: head/sys/x86/x86/intr_machdep.c
===================================================================
--- head/sys/x86/x86/intr_machdep.c
+++ head/sys/x86/x86/intr_machdep.c
@@ -38,6 +38,7 @@
#include "opt_atpic.h"
#include "opt_ddb.h"
+#include "opt_smp.h"
#include <sys/param.h>
#include <sys/bus.h>
@@ -45,6 +46,7 @@
#include <sys/ktr.h>
#include <sys/kernel.h>
#include <sys/lock.h>
+#include <sys/malloc.h>
#include <sys/mutex.h>
#include <sys/proc.h>
#include <sys/queue.h>
@@ -78,10 +80,9 @@
typedef void (*mask_fn)(void *);
static int intrcnt_index;
-static struct intsrc *interrupt_sources[NUM_IO_INTS];
+static struct intsrc **interrupt_sources;
#ifdef SMP
-static struct intsrc *interrupt_sorted[NUM_IO_INTS];
-CTASSERT(sizeof(interrupt_sources) == sizeof(interrupt_sorted));
+static struct intsrc **interrupt_sorted;
static int intrbalance;
SYSCTL_INT(_hw, OID_AUTO, intrbalance, CTLFLAG_RW, &intrbalance, 0,
"Interrupt auto-balance interval (seconds). Zero disables.");
@@ -91,16 +92,20 @@
static struct mtx intrpic_lock;
static struct mtx intrcnt_lock;
static TAILQ_HEAD(pics_head, pic) pics;
+u_int num_io_irqs;
#if defined(SMP) && !defined(EARLY_AP_STARTUP)
static int assign_cpu;
#endif
-u_long intrcnt[INTRCNT_COUNT];
-char intrnames[INTRCNT_COUNT * (MAXCOMLEN + 1)];
+u_long *intrcnt;
+char *intrnames;
size_t sintrcnt = sizeof(intrcnt);
size_t sintrnames = sizeof(intrnames);
+int nintrcnt;
+static MALLOC_DEFINE(M_INTR, "intr", "Interrupt Sources");
+
static int intr_assign_cpu(void *arg, int cpu);
static void intr_disable_src(void *arg);
static void intr_init(void *__dummy);
@@ -109,6 +114,18 @@
static void intrcnt_updatename(struct intsrc *is);
static void intrcnt_register(struct intsrc *is);
+/*
+ * SYSINIT levels for SI_SUB_INTR:
+ *
+ * SI_ORDER_FIRST: Initialize locks and pics TAILQ, xen_hvm_cpu_init
+ * SI_ORDER_SECOND: Xen PICs
+ * SI_ORDER_THIRD: Add I/O APIC PICs, alloc MSI and Xen IRQ ranges
+ * SI_ORDER_FOURTH: Add 8259A PICs
+ * SI_ORDER_FOURTH + 1: Finalize interrupt count and add interrupt sources
+ * SI_ORDER_MIDDLE: SMP interrupt counters
+ * SI_ORDER_ANY: Enable interrupts on BSP
+ */
+
static int
intr_pic_registered(struct pic *pic)
{
@@ -144,6 +161,58 @@
}
/*
+ * Allocate interrupt source arrays and register interrupt sources
+ * once the number of interrupts is known.
+ */
+static void
+intr_init_sources(void *arg)
+{
+ struct pic *pic;
+
+ MPASS(num_io_irqs > 0);
+
+ interrupt_sources = mallocarray(num_io_irqs, sizeof(*interrupt_sources),
+ M_INTR, M_WAITOK | M_ZERO);
+ interrupt_sorted = mallocarray(num_io_irqs, sizeof(*interrupt_sorted),
+ M_INTR, M_WAITOK | M_ZERO);
+
+ /*
+ * - 1 ??? dummy counter.
+ * - 2 counters for each I/O interrupt.
+ * - 1 counter for each CPU for lapic timer.
+ * - 1 counter for each CPU for the Hyper-V vmbus driver.
+ * - 8 counters for each CPU for IPI counters for SMP.
+ */
+ nintrcnt = 1 + num_io_irqs * 2 + mp_ncpus * 2;
+#ifdef COUNT_IPIS
+ if (mp_ncpus > 1)
+ nintrcnt += 8 * mp_ncpus;
+#endif
+ intrcnt = mallocarray(nintrcnt, sizeof(u_long), M_INTR, M_WAITOK |
+ M_ZERO);
+ intrnames = mallocarray(nintrcnt, MAXCOMLEN + 1, M_INTR, M_WAITOK |
+ M_ZERO);
+ sintrcnt = nintrcnt * sizeof(u_long);
+ sintrnames = nintrcnt * (MAXCOMLEN + 1);
+
+ intrcnt_setname("???", 0);
+ intrcnt_index = 1;
+
+ /*
+ * NB: intrpic_lock is not held here to avoid LORs due to
+ * malloc() in intr_register_source(). However, we are still
+ * single-threaded at this point in startup so the list of
+ * PICs shouldn't change.
+ */
+ TAILQ_FOREACH(pic, &pics, pics) {
+ if (pic->pic_register_sources != NULL)
+ pic->pic_register_sources(pic);
+ }
+}
+SYSINIT(intr_init_sources, SI_SUB_INTR, SI_ORDER_FOURTH + 1, intr_init_sources,
+ NULL);
+
+/*
* Register a new interrupt source with the global interrupt system.
* The global interrupts need to be disabled when this function is
* called.
@@ -155,6 +224,8 @@
KASSERT(intr_pic_registered(isrc->is_pic), ("unregistered PIC"));
vector = isrc->is_pic->pic_vector(isrc);
+ KASSERT(vector < num_io_irqs, ("IRQ %d too large (%u irqs)", vector,
+ num_io_irqs));
if (interrupt_sources[vector] != NULL)
return (EEXIST);
error = intr_event_create(&isrc->is_event, isrc, 0, vector,
@@ -180,7 +251,7 @@
intr_lookup_source(int vector)
{
- if (vector < 0 || vector >= nitems(interrupt_sources))
+ if (vector < 0 || vector >= num_io_irqs)
return (NULL);
return (interrupt_sources[vector]);
}
@@ -378,6 +449,7 @@
KASSERT(is->is_event != NULL, ("%s: isrc with no event", __func__));
mtx_lock_spin(&intrcnt_lock);
+ MPASS(intrcnt_index + 2 <= nintrcnt);
is->is_index = intrcnt_index;
intrcnt_index += 2;
snprintf(straystr, MAXCOMLEN + 1, "stray irq%d",
@@ -394,6 +466,7 @@
{
mtx_lock_spin(&intrcnt_lock);
+ MPASS(intrcnt_index < nintrcnt);
*countp = &intrcnt[intrcnt_index];
intrcnt_setname(name, intrcnt_index);
intrcnt_index++;
@@ -404,8 +477,6 @@
intr_init(void *dummy __unused)
{
- intrcnt_setname("???", 0);
- intrcnt_index = 1;
TAILQ_INIT(&pics);
mtx_init(&intrpic_lock, "intrpic", NULL, MTX_DEF);
sx_init(&intrsrc_lock, "intrsrc");
@@ -471,10 +542,10 @@
intr_reprogram(void)
{
struct intsrc *is;
- int v;
+ u_int v;
sx_xlock(&intrsrc_lock);
- for (v = 0; v < NUM_IO_INTS; v++) {
+ for (v = 0; v < num_io_irqs; v++) {
is = interrupt_sources[v];
if (is == NULL)
continue;
@@ -491,14 +562,15 @@
DB_SHOW_COMMAND(irqs, db_show_irqs)
{
struct intsrc **isrc;
- int i, verbose;
+ u_int i;
+ int verbose;
if (strcmp(modif, "v") == 0)
verbose = 1;
else
verbose = 0;
isrc = interrupt_sources;
- for (i = 0; i < NUM_IO_INTS && !db_pager_quit; i++, isrc++)
+ for (i = 0; i < num_io_irqs && !db_pager_quit; i++, isrc++)
if (*isrc != NULL)
db_dump_intr_event((*isrc)->is_event, verbose);
}
@@ -606,8 +678,7 @@
intr_shuffle_irqs(void *arg __unused)
{
struct intsrc *isrc;
- u_int cpu;
- int i;
+ u_int cpu, i;
intr_init_cpus();
/* Don't bother on UP. */
@@ -617,7 +688,7 @@
/* Round-robin assign a CPU to each enabled source. */
sx_xlock(&intrsrc_lock);
assign_cpu = 1;
- for (i = 0; i < NUM_IO_INTS; i++) {
+ for (i = 0; i < num_io_irqs; i++) {
isrc = interrupt_sources[i];
if (isrc != NULL && isrc->is_handlers > 0) {
/*
@@ -652,8 +723,8 @@
{
struct sbuf sbuf;
struct intsrc *isrc;
+ u_int i;
int error;
- int i;
error = sysctl_wire_old_buffer(req, 0);
if (error != 0)
@@ -661,7 +732,7 @@
sbuf_new_for_sysctl(&sbuf, NULL, 128, req);
sx_slock(&intrsrc_lock);
- for (i = 0; i < NUM_IO_INTS; i++) {
+ for (i = 0; i < num_io_irqs; i++) {
isrc = interrupt_sources[i];
if (isrc == NULL)
continue;
@@ -720,8 +791,9 @@
* Sort interrupts according to count.
*/
sx_xlock(&intrsrc_lock);
- memcpy(interrupt_sorted, interrupt_sources, sizeof(interrupt_sorted));
- qsort(interrupt_sorted, NUM_IO_INTS, sizeof(interrupt_sorted[0]),
+ memcpy(interrupt_sorted, interrupt_sources, num_io_irqs *
+ sizeof(interrupt_sorted[0]));
+ qsort(interrupt_sorted, num_io_irqs, sizeof(interrupt_sorted[0]),
intrcmp);
/*
@@ -733,7 +805,7 @@
/*
* Assign round-robin from most loaded to least.
*/
- for (i = NUM_IO_INTS - 1; i >= 0; i--) {
+ for (i = num_io_irqs - 1; i >= 0; i--) {
isrc = interrupt_sorted[i];
if (isrc == NULL || isrc->is_event->ie_cpu != NOCPU)
continue;
Index: head/sys/x86/x86/io_apic.c
===================================================================
--- head/sys/x86/x86/io_apic.c
+++ head/sys/x86/x86/io_apic.c
@@ -80,7 +80,7 @@
struct ioapic_intsrc {
struct intsrc io_intsrc;
- u_int io_irq;
+ int io_irq;
u_int io_intpin:8;
u_int io_vector:8;
u_int io_cpu;
@@ -112,6 +112,7 @@
static void ioapic_write(volatile ioapic_t *apic, int reg, u_int val);
static const char *ioapic_bus_string(int bus_type);
static void ioapic_print_irq(struct ioapic_intsrc *intpin);
+static void ioapic_register_sources(struct pic *pic);
static void ioapic_enable_source(struct intsrc *isrc);
static void ioapic_disable_source(struct intsrc *isrc, int eoi);
static void ioapic_eoi_source(struct intsrc *isrc);
@@ -128,6 +129,7 @@
static STAILQ_HEAD(,ioapic) ioapic_list = STAILQ_HEAD_INITIALIZER(ioapic_list);
struct pic ioapic_template = {
+ .pic_register_sources = ioapic_register_sources,
.pic_enable_source = ioapic_enable_source,
.pic_disable_source = ioapic_disable_source,
.pic_eoi_source = ioapic_eoi_source,
@@ -142,7 +144,7 @@
.pic_reprogram_pin = ioapic_reprogram_intpin,
};
-static int next_ioapic_base;
+static u_int next_ioapic_base;
static u_int next_id;
static int enable_extint;
@@ -250,7 +252,7 @@
printf("SMI");
break;
default:
- printf("%s IRQ %u", ioapic_bus_string(intpin->io_bus),
+ printf("%s IRQ %d", ioapic_bus_string(intpin->io_bus),
intpin->io_irq);
}
}
@@ -318,7 +320,7 @@
* been enabled yet, just ensure that the pin is masked.
*/
mtx_assert(&icu_lock, MA_OWNED);
- if (intpin->io_irq == IRQ_DISABLED || (intpin->io_irq < NUM_IO_INTS &&
+ if (intpin->io_irq == IRQ_DISABLED || (intpin->io_irq >= 0 &&
intpin->io_vector == 0)) {
low = ioapic_read(io->io_addr,
IOAPIC_REDTBL_LO(intpin->io_intpin));
@@ -651,6 +653,8 @@
io->io_id, intbase, next_ioapic_base);
io->io_intbase = intbase;
next_ioapic_base = intbase + numintr;
+ if (next_ioapic_base > num_io_irqs)
+ num_io_irqs = next_ioapic_base;
io->io_numintr = numintr;
io->io_addr = apic;
io->io_paddr = addr;
@@ -759,7 +763,7 @@
io = (struct ioapic *)cookie;
if (pin >= io->io_numintr || vector < 0)
return (EINVAL);
- if (io->io_pins[pin].io_irq >= NUM_IO_INTS)
+ if (io->io_pins[pin].io_irq < 0)
return (EINVAL);
io->io_pins[pin].io_irq = vector;
if (bootverbose)
@@ -778,7 +782,7 @@
io = (struct ioapic *)cookie;
if (pin >= io->io_numintr)
return (EINVAL);
- if (io->io_pins[pin].io_irq >= NUM_IO_INTS)
+ if (io->io_pins[pin].io_irq < 0)
return (EINVAL);
if (io->io_pins[pin].io_bus == bus_type)
return (0);
@@ -799,7 +803,7 @@
return (EINVAL);
if (io->io_pins[pin].io_irq == IRQ_NMI)
return (0);
- if (io->io_pins[pin].io_irq >= NUM_IO_INTS)
+ if (io->io_pins[pin].io_irq < 0)
return (EINVAL);
io->io_pins[pin].io_bus = APIC_BUS_UNKNOWN;
io->io_pins[pin].io_irq = IRQ_NMI;
@@ -822,7 +826,7 @@
return (EINVAL);
if (io->io_pins[pin].io_irq == IRQ_SMI)
return (0);
- if (io->io_pins[pin].io_irq >= NUM_IO_INTS)
+ if (io->io_pins[pin].io_irq < 0)
return (EINVAL);
io->io_pins[pin].io_bus = APIC_BUS_UNKNOWN;
io->io_pins[pin].io_irq = IRQ_SMI;
@@ -845,7 +849,7 @@
return (EINVAL);
if (io->io_pins[pin].io_irq == IRQ_EXTINT)
return (0);
- if (io->io_pins[pin].io_irq >= NUM_IO_INTS)
+ if (io->io_pins[pin].io_irq < 0)
return (EINVAL);
io->io_pins[pin].io_bus = APIC_BUS_UNKNOWN;
io->io_pins[pin].io_irq = IRQ_EXTINT;
@@ -870,7 +874,7 @@
io = (struct ioapic *)cookie;
if (pin >= io->io_numintr || pol == INTR_POLARITY_CONFORM)
return (EINVAL);
- if (io->io_pins[pin].io_irq >= NUM_IO_INTS)
+ if (io->io_pins[pin].io_irq < 0)
return (EINVAL);
activehi = (pol == INTR_POLARITY_HIGH);
if (io->io_pins[pin].io_activehi == activehi)
@@ -891,7 +895,7 @@
io = (struct ioapic *)cookie;
if (pin >= io->io_numintr || trigger == INTR_TRIGGER_CONFORM)
return (EINVAL);
- if (io->io_pins[pin].io_irq >= NUM_IO_INTS)
+ if (io->io_pins[pin].io_irq < 0)
return (EINVAL);
edgetrigger = (trigger == INTR_TRIGGER_EDGE);
if (io->io_pins[pin].io_edgetrigger == edgetrigger)
@@ -927,12 +931,26 @@
/*
* Reprogram pins to handle special case pins (such as NMI and
- * SMI) and register valid pins as interrupt sources.
+ * SMI) and disable normal pins until a handler is registered.
*/
intr_register_pic(&io->io_pic);
- for (i = 0, pin = io->io_pins; i < io->io_numintr; i++, pin++) {
+ for (i = 0, pin = io->io_pins; i < io->io_numintr; i++, pin++)
ioapic_reprogram_intpin(&pin->io_intsrc);
- if (pin->io_irq < NUM_IO_INTS)
+}
+
+/*
+ * Add interrupt sources for I/O APIC interrupt pins.
+ */
+static void
+ioapic_register_sources(struct pic *pic)
+{
+ struct ioapic_intsrc *pin;
+ struct ioapic *io;
+ int i;
+
+ io = (struct ioapic *)pic;
+ for (i = 0, pin = io->io_pins; i < io->io_numintr; i++, pin++) {
+ if (pin->io_irq >= 0)
intr_register_source(&pin->io_intsrc);
}
}
Index: head/sys/x86/x86/local_apic.c
===================================================================
--- head/sys/x86/x86/local_apic.c
+++ head/sys/x86/x86/local_apic.c
@@ -92,11 +92,16 @@
CTASSERT(APIC_LOCAL_INTS == 240);
CTASSERT(IPI_STOP < APIC_SPURIOUS_INT);
-/* Magic IRQ values for the timer and syscalls. */
-#define IRQ_TIMER (NUM_IO_INTS + 1)
-#define IRQ_SYSCALL (NUM_IO_INTS + 2)
-#define IRQ_DTRACE_RET (NUM_IO_INTS + 3)
-#define IRQ_EVTCHN (NUM_IO_INTS + 4)
+/*
+ * I/O interrupts use non-negative IRQ values. These values are used
+ * to mark unused IDT entries or IDT entries reserved for a non-I/O
+ * interrupt.
+ */
+#define IRQ_FREE -1
+#define IRQ_TIMER -2
+#define IRQ_SYSCALL -3
+#define IRQ_DTRACE_RET -4
+#define IRQ_EVTCHN -5
enum lat_timer_mode {
LAT_MODE_UNDEF = 0,
@@ -648,7 +653,7 @@
lapics[apic_id].la_elvts[i].lvt_active = 0;
}
for (i = 0; i <= APIC_NUM_IOINTS; i++)
- lapics[apic_id].la_ioint_irqs[i] = -1;
+ lapics[apic_id].la_ioint_irqs[i] = IRQ_FREE;
lapics[apic_id].la_ioint_irqs[IDT_SYSCALL - APIC_IO_INTS] = IRQ_SYSCALL;
lapics[apic_id].la_ioint_irqs[APIC_TIMER_INT - APIC_IO_INTS] =
IRQ_TIMER;
@@ -751,7 +756,6 @@
uint32_t version;
uint32_t maxlvt;
register_t saveintr;
- char buf[MAXCOMLEN + 1];
int elvt_count;
int i;
@@ -780,15 +784,11 @@
LAPIC_LVT_PCINT));
}
- /* Program timer LVT and setup handler. */
+ /* Program timer LVT. */
la->lvt_timer_base = lvt_mode(la, APIC_LVT_TIMER,
lapic_read32(LAPIC_LVT_TIMER));
la->lvt_timer_last = la->lvt_timer_base;
lapic_write32(LAPIC_LVT_TIMER, la->lvt_timer_base);
- if (boot) {
- snprintf(buf, sizeof(buf), "cpu%d:timer", PCPU_GET(cpuid));
- intrcnt_add(buf, &la->la_timer_count);
- }
/* Calibrate the timer parameters using BSP. */
if (boot && IS_BSP()) {
@@ -843,6 +843,28 @@
}
static void
+native_lapic_intrcnt(void *dummy __unused)
+{
+ struct pcpu *pc;
+ struct lapic *la;
+ char buf[MAXCOMLEN + 1];
+
+ /* If there are no APICs, skip this function. */
+ if (lapics == NULL)
+ return;
+
+ STAILQ_FOREACH(pc, &cpuhead, pc_allcpu) {
+ la = &lapics[pc->pc_apic_id];
+ KASSERT(la->la_present, ("missing APIC structure"));
+
+ snprintf(buf, sizeof(buf), "cpu%d:timer", pc->pc_cpuid);
+ intrcnt_add(buf, &la->la_timer_count);
+ }
+}
+SYSINIT(native_lapic_intrcnt, SI_SUB_INTR, SI_ORDER_MIDDLE, native_lapic_intrcnt,
+ NULL);
+
+static void
native_lapic_reenable_pmc(void)
{
#ifdef HWPMC_HOOKS
@@ -1493,7 +1515,7 @@
{
u_int vector;
- KASSERT(irq < NUM_IO_INTS, ("Invalid IRQ %u", irq));
+ KASSERT(irq < num_io_irqs, ("Invalid IRQ %u", irq));
/*
* Search for a free vector. Currently we just use a very simple
@@ -1501,7 +1523,7 @@
*/
mtx_lock_spin(&icu_lock);
for (vector = 0; vector < APIC_NUM_IOINTS; vector++) {
- if (lapics[apic_id].la_ioint_irqs[vector] != -1)
+ if (lapics[apic_id].la_ioint_irqs[vector] != IRQ_FREE)
continue;
lapics[apic_id].la_ioint_irqs[vector] = irq;
mtx_unlock_spin(&icu_lock);
@@ -1527,7 +1549,7 @@
KASSERT(align >= count, ("align < count"));
#ifdef INVARIANTS
for (run = 0; run < count; run++)
- KASSERT(irqs[run] < NUM_IO_INTS, ("Invalid IRQ %u at index %u",
+ KASSERT(irqs[run] < num_io_irqs, ("Invalid IRQ %u at index %u",
irqs[run], run));
#endif
@@ -1541,7 +1563,7 @@
for (vector = 0; vector < APIC_NUM_IOINTS; vector++) {
/* Vector is in use, end run. */
- if (lapics[apic_id].la_ioint_irqs[vector] != -1) {
+ if (lapics[apic_id].la_ioint_irqs[vector] != IRQ_FREE) {
run = 0;
first = 0;
continue;
@@ -1622,7 +1644,7 @@
KASSERT(vector >= APIC_IO_INTS && vector != IDT_SYSCALL &&
vector <= APIC_IO_INTS + APIC_NUM_IOINTS,
("Vector %u does not map to an IRQ line", vector));
- KASSERT(irq < NUM_IO_INTS, ("Invalid IRQ %u", irq));
+ KASSERT(irq < num_io_irqs, ("Invalid IRQ %u", irq));
KASSERT(lapics[apic_id].la_ioint_irqs[vector - APIC_IO_INTS] ==
irq, ("IRQ mismatch"));
#ifdef KDTRACE_HOOKS
@@ -1643,7 +1665,7 @@
thread_unlock(td);
}
mtx_lock_spin(&icu_lock);
- lapics[apic_id].la_ioint_irqs[vector - APIC_IO_INTS] = -1;
+ lapics[apic_id].la_ioint_irqs[vector - APIC_IO_INTS] = IRQ_FREE;
mtx_unlock_spin(&icu_lock);
if (!rebooting) {
thread_lock(td);
@@ -1694,7 +1716,7 @@
db_printf("Interrupts bound to lapic %u\n", apic_id);
for (i = 0; i < APIC_NUM_IOINTS + 1 && !db_pager_quit; i++) {
irq = lapics[apic_id].la_ioint_irqs[i];
- if (irq == -1 || irq == IRQ_SYSCALL)
+ if (irq == IRQ_FREE || irq == IRQ_SYSCALL)
continue;
#ifdef KDTRACE_HOOKS
if (irq == IRQ_DTRACE_RET)
@@ -1707,7 +1729,7 @@
db_printf("vec 0x%2x -> ", i + APIC_IO_INTS);
if (irq == IRQ_TIMER)
db_printf("lapic timer\n");
- else if (irq < NUM_IO_INTS) {
+ else if (irq < num_io_irqs) {
isrc = intr_lookup_source(irq);
if (isrc == NULL || verbose == 0)
db_printf("IRQ %u\n", irq);
@@ -1934,6 +1956,10 @@
/* Enable the MSI "pic". */
init_ops.msi_init();
+
+#ifdef XENHVM
+ xen_intr_alloc_irqs();
+#endif
}
SYSINIT(apic_setup_io, SI_SUB_INTR, SI_ORDER_THIRD, apic_setup_io, NULL);
Index: head/sys/x86/x86/msi.c
===================================================================
--- head/sys/x86/x86/msi.c
+++ head/sys/x86/x86/msi.c
@@ -120,7 +120,7 @@
u_int msi_cpu; /* Local APIC ID. (g) */
u_int msi_count:8; /* Messages in this group. (g) */
u_int msi_maxcount:8; /* Alignment for this group. (g) */
- int *msi_irqs; /* Group's IRQ list. (g) */
+ u_int *msi_irqs; /* Group's IRQ list. (g) */
u_int msi_remap_cookie;
};
@@ -151,6 +151,8 @@
.pic_reprogram_pin = NULL,
};
+u_int first_msi_irq;
+
#ifdef SMP
/**
* Xen hypervisors prior to 4.6.0 do not properly handle updates to
@@ -168,7 +170,7 @@
#endif
static int msi_enabled;
-static int msi_last_irq;
+static u_int msi_last_irq;
static struct mtx msi_lock;
static void
@@ -329,6 +331,10 @@
}
#endif
+ MPASS(num_io_irqs > 0);
+ first_msi_irq = max(MINIMUM_MSI_INT, num_io_irqs);
+ num_io_irqs = first_msi_irq + NUM_MSI_INTS;
+
msi_enabled = 1;
intr_register_pic(&msi_pic);
mtx_init(&msi_lock, "msi", NULL, MTX_DEF);
@@ -345,7 +351,7 @@
mtx_unlock(&msi_lock);
return;
}
- irq = msi_last_irq + FIRST_MSI_INT;
+ irq = msi_last_irq + first_msi_irq;
msi_last_irq++;
mtx_unlock(&msi_lock);
@@ -363,8 +369,8 @@
msi_alloc(device_t dev, int count, int maxcount, int *irqs)
{
struct msi_intsrc *msi, *fsrc;
- u_int cpu, domain;
- int cnt, i, *mirqs, vector;
+ u_int cpu, domain, *mirqs;
+ int cnt, i, vector;
#ifdef ACPI_DMAR
u_int cookies[count];
int error;
@@ -385,7 +391,7 @@
/* Try to find 'count' free IRQs. */
cnt = 0;
- for (i = FIRST_MSI_INT; i < FIRST_MSI_INT + NUM_MSI_INTS; i++) {
+ for (i = first_msi_irq; i < first_msi_irq + NUM_MSI_INTS; i++) {
msi = (struct msi_intsrc *)intr_lookup_source(i);
/* End of allocated sources, so break. */
@@ -404,7 +410,7 @@
/* Do we need to create some new sources? */
if (cnt < count) {
/* If we would exceed the max, give up. */
- if (i + (count - cnt) >= FIRST_MSI_INT + NUM_MSI_INTS) {
+ if (i + (count - cnt) >= first_msi_irq + NUM_MSI_INTS) {
mtx_unlock(&msi_lock);
free(mirqs, M_MSI);
return (ENXIO);
@@ -579,8 +585,8 @@
#ifdef ACPI_DMAR
if (!msi->msi_msix) {
- for (k = msi->msi_count - 1, i = FIRST_MSI_INT; k > 0 &&
- i < FIRST_MSI_INT + NUM_MSI_INTS; i++) {
+ for (k = msi->msi_count - 1, i = first_msi_irq; k > 0 &&
+ i < first_msi_irq + NUM_MSI_INTS; i++) {
if (i == msi->msi_irq)
continue;
msi1 = (struct msi_intsrc *)intr_lookup_source(i);
@@ -630,7 +636,7 @@
mtx_lock(&msi_lock);
/* Find a free IRQ. */
- for (i = FIRST_MSI_INT; i < FIRST_MSI_INT + NUM_MSI_INTS; i++) {
+ for (i = first_msi_irq; i < first_msi_irq + NUM_MSI_INTS; i++) {
msi = (struct msi_intsrc *)intr_lookup_source(i);
/* End of allocated sources, so break. */
@@ -645,7 +651,7 @@
/* Do we need to create a new source? */
if (msi == NULL) {
/* If we would exceed the max, give up. */
- if (i + 1 >= FIRST_MSI_INT + NUM_MSI_INTS) {
+ if (i + 1 >= first_msi_irq + NUM_MSI_INTS) {
mtx_unlock(&msi_lock);
return (ENXIO);
}
Index: head/sys/x86/x86/nexus.c
===================================================================
--- head/sys/x86/x86/nexus.c
+++ head/sys/x86/x86/nexus.c
@@ -223,7 +223,7 @@
irq_rman.rm_start = 0;
irq_rman.rm_type = RMAN_ARRAY;
irq_rman.rm_descr = "Interrupt request lines";
- irq_rman.rm_end = NUM_IO_INTS - 1;
+ irq_rman.rm_end = num_io_irqs - 1;
if (rman_init(&irq_rman))
panic("nexus_init_resources irq_rman");
@@ -231,7 +231,7 @@
* We search for regions of existing IRQs and add those to the IRQ
* resource manager.
*/
- for (irq = 0; irq < NUM_IO_INTS; irq++)
+ for (irq = 0; irq < num_io_irqs; irq++)
if (intr_lookup_source(irq) != NULL)
if (rman_manage_region(&irq_rman, irq, irq) != 0)
panic("nexus_init_resources irq_rman add");
Index: head/sys/x86/xen/xen_intr.c
===================================================================
--- head/sys/x86/xen/xen_intr.c
+++ head/sys/x86/xen/xen_intr.c
@@ -59,6 +59,7 @@
#include <machine/xen/synch_bitops.h>
#include <machine/xen/xen-os.h>
+#include <xen/xen-os.h>
#include <xen/hypervisor.h>
#include <xen/xen_intr.h>
#include <xen/evtchn/evtchnvar.h>
@@ -72,6 +73,8 @@
static MALLOC_DEFINE(M_XENINTR, "xen_intr", "Xen Interrupt Services");
+static u_int first_evtchn_irq;
+
/**
* Per-cpu event channel processing state.
*/
@@ -187,7 +190,7 @@
};
static struct mtx xen_intr_isrc_lock;
-static int xen_intr_auto_vector_count;
+static u_int xen_intr_auto_vector_count;
static struct xenisrc *xen_intr_port_to_isrc[NR_EVENT_CHANNELS];
static u_long *xen_intr_pirq_eoi_map;
static boolean_t xen_intr_pirq_eoi_map_enabled;
@@ -276,7 +279,7 @@
struct xenisrc *isrc;
u_int vector;
- vector = FIRST_EVTCHN_INT + isrc_idx;
+ vector = first_evtchn_irq + isrc_idx;
isrc = (struct xenisrc *)intr_lookup_source(vector);
if (isrc != NULL
&& isrc->xi_type == EVTCHN_TYPE_UNBOUND) {
@@ -314,7 +317,7 @@
}
if (type != EVTCHN_TYPE_PIRQ) {
- vector = FIRST_EVTCHN_INT + xen_intr_auto_vector_count;
+ vector = first_evtchn_irq + xen_intr_auto_vector_count;
xen_intr_auto_vector_count++;
}
@@ -473,8 +476,8 @@
return (NULL);
vector = *(int *)handle;
- KASSERT(vector >= FIRST_EVTCHN_INT &&
- vector < (FIRST_EVTCHN_INT + xen_intr_auto_vector_count),
+ KASSERT(vector >= first_evtchn_irq &&
+ vector < (first_evtchn_irq + xen_intr_auto_vector_count),
("Xen interrupt vector is out of range"));
return ((struct xenisrc *)intr_lookup_source(vector));
@@ -631,17 +634,13 @@
mtx_init(&xen_intr_isrc_lock, "xen-irq-lock", NULL, MTX_DEF);
/*
- * Register interrupt count manually as we aren't
- * guaranteed to see a call to xen_intr_assign_cpu()
- * before our first interrupt. Also set the per-cpu
- * mask of CPU#0 to enable all, since by default
- * all event channels are bound to CPU#0.
+ * Set the per-cpu mask of CPU#0 to enable all, since by default all
+ * event channels are bound to CPU#0.
*/
CPU_FOREACH(i) {
pcpu = DPCPU_ID_PTR(i, xen_intr_pcpu);
memset(pcpu->evtchn_enabled, i == 0 ? ~0 : 0,
sizeof(pcpu->evtchn_enabled));
- xen_intr_intrcnt_add(i);
}
for (i = 0; i < nitems(s->evtchn_mask); i++)
@@ -666,6 +665,31 @@
}
SYSINIT(xen_intr_init, SI_SUB_INTR, SI_ORDER_SECOND, xen_intr_init, NULL);
+static void
+xen_intrcnt_init(void *dummy __unused)
+{
+ unsigned int i;
+
+ if (!xen_domain())
+ return;
+
+ /*
+ * Register interrupt count manually as we aren't guaranteed to see a
+ * call to xen_intr_assign_cpu() before our first interrupt.
+ */
+ CPU_FOREACH(i)
+ xen_intr_intrcnt_add(i);
+}
+SYSINIT(xen_intrcnt_init, SI_SUB_INTR, SI_ORDER_MIDDLE, xen_intrcnt_init, NULL);
+
+void
+xen_intr_alloc_irqs(void)
+{
+
+ first_evtchn_irq = num_io_irqs;
+ num_io_irqs += NR_EVENT_CHANNELS;
+}
+
/*--------------------------- Common PIC Functions ---------------------------*/
/**
* Prepare this PIC for system suspension.
@@ -768,7 +792,7 @@
for (isrc_idx = 0; isrc_idx < xen_intr_auto_vector_count; isrc_idx++) {
u_int vector;
- vector = FIRST_EVTCHN_INT + isrc_idx;
+ vector = first_evtchn_irq + isrc_idx;
isrc = (struct xenisrc *)intr_lookup_source(vector);
if (isrc != NULL) {
isrc->xi_port = 0;
@@ -872,7 +896,6 @@
to_cpu = apic_cpuid(apic_id);
vcpu_id = pcpu_find(to_cpu)->pc_vcpu_id;
- xen_intr_intrcnt_add(to_cpu);
mtx_lock(&xen_intr_isrc_lock);
isrc = (struct xenisrc *)base_isrc;
@@ -1273,9 +1296,6 @@
struct evtchn_bind_virq bind_virq = { .virq = virq, .vcpu = vcpu_id };
int error;
- /* Ensure the target CPU is ready to handle evtchn interrupts. */
- xen_intr_intrcnt_add(cpu);
-
isrc = NULL;
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_virq, &bind_virq);
if (error != 0) {
@@ -1337,9 +1357,6 @@
/* Same size as the one used by intr_handler->ih_name. */
char name[MAXCOMLEN + 1];
int error;
-
- /* Ensure the target CPU is ready to handle evtchn interrupts. */
- xen_intr_intrcnt_add(cpu);
isrc = NULL;
error = HYPERVISOR_event_channel_op(EVTCHNOP_bind_ipi, &bind_ipi);
Index: head/sys/x86/xen/xen_msi.c
===================================================================
--- head/sys/x86/xen/xen_msi.c
+++ head/sys/x86/xen/xen_msi.c
@@ -44,16 +44,21 @@
#include <machine/specialreg.h>
#include <dev/pci/pcivar.h>
+#include <xen/xen-os.h>
#include <xen/xen_intr.h>
#include <xen/xen_msi.h>
static struct mtx msi_lock;
-static int msi_last_irq;
+static u_int msi_last_irq;
void
xen_msi_init(void)
{
+ MPASS(num_io_irqs > 0);
+ first_msi_irq = min(MINIMUM_MSI_INT, num_io_irqs);
+ num_io_irqs = first_msi_irq + NUM_MSI_INTS;
+
mtx_init(&msi_lock, "msi", NULL, MTX_DEF);
}
@@ -75,7 +80,7 @@
/* Allocate MSI vectors */
for (i = 0; i < count; i++)
- irqs[i] = FIRST_MSI_INT + msi_last_irq++;
+ irqs[i] = first_msi_irq + msi_last_irq++;
mtx_unlock(&msi_lock);
Index: head/sys/x86/xen/xen_nexus.c
===================================================================
--- head/sys/x86/xen/xen_nexus.c
+++ head/sys/x86/xen/xen_nexus.c
@@ -99,7 +99,7 @@
* ISA and PCI intline IRQs are not preregistered on Xen, so
* intercept calls to configure those and register them on the fly.
*/
- if ((irq < FIRST_MSI_INT) && (intr_lookup_source(irq) == NULL)) {
+ if ((irq < first_msi_irq) && (intr_lookup_source(irq) == NULL)) {
ret = xen_register_pirq(irq, trig, pol);
if (ret != 0)
return (ret);
Index: head/usr.bin/vmstat/vmstat.c
===================================================================
--- head/usr.bin/vmstat/vmstat.c
+++ head/usr.bin/vmstat/vmstat.c
@@ -86,7 +86,7 @@
static char da[] = "da";
enum x_stats { X_SUM, X_HZ, X_STATHZ, X_NCHSTATS, X_INTRNAMES, X_SINTRNAMES,
- X_INTRCNT, X_SINTRCNT };
+ X_INTRCNT, X_SINTRCNT, X_NINTRCNT };
static struct nlist namelist[] = {
[X_SUM] = { .n_name = "_vm_cnt", },
@@ -97,6 +97,7 @@
[X_SINTRNAMES] = { .n_name = "_sintrnames", },
[X_INTRCNT] = { .n_name = "_intrcnt", },
[X_SINTRCNT] = { .n_name = "_sintrcnt", },
+ [X_NINTRCNT] = { .n_name = "_nintrcnt", },
{ .n_name = NULL, },
};
@@ -196,6 +197,7 @@
static void domemstat_zone(void);
static void kread(int, void *, size_t);
static void kreado(int, void *, size_t, size_t);
+static void kreadptr(uintptr_t, void *, size_t);
static void needhdr(int);
static void needresize(int);
static void doresize(void);
@@ -318,6 +320,13 @@
goto retry_nlist;
}
+ /*
+ * 'nintrcnt' doesn't exist in older kernels, but
+ * that isn't fatal.
+ */
+ if (namelist[X_NINTRCNT].n_type == 0 && c == 1)
+ goto nlist_ok;
+
for (c = 0; c < (int)(nitems(namelist)); c++)
if (namelist[c].n_type == 0)
bufsize += strlen(namelist[c].n_name)
@@ -341,6 +350,7 @@
xo_finish();
exit(1);
}
+nlist_ok:
if (kd && Pflag)
xo_errx(1, "Cannot use -P with crash dumps");
@@ -1232,12 +1242,18 @@
read_intrcnts(unsigned long **intrcnts)
{
size_t intrcntlen;
+ uintptr_t kaddr;
if (kd != NULL) {
kread(X_SINTRCNT, &intrcntlen, sizeof(intrcntlen));
if ((*intrcnts = malloc(intrcntlen)) == NULL)
err(1, "malloc()");
- kread(X_INTRCNT, *intrcnts, intrcntlen);
+ if (namelist[X_NINTRCNT].n_type == 0)
+ kread(X_INTRCNT, *intrcnts, intrcntlen);
+ else {
+ kread(X_INTRCNT, &kaddr, sizeof(kaddr));
+ kreadptr(kaddr, *intrcnts, intrcntlen);
+ }
} else {
for (*intrcnts = NULL, intrcntlen = 1024; ; intrcntlen *= 2) {
*intrcnts = reallocf(*intrcnts, intrcntlen);
@@ -1294,6 +1310,7 @@
char *intrname, *intrnames;
long long period_ms, old_uptime, uptime;
size_t clen, inamlen, istrnamlen;
+ uintptr_t kaddr;
unsigned int nintr;
old_intrcnts = NULL;
@@ -1304,7 +1321,12 @@
kread(X_SINTRNAMES, &inamlen, sizeof(inamlen));
if ((intrnames = malloc(inamlen)) == NULL)
xo_err(1, "malloc()");
- kread(X_INTRNAMES, intrnames, inamlen);
+ if (namelist[X_NINTRCNT].n_type == 0)
+ kread(X_INTRNAMES, intrnames, inamlen);
+ else {
+ kread(X_INTRNAMES, &kaddr, sizeof(kaddr));
+ kreadptr(kaddr, intrnames, inamlen);
+ }
} else {
for (intrnames = NULL, inamlen = 1024; ; inamlen *= 2) {
if ((intrnames = reallocf(intrnames, inamlen)) == NULL)
@@ -1644,6 +1666,14 @@
{
kreado(nlx, addr, size, 0);
+}
+
+static void
+kreadptr(uintptr_t addr, void *buf, size_t size)
+{
+
+ if ((size_t)kvm_read(kd, addr, buf, size) != size)
+ xo_errx(1, "%s", kvm_geterr(kd));
}
static void __dead2
File Metadata
Details
Attached
Mime Type
text/plain
Expires
Tue, Feb 4, 11:50 PM (16 h, 58 m)
Storage Engine
blob
Storage Format
Raw Data
Storage Handle
16463178
Default Alt Text
D16861.id.diff (39 KB)
Attached To
Mode
D16861: Dynamically allocate IRQ ranges on x86.
Attached
Detach File
Event Timeline
Log In to Comment