2
0
mirror of https://github.com/edk2-porting/linux-next.git synced 2024-12-15 16:53:54 +08:00

genirq: Fix typos and misspellings in comments

No functional change.

Signed-off-by: Krzysztof Kozlowski <krzysztof.kozlowski@canonical.com>
Signed-off-by: Thomas Gleixner <tglx@linutronix.de>
Link: https://lore.kernel.org/r/20210316100205.23492-1-krzysztof.kozlowski@canonical.com
This commit is contained in:
Krzysztof Kozlowski 2021-03-16 11:02:05 +01:00 committed by Thomas Gleixner
parent 3a0ade0c52
commit 5c982c5875
7 changed files with 13 additions and 13 deletions

View File

@ -761,7 +761,7 @@ EXPORT_SYMBOL_GPL(handle_fasteoi_nmi);
* handle_edge_irq - edge type IRQ handler * handle_edge_irq - edge type IRQ handler
* @desc: the interrupt description structure for this irq * @desc: the interrupt description structure for this irq
* *
* Interrupt occures on the falling and/or rising edge of a hardware * Interrupt occurs on the falling and/or rising edge of a hardware
* signal. The occurrence is latched into the irq controller hardware * signal. The occurrence is latched into the irq controller hardware
* and must be acked in order to be reenabled. After the ack another * and must be acked in order to be reenabled. After the ack another
* interrupt can happen on the same source even before the first one * interrupt can happen on the same source even before the first one
@ -1419,7 +1419,7 @@ EXPORT_SYMBOL_GPL(irq_chip_eoi_parent);
* @dest: The affinity mask to set * @dest: The affinity mask to set
* @force: Flag to enforce setting (disable online checks) * @force: Flag to enforce setting (disable online checks)
* *
* Conditinal, as the underlying parent chip might not implement it. * Conditional, as the underlying parent chip might not implement it.
*/ */
int irq_chip_set_affinity_parent(struct irq_data *data, int irq_chip_set_affinity_parent(struct irq_data *data,
const struct cpumask *dest, bool force) const struct cpumask *dest, bool force)
@ -1531,7 +1531,7 @@ EXPORT_SYMBOL_GPL(irq_chip_release_resources_parent);
#endif #endif
/** /**
* irq_chip_compose_msi_msg - Componse msi message for a irq chip * irq_chip_compose_msi_msg - Compose msi message for a irq chip
* @data: Pointer to interrupt specific data * @data: Pointer to interrupt specific data
* @msg: Pointer to the MSI message * @msg: Pointer to the MSI message
* *

View File

@ -107,7 +107,7 @@ free_descs:
* @irq: linux irq number to be destroyed * @irq: linux irq number to be destroyed
* @dest: cpumask of cpus which should have the IPI removed * @dest: cpumask of cpus which should have the IPI removed
* *
* The IPIs allocated with irq_reserve_ipi() are retuerned to the system * The IPIs allocated with irq_reserve_ipi() are returned to the system
* destroying all virqs associated with them. * destroying all virqs associated with them.
* *
* Return 0 on success or error code on failure. * Return 0 on success or error code on failure.

View File

@ -179,7 +179,7 @@ bool irq_can_set_affinity_usr(unsigned int irq)
/** /**
* irq_set_thread_affinity - Notify irq threads to adjust affinity * irq_set_thread_affinity - Notify irq threads to adjust affinity
* @desc: irq descriptor which has affitnity changed * @desc: irq descriptor which has affinity changed
* *
* We just set IRQTF_AFFINITY and delegate the affinity setting * We just set IRQTF_AFFINITY and delegate the affinity setting
* to the interrupt thread itself. We can not call * to the interrupt thread itself. We can not call
@ -1153,7 +1153,7 @@ irq_forced_thread_fn(struct irq_desc *desc, struct irqaction *action)
/* /*
* Interrupts explicitly requested as threaded interrupts want to be * Interrupts explicitly requested as threaded interrupts want to be
* preemtible - many of them need to sleep and wait for slow busses to * preemptible - many of them need to sleep and wait for slow busses to
* complete. * complete.
*/ */
static irqreturn_t irq_thread_fn(struct irq_desc *desc, static irqreturn_t irq_thread_fn(struct irq_desc *desc,
@ -2749,7 +2749,7 @@ int __irq_get_irqchip_state(struct irq_data *data, enum irqchip_irq_state which,
* irq_get_irqchip_state - returns the irqchip state of a interrupt. * irq_get_irqchip_state - returns the irqchip state of a interrupt.
* @irq: Interrupt line that is forwarded to a VM * @irq: Interrupt line that is forwarded to a VM
* @which: One of IRQCHIP_STATE_* the caller wants to know about * @which: One of IRQCHIP_STATE_* the caller wants to know about
* @state: a pointer to a boolean where the state is to be storeed * @state: a pointer to a boolean where the state is to be stored
* *
* This call snapshots the internal irqchip state of an * This call snapshots the internal irqchip state of an
* interrupt, returning into @state the bit corresponding to * interrupt, returning into @state the bit corresponding to

View File

@ -356,7 +356,7 @@ void irq_matrix_reserve(struct irq_matrix *m)
* irq_matrix_remove_reserved - Remove interrupt reservation * irq_matrix_remove_reserved - Remove interrupt reservation
* @m: Matrix pointer * @m: Matrix pointer
* *
* This is merily a book keeping call. It decrements the number of globally * This is merely a book keeping call. It decrements the number of globally
* reserved interrupt bits. This is used to undo irq_matrix_reserve() when the * reserved interrupt bits. This is used to undo irq_matrix_reserve() when the
* interrupt was never in use and a real vector allocated, which undid the * interrupt was never in use and a real vector allocated, which undid the
* reservation. * reservation.

View File

@ -7,7 +7,7 @@
/** /**
* irq_fixup_move_pending - Cleanup irq move pending from a dying CPU * irq_fixup_move_pending - Cleanup irq move pending from a dying CPU
* @desc: Interrupt descpriptor to clean up * @desc: Interrupt descriptor to clean up
* @force_clear: If set clear the move pending bit unconditionally. * @force_clear: If set clear the move pending bit unconditionally.
* If not set, clear it only when the dying CPU is the * If not set, clear it only when the dying CPU is the
* last one in the pending mask. * last one in the pending mask.

View File

@ -128,7 +128,7 @@ int check_irq_resend(struct irq_desc *desc, bool inject)
if (!try_retrigger(desc)) if (!try_retrigger(desc))
err = irq_sw_resend(desc); err = irq_sw_resend(desc);
/* If the retrigger was successfull, mark it with the REPLAY bit */ /* If the retrigger was successful, mark it with the REPLAY bit */
if (!err) if (!err)
desc->istate |= IRQS_REPLAY; desc->istate |= IRQS_REPLAY;
return err; return err;

View File

@ -84,7 +84,7 @@ void irq_timings_disable(void)
* 2. Log interval * 2. Log interval
* *
* We saw the irq timings allow to compute the interval of the * We saw the irq timings allow to compute the interval of the
* occurrences for a specific interrupt. We can reasonibly assume the * occurrences for a specific interrupt. We can reasonably assume the
* longer is the interval, the higher is the error for the next event * longer is the interval, the higher is the error for the next event
* and we can consider storing those interval values into an array * and we can consider storing those interval values into an array
* where each slot in the array correspond to an interval at the power * where each slot in the array correspond to an interval at the power
@ -416,7 +416,7 @@ static u64 __irq_timings_next_event(struct irqt_stat *irqs, int irq, u64 now)
* Copy the content of the circular buffer into another buffer * Copy the content of the circular buffer into another buffer
* in order to linearize the buffer instead of dealing with * in order to linearize the buffer instead of dealing with
* wrapping indexes and shifted array which will be prone to * wrapping indexes and shifted array which will be prone to
* error and extremelly difficult to debug. * error and extremely difficult to debug.
*/ */
for (i = 0; i < count; i++) { for (i = 0; i < count; i++) {
int index = (start + i) & IRQ_TIMINGS_MASK; int index = (start + i) & IRQ_TIMINGS_MASK;
@ -514,7 +514,7 @@ static inline void irq_timings_store(int irq, struct irqt_stat *irqs, u64 ts)
* If more than the array size interrupts happened during the * If more than the array size interrupts happened during the
* last busy/idle cycle, the index wrapped up and we have to * last busy/idle cycle, the index wrapped up and we have to
* begin with the next element in the array which is the last one * begin with the next element in the array which is the last one
* in the sequence, otherwise it is a the index 0. * in the sequence, otherwise it is at the index 0.
* *
* - have an indication of the interrupts activity on this CPU * - have an indication of the interrupts activity on this CPU
* (eg. irq/sec) * (eg. irq/sec)