mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-15 08:44:14 +08:00
genirq: Fix typos and misspellings in comments
No functional change. Signed-off-by: Krzysztof Kozlowski <krzysztof.kozlowski@canonical.com> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Link: https://lore.kernel.org/r/20210316100205.23492-1-krzysztof.kozlowski@canonical.com
This commit is contained in:
parent
3a0ade0c52
commit
5c982c5875
@ -761,7 +761,7 @@ EXPORT_SYMBOL_GPL(handle_fasteoi_nmi);
|
||||
* handle_edge_irq - edge type IRQ handler
|
||||
* @desc: the interrupt description structure for this irq
|
||||
*
|
||||
* Interrupt occures on the falling and/or rising edge of a hardware
|
||||
* Interrupt occurs on the falling and/or rising edge of a hardware
|
||||
* signal. The occurrence is latched into the irq controller hardware
|
||||
* and must be acked in order to be reenabled. After the ack another
|
||||
* interrupt can happen on the same source even before the first one
|
||||
@ -1419,7 +1419,7 @@ EXPORT_SYMBOL_GPL(irq_chip_eoi_parent);
|
||||
* @dest: The affinity mask to set
|
||||
* @force: Flag to enforce setting (disable online checks)
|
||||
*
|
||||
* Conditinal, as the underlying parent chip might not implement it.
|
||||
* Conditional, as the underlying parent chip might not implement it.
|
||||
*/
|
||||
int irq_chip_set_affinity_parent(struct irq_data *data,
|
||||
const struct cpumask *dest, bool force)
|
||||
@ -1531,7 +1531,7 @@ EXPORT_SYMBOL_GPL(irq_chip_release_resources_parent);
|
||||
#endif
|
||||
|
||||
/**
|
||||
* irq_chip_compose_msi_msg - Componse msi message for a irq chip
|
||||
* irq_chip_compose_msi_msg - Compose msi message for a irq chip
|
||||
* @data: Pointer to interrupt specific data
|
||||
* @msg: Pointer to the MSI message
|
||||
*
|
||||
|
@ -107,7 +107,7 @@ free_descs:
|
||||
* @irq: linux irq number to be destroyed
|
||||
* @dest: cpumask of cpus which should have the IPI removed
|
||||
*
|
||||
* The IPIs allocated with irq_reserve_ipi() are retuerned to the system
|
||||
* The IPIs allocated with irq_reserve_ipi() are returned to the system
|
||||
* destroying all virqs associated with them.
|
||||
*
|
||||
* Return 0 on success or error code on failure.
|
||||
|
@ -179,7 +179,7 @@ bool irq_can_set_affinity_usr(unsigned int irq)
|
||||
|
||||
/**
|
||||
* irq_set_thread_affinity - Notify irq threads to adjust affinity
|
||||
* @desc: irq descriptor which has affitnity changed
|
||||
* @desc: irq descriptor which has affinity changed
|
||||
*
|
||||
* We just set IRQTF_AFFINITY and delegate the affinity setting
|
||||
* to the interrupt thread itself. We can not call
|
||||
@ -1153,7 +1153,7 @@ irq_forced_thread_fn(struct irq_desc *desc, struct irqaction *action)
|
||||
|
||||
/*
|
||||
* Interrupts explicitly requested as threaded interrupts want to be
|
||||
* preemtible - many of them need to sleep and wait for slow busses to
|
||||
* preemptible - many of them need to sleep and wait for slow busses to
|
||||
* complete.
|
||||
*/
|
||||
static irqreturn_t irq_thread_fn(struct irq_desc *desc,
|
||||
@ -2749,7 +2749,7 @@ int __irq_get_irqchip_state(struct irq_data *data, enum irqchip_irq_state which,
|
||||
* irq_get_irqchip_state - returns the irqchip state of a interrupt.
|
||||
* @irq: Interrupt line that is forwarded to a VM
|
||||
* @which: One of IRQCHIP_STATE_* the caller wants to know about
|
||||
* @state: a pointer to a boolean where the state is to be storeed
|
||||
* @state: a pointer to a boolean where the state is to be stored
|
||||
*
|
||||
* This call snapshots the internal irqchip state of an
|
||||
* interrupt, returning into @state the bit corresponding to
|
||||
|
@ -356,7 +356,7 @@ void irq_matrix_reserve(struct irq_matrix *m)
|
||||
* irq_matrix_remove_reserved - Remove interrupt reservation
|
||||
* @m: Matrix pointer
|
||||
*
|
||||
* This is merily a book keeping call. It decrements the number of globally
|
||||
* This is merely a book keeping call. It decrements the number of globally
|
||||
* reserved interrupt bits. This is used to undo irq_matrix_reserve() when the
|
||||
* interrupt was never in use and a real vector allocated, which undid the
|
||||
* reservation.
|
||||
|
@ -7,7 +7,7 @@
|
||||
|
||||
/**
|
||||
* irq_fixup_move_pending - Cleanup irq move pending from a dying CPU
|
||||
* @desc: Interrupt descpriptor to clean up
|
||||
* @desc: Interrupt descriptor to clean up
|
||||
* @force_clear: If set clear the move pending bit unconditionally.
|
||||
* If not set, clear it only when the dying CPU is the
|
||||
* last one in the pending mask.
|
||||
|
@ -128,7 +128,7 @@ int check_irq_resend(struct irq_desc *desc, bool inject)
|
||||
if (!try_retrigger(desc))
|
||||
err = irq_sw_resend(desc);
|
||||
|
||||
/* If the retrigger was successfull, mark it with the REPLAY bit */
|
||||
/* If the retrigger was successful, mark it with the REPLAY bit */
|
||||
if (!err)
|
||||
desc->istate |= IRQS_REPLAY;
|
||||
return err;
|
||||
|
@ -84,7 +84,7 @@ void irq_timings_disable(void)
|
||||
* 2. Log interval
|
||||
*
|
||||
* We saw the irq timings allow to compute the interval of the
|
||||
* occurrences for a specific interrupt. We can reasonibly assume the
|
||||
* occurrences for a specific interrupt. We can reasonably assume the
|
||||
* longer is the interval, the higher is the error for the next event
|
||||
* and we can consider storing those interval values into an array
|
||||
* where each slot in the array correspond to an interval at the power
|
||||
@ -416,7 +416,7 @@ static u64 __irq_timings_next_event(struct irqt_stat *irqs, int irq, u64 now)
|
||||
* Copy the content of the circular buffer into another buffer
|
||||
* in order to linearize the buffer instead of dealing with
|
||||
* wrapping indexes and shifted array which will be prone to
|
||||
* error and extremelly difficult to debug.
|
||||
* error and extremely difficult to debug.
|
||||
*/
|
||||
for (i = 0; i < count; i++) {
|
||||
int index = (start + i) & IRQ_TIMINGS_MASK;
|
||||
@ -514,7 +514,7 @@ static inline void irq_timings_store(int irq, struct irqt_stat *irqs, u64 ts)
|
||||
* If more than the array size interrupts happened during the
|
||||
* last busy/idle cycle, the index wrapped up and we have to
|
||||
* begin with the next element in the array which is the last one
|
||||
* in the sequence, otherwise it is a the index 0.
|
||||
* in the sequence, otherwise it is at the index 0.
|
||||
*
|
||||
* - have an indication of the interrupts activity on this CPU
|
||||
* (eg. irq/sec)
|
||||
|
Loading…
Reference in New Issue
Block a user