mirror of
https://github.com/edk2-porting/linux-next.git
synced 2024-12-15 08:44:14 +08:00
genirq: Fix typos and misspellings in comments
No functional change. Signed-off-by: Krzysztof Kozlowski <krzysztof.kozlowski@canonical.com> Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Link: https://lore.kernel.org/r/20210316100205.23492-1-krzysztof.kozlowski@canonical.com
This commit is contained in:
parent
3a0ade0c52
commit
5c982c5875
@ -761,7 +761,7 @@ EXPORT_SYMBOL_GPL(handle_fasteoi_nmi);
|
|||||||
* handle_edge_irq - edge type IRQ handler
|
* handle_edge_irq - edge type IRQ handler
|
||||||
* @desc: the interrupt description structure for this irq
|
* @desc: the interrupt description structure for this irq
|
||||||
*
|
*
|
||||||
* Interrupt occures on the falling and/or rising edge of a hardware
|
* Interrupt occurs on the falling and/or rising edge of a hardware
|
||||||
* signal. The occurrence is latched into the irq controller hardware
|
* signal. The occurrence is latched into the irq controller hardware
|
||||||
* and must be acked in order to be reenabled. After the ack another
|
* and must be acked in order to be reenabled. After the ack another
|
||||||
* interrupt can happen on the same source even before the first one
|
* interrupt can happen on the same source even before the first one
|
||||||
@ -1419,7 +1419,7 @@ EXPORT_SYMBOL_GPL(irq_chip_eoi_parent);
|
|||||||
* @dest: The affinity mask to set
|
* @dest: The affinity mask to set
|
||||||
* @force: Flag to enforce setting (disable online checks)
|
* @force: Flag to enforce setting (disable online checks)
|
||||||
*
|
*
|
||||||
* Conditinal, as the underlying parent chip might not implement it.
|
* Conditional, as the underlying parent chip might not implement it.
|
||||||
*/
|
*/
|
||||||
int irq_chip_set_affinity_parent(struct irq_data *data,
|
int irq_chip_set_affinity_parent(struct irq_data *data,
|
||||||
const struct cpumask *dest, bool force)
|
const struct cpumask *dest, bool force)
|
||||||
@ -1531,7 +1531,7 @@ EXPORT_SYMBOL_GPL(irq_chip_release_resources_parent);
|
|||||||
#endif
|
#endif
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* irq_chip_compose_msi_msg - Componse msi message for a irq chip
|
* irq_chip_compose_msi_msg - Compose msi message for a irq chip
|
||||||
* @data: Pointer to interrupt specific data
|
* @data: Pointer to interrupt specific data
|
||||||
* @msg: Pointer to the MSI message
|
* @msg: Pointer to the MSI message
|
||||||
*
|
*
|
||||||
|
@ -107,7 +107,7 @@ free_descs:
|
|||||||
* @irq: linux irq number to be destroyed
|
* @irq: linux irq number to be destroyed
|
||||||
* @dest: cpumask of cpus which should have the IPI removed
|
* @dest: cpumask of cpus which should have the IPI removed
|
||||||
*
|
*
|
||||||
* The IPIs allocated with irq_reserve_ipi() are retuerned to the system
|
* The IPIs allocated with irq_reserve_ipi() are returned to the system
|
||||||
* destroying all virqs associated with them.
|
* destroying all virqs associated with them.
|
||||||
*
|
*
|
||||||
* Return 0 on success or error code on failure.
|
* Return 0 on success or error code on failure.
|
||||||
|
@ -179,7 +179,7 @@ bool irq_can_set_affinity_usr(unsigned int irq)
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* irq_set_thread_affinity - Notify irq threads to adjust affinity
|
* irq_set_thread_affinity - Notify irq threads to adjust affinity
|
||||||
* @desc: irq descriptor which has affitnity changed
|
* @desc: irq descriptor which has affinity changed
|
||||||
*
|
*
|
||||||
* We just set IRQTF_AFFINITY and delegate the affinity setting
|
* We just set IRQTF_AFFINITY and delegate the affinity setting
|
||||||
* to the interrupt thread itself. We can not call
|
* to the interrupt thread itself. We can not call
|
||||||
@ -1153,7 +1153,7 @@ irq_forced_thread_fn(struct irq_desc *desc, struct irqaction *action)
|
|||||||
|
|
||||||
/*
|
/*
|
||||||
* Interrupts explicitly requested as threaded interrupts want to be
|
* Interrupts explicitly requested as threaded interrupts want to be
|
||||||
* preemtible - many of them need to sleep and wait for slow busses to
|
* preemptible - many of them need to sleep and wait for slow busses to
|
||||||
* complete.
|
* complete.
|
||||||
*/
|
*/
|
||||||
static irqreturn_t irq_thread_fn(struct irq_desc *desc,
|
static irqreturn_t irq_thread_fn(struct irq_desc *desc,
|
||||||
@ -2749,7 +2749,7 @@ int __irq_get_irqchip_state(struct irq_data *data, enum irqchip_irq_state which,
|
|||||||
* irq_get_irqchip_state - returns the irqchip state of a interrupt.
|
* irq_get_irqchip_state - returns the irqchip state of a interrupt.
|
||||||
* @irq: Interrupt line that is forwarded to a VM
|
* @irq: Interrupt line that is forwarded to a VM
|
||||||
* @which: One of IRQCHIP_STATE_* the caller wants to know about
|
* @which: One of IRQCHIP_STATE_* the caller wants to know about
|
||||||
* @state: a pointer to a boolean where the state is to be storeed
|
* @state: a pointer to a boolean where the state is to be stored
|
||||||
*
|
*
|
||||||
* This call snapshots the internal irqchip state of an
|
* This call snapshots the internal irqchip state of an
|
||||||
* interrupt, returning into @state the bit corresponding to
|
* interrupt, returning into @state the bit corresponding to
|
||||||
|
@ -356,7 +356,7 @@ void irq_matrix_reserve(struct irq_matrix *m)
|
|||||||
* irq_matrix_remove_reserved - Remove interrupt reservation
|
* irq_matrix_remove_reserved - Remove interrupt reservation
|
||||||
* @m: Matrix pointer
|
* @m: Matrix pointer
|
||||||
*
|
*
|
||||||
* This is merily a book keeping call. It decrements the number of globally
|
* This is merely a book keeping call. It decrements the number of globally
|
||||||
* reserved interrupt bits. This is used to undo irq_matrix_reserve() when the
|
* reserved interrupt bits. This is used to undo irq_matrix_reserve() when the
|
||||||
* interrupt was never in use and a real vector allocated, which undid the
|
* interrupt was never in use and a real vector allocated, which undid the
|
||||||
* reservation.
|
* reservation.
|
||||||
|
@ -7,7 +7,7 @@
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* irq_fixup_move_pending - Cleanup irq move pending from a dying CPU
|
* irq_fixup_move_pending - Cleanup irq move pending from a dying CPU
|
||||||
* @desc: Interrupt descpriptor to clean up
|
* @desc: Interrupt descriptor to clean up
|
||||||
* @force_clear: If set clear the move pending bit unconditionally.
|
* @force_clear: If set clear the move pending bit unconditionally.
|
||||||
* If not set, clear it only when the dying CPU is the
|
* If not set, clear it only when the dying CPU is the
|
||||||
* last one in the pending mask.
|
* last one in the pending mask.
|
||||||
|
@ -128,7 +128,7 @@ int check_irq_resend(struct irq_desc *desc, bool inject)
|
|||||||
if (!try_retrigger(desc))
|
if (!try_retrigger(desc))
|
||||||
err = irq_sw_resend(desc);
|
err = irq_sw_resend(desc);
|
||||||
|
|
||||||
/* If the retrigger was successfull, mark it with the REPLAY bit */
|
/* If the retrigger was successful, mark it with the REPLAY bit */
|
||||||
if (!err)
|
if (!err)
|
||||||
desc->istate |= IRQS_REPLAY;
|
desc->istate |= IRQS_REPLAY;
|
||||||
return err;
|
return err;
|
||||||
|
@ -84,7 +84,7 @@ void irq_timings_disable(void)
|
|||||||
* 2. Log interval
|
* 2. Log interval
|
||||||
*
|
*
|
||||||
* We saw the irq timings allow to compute the interval of the
|
* We saw the irq timings allow to compute the interval of the
|
||||||
* occurrences for a specific interrupt. We can reasonibly assume the
|
* occurrences for a specific interrupt. We can reasonably assume the
|
||||||
* longer is the interval, the higher is the error for the next event
|
* longer is the interval, the higher is the error for the next event
|
||||||
* and we can consider storing those interval values into an array
|
* and we can consider storing those interval values into an array
|
||||||
* where each slot in the array correspond to an interval at the power
|
* where each slot in the array correspond to an interval at the power
|
||||||
@ -416,7 +416,7 @@ static u64 __irq_timings_next_event(struct irqt_stat *irqs, int irq, u64 now)
|
|||||||
* Copy the content of the circular buffer into another buffer
|
* Copy the content of the circular buffer into another buffer
|
||||||
* in order to linearize the buffer instead of dealing with
|
* in order to linearize the buffer instead of dealing with
|
||||||
* wrapping indexes and shifted array which will be prone to
|
* wrapping indexes and shifted array which will be prone to
|
||||||
* error and extremelly difficult to debug.
|
* error and extremely difficult to debug.
|
||||||
*/
|
*/
|
||||||
for (i = 0; i < count; i++) {
|
for (i = 0; i < count; i++) {
|
||||||
int index = (start + i) & IRQ_TIMINGS_MASK;
|
int index = (start + i) & IRQ_TIMINGS_MASK;
|
||||||
@ -514,7 +514,7 @@ static inline void irq_timings_store(int irq, struct irqt_stat *irqs, u64 ts)
|
|||||||
* If more than the array size interrupts happened during the
|
* If more than the array size interrupts happened during the
|
||||||
* last busy/idle cycle, the index wrapped up and we have to
|
* last busy/idle cycle, the index wrapped up and we have to
|
||||||
* begin with the next element in the array which is the last one
|
* begin with the next element in the array which is the last one
|
||||||
* in the sequence, otherwise it is a the index 0.
|
* in the sequence, otherwise it is at the index 0.
|
||||||
*
|
*
|
||||||
* - have an indication of the interrupts activity on this CPU
|
* - have an indication of the interrupts activity on this CPU
|
||||||
* (eg. irq/sec)
|
* (eg. irq/sec)
|
||||||
|
Loading…
Reference in New Issue
Block a user