On Thu, 5 May 2022, Joel Stanley wrote: > On Sat, 30 Apr 2022 at 18:58, Julia Lawall <Julia.Lawall@xxxxxxxx> wrote: > > > > Various spelling mistakes in comments. > > Detected with the help of Coccinelle. > > > > Signed-off-by: Julia Lawall <Julia.Lawall@xxxxxxxx> > > I read the patch and it appears that all of the corrections are good. > Thanks for sending it Julia. > > I'm not sure that one mega patch is the right way to go or not, so > I'll leave that to mpe. I tried some other options, but wasn't satisfied with any of them. But just let me know if you want something else (by file? by subdirectory?) julia > > Reviewed-by: Joel Stanley <joel@xxxxxxxxx> > > > > > --- > > arch/powerpc/boot/cuboot-hotfoot.c | 2 +- > > arch/powerpc/crypto/aes-spe-glue.c | 2 +- > > arch/powerpc/kernel/cputable.c | 2 +- > > arch/powerpc/kernel/dawr.c | 2 +- > > arch/powerpc/kernel/eeh.c | 4 ++-- > > arch/powerpc/kernel/eeh_event.c | 2 +- > > arch/powerpc/kernel/fadump.c | 10 +++++----- > > arch/powerpc/kernel/module_32.c | 2 +- > > arch/powerpc/kernel/module_64.c | 4 ++-- > > arch/powerpc/kernel/pci-common.c | 2 +- > > arch/powerpc/kernel/pci_of_scan.c | 2 +- > > arch/powerpc/kernel/process.c | 4 ++-- > > arch/powerpc/kernel/prom_init.c | 2 +- > > arch/powerpc/kernel/ptrace/ptrace-view.c | 2 +- > > arch/powerpc/kernel/rtas_flash.c | 2 +- > > arch/powerpc/kernel/setup-common.c | 2 +- > > arch/powerpc/kernel/signal_64.c | 2 +- > > arch/powerpc/kernel/smp.c | 2 +- > > arch/powerpc/kernel/time.c | 4 ++-- > > arch/powerpc/kernel/watchdog.c | 2 +- > > arch/powerpc/kexec/core_64.c | 2 +- > > arch/powerpc/kvm/book3s_64_mmu_hv.c | 2 +- > > arch/powerpc/kvm/book3s_64_vio_hv.c | 2 +- > > arch/powerpc/kvm/book3s_emulate.c | 2 +- > > arch/powerpc/kvm/book3s_hv_p9_entry.c | 2 +- > > arch/powerpc/kvm/book3s_hv_uvmem.c | 2 +- > > arch/powerpc/kvm/book3s_pr.c | 2 +- > > arch/powerpc/kvm/book3s_xics.c | 2 +- > > arch/powerpc/kvm/book3s_xive.c | 6 +++--- > > arch/powerpc/kvm/e500mc.c | 2 +- > > arch/powerpc/mm/book3s64/hash_pgtable.c | 2 +- > > arch/powerpc/mm/book3s64/hash_utils.c | 4 ++-- > > arch/powerpc/mm/book3s64/pgtable.c | 2 +- > > arch/powerpc/mm/book3s64/radix_pgtable.c | 2 +- > > arch/powerpc/mm/book3s64/radix_tlb.c | 2 +- > > arch/powerpc/mm/book3s64/slb.c | 4 ++-- > > arch/powerpc/mm/init_64.c | 4 ++-- > > arch/powerpc/mm/nohash/book3e_hugetlbpage.c | 2 +- > > arch/powerpc/mm/nohash/kaslr_booke.c | 2 +- > > arch/powerpc/mm/pgtable-frag.c | 2 +- > > arch/powerpc/perf/8xx-pmu.c | 2 +- > > arch/powerpc/perf/core-book3s.c | 6 +++--- > > arch/powerpc/perf/imc-pmu.c | 4 ++-- > > arch/powerpc/perf/isa207-common.c | 6 +++--- > > arch/powerpc/platforms/512x/clock-commonclk.c | 2 +- > > arch/powerpc/platforms/512x/mpc512x_shared.c | 2 +- > > arch/powerpc/platforms/52xx/mpc52xx_common.c | 2 +- > > arch/powerpc/platforms/52xx/mpc52xx_gpt.c | 2 +- > > arch/powerpc/platforms/52xx/mpc52xx_lpbfifo.c | 2 +- > > arch/powerpc/platforms/85xx/mpc85xx_cds.c | 2 +- > > arch/powerpc/platforms/86xx/gef_ppc9a.c | 2 +- > > arch/powerpc/platforms/86xx/gef_sbc310.c | 2 +- > > arch/powerpc/platforms/86xx/gef_sbc610.c | 2 +- > > arch/powerpc/platforms/book3s/vas-api.c | 2 +- > > arch/powerpc/platforms/cell/cbe_regs.c | 2 +- > > arch/powerpc/platforms/cell/iommu.c | 2 +- > > arch/powerpc/platforms/cell/spider-pci.c | 2 +- > > arch/powerpc/platforms/cell/spu_manage.c | 2 +- > > arch/powerpc/platforms/powermac/low_i2c.c | 2 +- > > arch/powerpc/platforms/powernv/eeh-powernv.c | 10 +++++----- > > arch/powerpc/platforms/powernv/idle.c | 4 ++-- > > arch/powerpc/platforms/powernv/ocxl.c | 2 +- > > arch/powerpc/platforms/powernv/opal-fadump.c | 2 +- > > arch/powerpc/platforms/powernv/opal-lpc.c | 2 +- > > arch/powerpc/platforms/powernv/opal-memory-errors.c | 2 +- > > arch/powerpc/platforms/powernv/pci-sriov.c | 2 +- > > arch/powerpc/platforms/ps3/mm.c | 2 +- > > arch/powerpc/platforms/ps3/system-bus.c | 2 +- > > arch/powerpc/platforms/pseries/eeh_pseries.c | 2 +- > > arch/powerpc/platforms/pseries/iommu.c | 2 +- > > arch/powerpc/platforms/pseries/setup.c | 4 ++-- > > arch/powerpc/platforms/pseries/vas-sysfs.c | 2 +- > > arch/powerpc/platforms/pseries/vas.c | 2 +- > > arch/powerpc/sysdev/fsl_lbc.c | 2 +- > > arch/powerpc/sysdev/fsl_pci.c | 2 +- > > arch/powerpc/sysdev/ge/ge_pic.c | 2 +- > > arch/powerpc/sysdev/mpic_msgr.c | 2 +- > > arch/powerpc/sysdev/mpic_msi.c | 2 +- > > arch/powerpc/sysdev/mpic_timer.c | 2 +- > > arch/powerpc/sysdev/mpic_u3msi.c | 2 +- > > arch/powerpc/sysdev/xive/native.c | 2 +- > > arch/powerpc/xmon/ppc-opc.c | 2 +- > > arch/powerpc/xmon/xmon.c | 2 +- > > 83 files changed, 107 insertions(+), 107 deletions(-) > > > > diff --git a/arch/powerpc/kvm/book3s_64_mmu_hv.c b/arch/powerpc/kvm/book3s_64_mmu_hv.c > > index 0aeb51738ca9..1137c4df726c 100644 > > --- a/arch/powerpc/kvm/book3s_64_mmu_hv.c > > +++ b/arch/powerpc/kvm/book3s_64_mmu_hv.c > > @@ -58,7 +58,7 @@ struct kvm_resize_hpt { > > /* Possible values and their usage: > > * <0 an error occurred during allocation, > > * -EBUSY allocation is in the progress, > > - * 0 allocation made successfuly. > > + * 0 allocation made successfully. > > */ > > int error; > > > > diff --git a/arch/powerpc/kvm/book3s_64_vio_hv.c b/arch/powerpc/kvm/book3s_64_vio_hv.c > > index fdeda6a9cff4..fdcc7b287dd8 100644 > > --- a/arch/powerpc/kvm/book3s_64_vio_hv.c > > +++ b/arch/powerpc/kvm/book3s_64_vio_hv.c > > @@ -453,7 +453,7 @@ static long kvmppc_rm_ua_to_hpa(struct kvm_vcpu *vcpu, unsigned long mmu_seq, > > * we are doing this on secondary cpus and current task there > > * is not the hypervisor. Also this is safe against THP in the > > * host, because an IPI to primary thread will wait for the secondary > > - * to exit which will agains result in the below page table walk > > + * to exit which will again result in the below page table walk > > * to finish. > > */ > > /* an rmap lock won't make it safe. because that just ensure hash > > diff --git a/arch/powerpc/kvm/book3s_emulate.c b/arch/powerpc/kvm/book3s_emulate.c > > index fdb57be71aa6..5bbfb2eed127 100644 > > --- a/arch/powerpc/kvm/book3s_emulate.c > > +++ b/arch/powerpc/kvm/book3s_emulate.c > > @@ -268,7 +268,7 @@ int kvmppc_core_emulate_op_pr(struct kvm_vcpu *vcpu, > > > > /* > > * add rules to fit in ISA specification regarding TM > > - * state transistion in TM disable/Suspended state, > > + * state transition in TM disable/Suspended state, > > * and target TM state is TM inactive(00) state. (the > > * change should be suppressed). > > */ > > diff --git a/arch/powerpc/kvm/book3s_hv_p9_entry.c b/arch/powerpc/kvm/book3s_hv_p9_entry.c > > index a28e5b3daabd..ac38c1cad378 100644 > > --- a/arch/powerpc/kvm/book3s_hv_p9_entry.c > > +++ b/arch/powerpc/kvm/book3s_hv_p9_entry.c > > @@ -379,7 +379,7 @@ void restore_p9_host_os_sprs(struct kvm_vcpu *vcpu, > > { > > /* > > * current->thread.xxx registers must all be restored to host > > - * values before a potential context switch, othrewise the context > > + * values before a potential context switch, otherwise the context > > * switch itself will overwrite current->thread.xxx with the values > > * from the guest SPRs. > > */ > > diff --git a/arch/powerpc/kvm/book3s_hv_uvmem.c b/arch/powerpc/kvm/book3s_hv_uvmem.c > > index 45c993dd05f5..0f66f7fbfb80 100644 > > --- a/arch/powerpc/kvm/book3s_hv_uvmem.c > > +++ b/arch/powerpc/kvm/book3s_hv_uvmem.c > > @@ -120,7 +120,7 @@ static DEFINE_SPINLOCK(kvmppc_uvmem_bitmap_lock); > > * content is un-encrypted. > > * > > * (c) Normal - The GFN is a normal. The GFN is associated with > > - * a normal VM. The contents of the GFN is accesible to > > + * a normal VM. The contents of the GFN is accessible to > > * the Hypervisor. Its content is never encrypted. > > * > > * States of a VM. > > diff --git a/arch/powerpc/kvm/book3s_pr.c b/arch/powerpc/kvm/book3s_pr.c > > index 7bf9e6ca5c2d..d6abed6e51e6 100644 > > --- a/arch/powerpc/kvm/book3s_pr.c > > +++ b/arch/powerpc/kvm/book3s_pr.c > > @@ -1287,7 +1287,7 @@ int kvmppc_handle_exit_pr(struct kvm_vcpu *vcpu, unsigned int exit_nr) > > > > /* Get last sc for papr */ > > if (vcpu->arch.papr_enabled) { > > - /* The sc instuction points SRR0 to the next inst */ > > + /* The sc instruction points SRR0 to the next inst */ > > emul = kvmppc_get_last_inst(vcpu, INST_SC, &last_sc); > > if (emul != EMULATE_DONE) { > > kvmppc_set_pc(vcpu, kvmppc_get_pc(vcpu) - 4); > > diff --git a/arch/powerpc/kvm/book3s_xics.c b/arch/powerpc/kvm/book3s_xics.c > > index ab6d37d78c62..589a8f257120 100644 > > --- a/arch/powerpc/kvm/book3s_xics.c > > +++ b/arch/powerpc/kvm/book3s_xics.c > > @@ -462,7 +462,7 @@ static void icp_deliver_irq(struct kvmppc_xics *xics, struct kvmppc_icp *icp, > > * new guy. We cannot assume that the rejected interrupt is less > > * favored than the new one, and thus doesn't need to be delivered, > > * because by the time we exit icp_try_to_deliver() the target > > - * processor may well have alrady consumed & completed it, and thus > > + * processor may well have already consumed & completed it, and thus > > * the rejected interrupt might actually be already acceptable. > > */ > > if (icp_try_to_deliver(icp, new_irq, state->priority, &reject)) { > > diff --git a/arch/powerpc/kvm/book3s_xive.c b/arch/powerpc/kvm/book3s_xive.c > > index c0ce5531d9bc..24d434f1f012 100644 > > --- a/arch/powerpc/kvm/book3s_xive.c > > +++ b/arch/powerpc/kvm/book3s_xive.c > > @@ -124,7 +124,7 @@ void kvmppc_xive_push_vcpu(struct kvm_vcpu *vcpu) > > * interrupt might have fired and be on its way to the > > * host queue while we mask it, and if we unmask it > > * early enough (re-cede right away), there is a > > - * theorical possibility that it fires again, thus > > + * theoretical possibility that it fires again, thus > > * landing in the target queue more than once which is > > * a big no-no. > > * > > @@ -622,7 +622,7 @@ static int xive_target_interrupt(struct kvm *kvm, > > > > /* > > * Targetting rules: In order to avoid losing track of > > - * pending interrupts accross mask and unmask, which would > > + * pending interrupts across mask and unmask, which would > > * allow queue overflows, we implement the following rules: > > * > > * - Unless it was never enabled (or we run out of capacity) > > @@ -1073,7 +1073,7 @@ int kvmppc_xive_clr_mapped(struct kvm *kvm, unsigned long guest_irq, > > /* > > * If old_p is set, the interrupt is pending, we switch it to > > * PQ=11. This will force a resend in the host so the interrupt > > - * isn't lost to whatver host driver may pick it up > > + * isn't lost to whatever host driver may pick it up > > */ > > if (state->old_p) > > xive_vm_esb_load(state->pt_data, XIVE_ESB_SET_PQ_11); > > diff --git a/arch/powerpc/kvm/e500mc.c b/arch/powerpc/kvm/e500mc.c > > index fa0d8dbbe484..4ff1372e48d4 100644 > > --- a/arch/powerpc/kvm/e500mc.c > > +++ b/arch/powerpc/kvm/e500mc.c > > @@ -309,7 +309,7 @@ static int kvmppc_core_vcpu_create_e500mc(struct kvm_vcpu *vcpu) > > BUILD_BUG_ON(offsetof(struct kvmppc_vcpu_e500, vcpu) != 0); > > vcpu_e500 = to_e500(vcpu); > > > > - /* Invalid PIR value -- this LPID dosn't have valid state on any cpu */ > > + /* Invalid PIR value -- this LPID doesn't have valid state on any cpu */ > > vcpu->arch.oldpir = 0xffffffff; > > > > err = kvmppc_e500_tlb_init(vcpu_e500); > > diff --git a/arch/powerpc/kernel/eeh.c b/arch/powerpc/kernel/eeh.c > > index 28bb1e7263a6..ab316e155ea9 100644 > > --- a/arch/powerpc/kernel/eeh.c > > +++ b/arch/powerpc/kernel/eeh.c > > @@ -1329,7 +1329,7 @@ int eeh_pe_set_option(struct eeh_pe *pe, int option) > > > > /* > > * EEH functionality could possibly be disabled, just > > - * return error for the case. And the EEH functinality > > + * return error for the case. And the EEH functionality > > * isn't expected to be disabled on one specific PE. > > */ > > switch (option) { > > @@ -1804,7 +1804,7 @@ static int eeh_debugfs_break_device(struct pci_dev *pdev) > > * PE freeze. Using the in_8() accessor skips the eeh detection hook > > * so the freeze hook so the EEH Detection machinery won't be > > * triggered here. This is to match the usual behaviour of EEH > > - * where the HW will asyncronously freeze a PE and it's up to > > + * where the HW will asynchronously freeze a PE and it's up to > > * the kernel to notice and deal with it. > > * > > * 3. Turn Memory space back on. This is more important for VFs > > diff --git a/arch/powerpc/kernel/eeh_event.c b/arch/powerpc/kernel/eeh_event.c > > index a7a8dc182efb..c23a454af08a 100644 > > --- a/arch/powerpc/kernel/eeh_event.c > > +++ b/arch/powerpc/kernel/eeh_event.c > > @@ -143,7 +143,7 @@ int __eeh_send_failure_event(struct eeh_pe *pe) > > int eeh_send_failure_event(struct eeh_pe *pe) > > { > > /* > > - * If we've manually supressed recovery events via debugfs > > + * If we've manually suppressed recovery events via debugfs > > * then just drop it on the floor. > > */ > > if (eeh_debugfs_no_recover) { > > diff --git a/arch/powerpc/platforms/powernv/eeh-powernv.c b/arch/powerpc/platforms/powernv/eeh-powernv.c > > index 89e22c460ebf..33f7b959c810 100644 > > --- a/arch/powerpc/platforms/powernv/eeh-powernv.c > > +++ b/arch/powerpc/platforms/powernv/eeh-powernv.c > > @@ -390,7 +390,7 @@ static struct eeh_dev *pnv_eeh_probe(struct pci_dev *pdev) > > * should be blocked until PE reset. MMIO access is dropped > > * by hardware certainly. In order to drop PCI config requests, > > * one more flag (EEH_PE_CFG_RESTRICTED) is introduced, which > > - * will be checked in the backend for PE state retrival. If > > + * will be checked in the backend for PE state retrieval. If > > * the PE becomes frozen for the first time and the flag has > > * been set for the PE, we will set EEH_PE_CFG_BLOCKED for > > * that PE to block its config space. > > @@ -981,7 +981,7 @@ static int pnv_eeh_do_af_flr(struct pci_dn *pdn, int option) > > case EEH_RESET_FUNDAMENTAL: > > /* > > * Wait for Transaction Pending bit to clear. A word-aligned > > - * test is used, so we use the conrol offset rather than status > > + * test is used, so we use the control offset rather than status > > * and shift the test bit to match. > > */ > > pnv_eeh_wait_for_pending(pdn, "AF", > > @@ -1048,7 +1048,7 @@ static int pnv_eeh_reset(struct eeh_pe *pe, int option) > > * frozen state during PE reset. However, the good idea here from > > * benh is to keep frozen state before we get PE reset done completely > > * (until BAR restore). With the frozen state, HW drops illegal IO > > - * or MMIO access, which can incur recrusive frozen PE during PE > > + * or MMIO access, which can incur recursive frozen PE during PE > > * reset. The side effect is that EEH core has to clear the frozen > > * state explicitly after BAR restore. > > */ > > @@ -1095,8 +1095,8 @@ static int pnv_eeh_reset(struct eeh_pe *pe, int option) > > * bus is behind a hotplug slot and it will use the slot provided > > * reset methods to prevent spurious hotplug events during the reset. > > * > > - * Fundemental resets need to be handled internally to EEH since the > > - * PCI core doesn't really have a concept of a fundemental reset, > > + * Fundamental resets need to be handled internally to EEH since the > > + * PCI core doesn't really have a concept of a fundamental reset, > > * mainly because there's no standard way to generate one. Only a > > * few devices require an FRESET so it should be fine. > > */ > > diff --git a/arch/powerpc/platforms/pseries/eeh_pseries.c b/arch/powerpc/platforms/pseries/eeh_pseries.c > > index 09fafcf2d3a0..f0b2bca750da 100644 > > --- a/arch/powerpc/platforms/pseries/eeh_pseries.c > > +++ b/arch/powerpc/platforms/pseries/eeh_pseries.c > > @@ -510,7 +510,7 @@ static int pseries_eeh_set_option(struct eeh_pe *pe, int option) > > int ret = 0; > > > > /* > > - * When we're enabling or disabling EEH functioality on > > + * When we're enabling or disabling EEH functionality on > > * the particular PE, the PE config address is possibly > > * unavailable. Therefore, we have to figure it out from > > * the FDT node. > > diff --git a/arch/powerpc/platforms/cell/cbe_regs.c b/arch/powerpc/platforms/cell/cbe_regs.c > > index 1c4c53bec66c..03512a41bd7e 100644 > > --- a/arch/powerpc/platforms/cell/cbe_regs.c > > +++ b/arch/powerpc/platforms/cell/cbe_regs.c > > @@ -23,7 +23,7 @@ > > * Current implementation uses "cpu" nodes. We build our own mapping > > * array of cpu numbers to cpu nodes locally for now to allow interrupt > > * time code to have a fast path rather than call of_get_cpu_node(). If > > - * we implement cpu hotplug, we'll have to install an appropriate norifier > > + * we implement cpu hotplug, we'll have to install an appropriate notifier > > * in order to release references to the cpu going away > > */ > > static struct cbe_regs_map > > diff --git a/arch/powerpc/platforms/cell/iommu.c b/arch/powerpc/platforms/cell/iommu.c > > index 25e726bf0172..3f141cf5e580 100644 > > --- a/arch/powerpc/platforms/cell/iommu.c > > +++ b/arch/powerpc/platforms/cell/iommu.c > > @@ -582,7 +582,7 @@ static int cell_of_bus_notify(struct notifier_block *nb, unsigned long action, > > { > > struct device *dev = data; > > > > - /* We are only intereted in device addition */ > > + /* We are only interested in device addition */ > > if (action != BUS_NOTIFY_ADD_DEVICE) > > return 0; > > > > diff --git a/arch/powerpc/platforms/cell/spider-pci.c b/arch/powerpc/platforms/cell/spider-pci.c > > index a1c293f42a1f..3a2ea8376e32 100644 > > --- a/arch/powerpc/platforms/cell/spider-pci.c > > +++ b/arch/powerpc/platforms/cell/spider-pci.c > > @@ -81,7 +81,7 @@ static int __init spiderpci_pci_setup_chip(struct pci_controller *phb, > > /* > > * On CellBlade, we can't know that which XDR memory is used by > > * kmalloc() to allocate dummy_page_va. > > - * In order to imporve the performance, the XDR which is used to > > + * In order to improve the performance, the XDR which is used to > > * allocate dummy_page_va is the nearest the spider-pci. > > * We have to select the CBE which is the nearest the spider-pci > > * to allocate memory from the best XDR, but I don't know that > > diff --git a/arch/powerpc/platforms/cell/spu_manage.c b/arch/powerpc/platforms/cell/spu_manage.c > > index ddf8742f09a3..080ed2d2c682 100644 > > --- a/arch/powerpc/platforms/cell/spu_manage.c > > +++ b/arch/powerpc/platforms/cell/spu_manage.c > > @@ -457,7 +457,7 @@ static void __init init_affinity_node(int cbe) > > > > /* > > * Walk through each phandle in vicinity property of the spu > > - * (tipically two vicinity phandles per spe node) > > + * (typically two vicinity phandles per spe node) > > */ > > for (i = 0; i < (lenp / sizeof(phandle)); i++) { > > if (vic_handles[i] == avoid_ph) > > diff --git a/arch/powerpc/platforms/powermac/low_i2c.c b/arch/powerpc/platforms/powermac/low_i2c.c > > index df89d916236d..9aded0188ce8 100644 > > --- a/arch/powerpc/platforms/powermac/low_i2c.c > > +++ b/arch/powerpc/platforms/powermac/low_i2c.c > > @@ -1472,7 +1472,7 @@ int __init pmac_i2c_init(void) > > smu_i2c_probe(); > > #endif > > > > - /* Now add plaform functions for some known devices */ > > + /* Now add platform functions for some known devices */ > > pmac_i2c_devscan(pmac_i2c_dev_create); > > > > return 0; > > diff --git a/arch/powerpc/platforms/85xx/mpc85xx_cds.c b/arch/powerpc/platforms/85xx/mpc85xx_cds.c > > index 5bd487030256..ad0b88ecfd0c 100644 > > --- a/arch/powerpc/platforms/85xx/mpc85xx_cds.c > > +++ b/arch/powerpc/platforms/85xx/mpc85xx_cds.c > > @@ -151,7 +151,7 @@ static void __init mpc85xx_cds_pci_irq_fixup(struct pci_dev *dev) > > */ > > case PCI_DEVICE_ID_VIA_82C586_2: > > /* There are two USB controllers. > > - * Identify them by functon number > > + * Identify them by function number > > */ > > if (PCI_FUNC(dev->devfn) == 3) > > dev->irq = 11; > > diff --git a/arch/powerpc/crypto/aes-spe-glue.c b/arch/powerpc/crypto/aes-spe-glue.c > > index c2b23b69d7b1..e8dfe9fb0266 100644 > > --- a/arch/powerpc/crypto/aes-spe-glue.c > > +++ b/arch/powerpc/crypto/aes-spe-glue.c > > @@ -404,7 +404,7 @@ static int ppc_xts_decrypt(struct skcipher_request *req) > > > > /* > > * Algorithm definitions. Disabling alignment (cra_alignmask=0) was chosen > > - * because the e500 platform can handle unaligned reads/writes very efficently. > > + * because the e500 platform can handle unaligned reads/writes very efficiently. > > * This improves IPsec thoughput by another few percent. Additionally we assume > > * that AES context is always aligned to at least 8 bytes because it is created > > * with kmalloc() in the crypto infrastructure > > diff --git a/arch/powerpc/platforms/ps3/mm.c b/arch/powerpc/platforms/ps3/mm.c > > index 5ce924611b94..63ef61ed7597 100644 > > --- a/arch/powerpc/platforms/ps3/mm.c > > +++ b/arch/powerpc/platforms/ps3/mm.c > > @@ -364,7 +364,7 @@ static void __maybe_unused _dma_dump_region(const struct ps3_dma_region *r, > > * @bus_addr: Starting ioc bus address of the area to map. > > * @len: Length in bytes of the area to map. > > * @link: A struct list_head used with struct ps3_dma_region.chunk_list, the > > - * list of all chuncks owned by the region. > > + * list of all chunks owned by the region. > > * > > * This implementation uses a very simple dma page manager > > * based on the dma_chunk structure. This scheme assumes > > diff --git a/arch/powerpc/platforms/ps3/system-bus.c b/arch/powerpc/platforms/ps3/system-bus.c > > index b637bf292047..2502e9b17df4 100644 > > --- a/arch/powerpc/platforms/ps3/system-bus.c > > +++ b/arch/powerpc/platforms/ps3/system-bus.c > > @@ -601,7 +601,7 @@ static dma_addr_t ps3_ioc0_map_page(struct device *_dev, struct page *page, > > iopte_flag |= CBE_IOPTE_PP_W | CBE_IOPTE_SO_RW; > > break; > > default: > > - /* not happned */ > > + /* not happened */ > > BUG(); > > } > > result = ps3_dma_map(dev->d_region, (unsigned long)ptr, size, > > diff --git a/arch/powerpc/platforms/512x/clock-commonclk.c b/arch/powerpc/platforms/512x/clock-commonclk.c > > index 0b03d812baae..0652c7e69225 100644 > > --- a/arch/powerpc/platforms/512x/clock-commonclk.c > > +++ b/arch/powerpc/platforms/512x/clock-commonclk.c > > @@ -663,7 +663,7 @@ static void __init mpc512x_clk_setup_mclk(struct mclk_setup_data *entry, size_t > > * the PSC/MSCAN/SPDIF (serial drivers et al) need the MCLK > > * for their bitrate > > * - in the absence of "aliases" for clocks we need to create > > - * individial 'struct clk' items for whatever might get > > + * individual 'struct clk' items for whatever might get > > * referenced or looked up, even if several of those items are > > * identical from the logical POV (their rate value) > > * - for easier future maintenance and for better reflection of > > diff --git a/arch/powerpc/platforms/512x/mpc512x_shared.c b/arch/powerpc/platforms/512x/mpc512x_shared.c > > index e3411663edad..96d9cf49560d 100644 > > --- a/arch/powerpc/platforms/512x/mpc512x_shared.c > > +++ b/arch/powerpc/platforms/512x/mpc512x_shared.c > > @@ -289,7 +289,7 @@ static void __init mpc512x_setup_diu(void) > > > > /* > > * We do not allocate and configure new area for bitmap buffer > > - * because it would requere copying bitmap data (splash image) > > + * because it would require copying bitmap data (splash image) > > * and so negatively affect boot time. Instead we reserve the > > * already configured frame buffer area so that it won't be > > * destroyed. The starting address of the area to reserve and > > diff --git a/arch/powerpc/platforms/52xx/mpc52xx_common.c b/arch/powerpc/platforms/52xx/mpc52xx_common.c > > index 565e3a83dc9e..60aa6015e284 100644 > > --- a/arch/powerpc/platforms/52xx/mpc52xx_common.c > > +++ b/arch/powerpc/platforms/52xx/mpc52xx_common.c > > @@ -308,7 +308,7 @@ int mpc5200_psc_ac97_gpio_reset(int psc_number) > > > > spin_lock_irqsave(&gpio_lock, flags); > > > > - /* Reconfiure pin-muxing to gpio */ > > + /* Reconfigure pin-muxing to gpio */ > > mux = in_be32(&simple_gpio->port_config); > > out_be32(&simple_gpio->port_config, mux & (~gpio)); > > > > diff --git a/arch/powerpc/platforms/52xx/mpc52xx_gpt.c b/arch/powerpc/platforms/52xx/mpc52xx_gpt.c > > index f862b48b4824..7252d992ca9f 100644 > > --- a/arch/powerpc/platforms/52xx/mpc52xx_gpt.c > > +++ b/arch/powerpc/platforms/52xx/mpc52xx_gpt.c > > @@ -398,7 +398,7 @@ static int mpc52xx_gpt_do_start(struct mpc52xx_gpt_priv *gpt, u64 period, > > set |= MPC52xx_GPT_MODE_CONTINUOUS; > > > > /* Determine the number of clocks in the requested period. 64 bit > > - * arithmatic is done here to preserve the precision until the value > > + * arithmetic is done here to preserve the precision until the value > > * is scaled back down into the u32 range. Period is in 'ns', bus > > * frequency is in Hz. */ > > clocks = period * (u64)gpt->ipb_freq; > > diff --git a/arch/powerpc/platforms/52xx/mpc52xx_lpbfifo.c b/arch/powerpc/platforms/52xx/mpc52xx_lpbfifo.c > > index b91ebebd9ff2..54dfc63809d3 100644 > > --- a/arch/powerpc/platforms/52xx/mpc52xx_lpbfifo.c > > +++ b/arch/powerpc/platforms/52xx/mpc52xx_lpbfifo.c > > @@ -104,7 +104,7 @@ static void mpc52xx_lpbfifo_kick(struct mpc52xx_lpbfifo_request *req) > > * > > * Configure the watermarks so DMA will always complete correctly. > > * It may be worth experimenting with the ALARM value to see if > > - * there is a performance impacit. However, if it is wrong there > > + * there is a performance impact. However, if it is wrong there > > * is a risk of DMA not transferring the last chunk of data > > */ > > if (write) { > > diff --git a/arch/powerpc/boot/cuboot-hotfoot.c b/arch/powerpc/boot/cuboot-hotfoot.c > > index 888a6b9bfead..0e5532f855d6 100644 > > --- a/arch/powerpc/boot/cuboot-hotfoot.c > > +++ b/arch/powerpc/boot/cuboot-hotfoot.c > > @@ -70,7 +70,7 @@ static void hotfoot_fixups(void) > > > > printf("Fixing devtree for 4M Flash\n"); > > > > - /* First fix up the base addresse */ > > + /* First fix up the base address */ > > getprop(devp, "reg", regs, sizeof(regs)); > > regs[0] = 0; > > regs[1] = 0xffc00000; > > diff --git a/arch/powerpc/kernel/cputable.c b/arch/powerpc/kernel/cputable.c > > index ae0fdef0ac11..2a271a6d6924 100644 > > --- a/arch/powerpc/kernel/cputable.c > > +++ b/arch/powerpc/kernel/cputable.c > > @@ -2025,7 +2025,7 @@ static struct cpu_spec * __init setup_cpu_spec(unsigned long offset, > > * oprofile_cpu_type already has a value, then we are > > * possibly overriding a real PVR with a logical one, > > * and, in that case, keep the current value for > > - * oprofile_cpu_type. Futhermore, let's ensure that the > > + * oprofile_cpu_type. Furthermore, let's ensure that the > > * fix for the PMAO bug is enabled on compatibility mode. > > */ > > if (old.oprofile_cpu_type != NULL) { > > diff --git a/arch/powerpc/kernel/dawr.c b/arch/powerpc/kernel/dawr.c > > index 64e423d2fe0f..30d4eca88d17 100644 > > --- a/arch/powerpc/kernel/dawr.c > > +++ b/arch/powerpc/kernel/dawr.c > > @@ -27,7 +27,7 @@ int set_dawr(int nr, struct arch_hw_breakpoint *brk) > > dawrx |= (brk->type & (HW_BRK_TYPE_PRIV_ALL)) >> 3; > > /* > > * DAWR length is stored in field MDR bits 48:53. Matches range in > > - * doublewords (64 bits) baised by -1 eg. 0b000000=1DW and > > + * doublewords (64 bits) biased by -1 eg. 0b000000=1DW and > > * 0b111111=64DW. > > * brk->hw_len is in bytes. > > * This aligns up to double word size, shifts and does the bias. > > diff --git a/arch/powerpc/kernel/fadump.c b/arch/powerpc/kernel/fadump.c > > index 4c09c6688ac6..2451fabe4d45 100644 > > --- a/arch/powerpc/kernel/fadump.c > > +++ b/arch/powerpc/kernel/fadump.c > > @@ -73,9 +73,9 @@ static struct cma *fadump_cma; > > * The total size of fadump reserved memory covers for boot memory size > > * + cpu data size + hpte size and metadata. > > * Initialize only the area equivalent to boot memory size for CMA use. > > - * The reamining portion of fadump reserved memory will be not given > > - * to CMA and pages for thoes will stay reserved. boot memory size is > > - * aligned per CMA requirement to satisy cma_init_reserved_mem() call. > > + * The remaining portion of fadump reserved memory will be not given > > + * to CMA and pages for those will stay reserved. boot memory size is > > + * aligned per CMA requirement to satisfy cma_init_reserved_mem() call. > > * But for some reason even if it fails we still have the memory reservation > > * with us and we can still continue doing fadump. > > */ > > @@ -1661,8 +1661,8 @@ int __init setup_fadump(void) > > } > > /* > > * Use subsys_initcall_sync() here because there is dependency with > > - * crash_save_vmcoreinfo_init(), which mush run first to ensure vmcoreinfo initialization > > - * is done before regisering with f/w. > > + * crash_save_vmcoreinfo_init(), which must run first to ensure vmcoreinfo initialization > > + * is done before registering with f/w. > > */ > > subsys_initcall_sync(setup_fadump); > > #else /* !CONFIG_PRESERVE_FA_DUMP */ > > diff --git a/arch/powerpc/kernel/module_32.c b/arch/powerpc/kernel/module_32.c > > index a0432ef46967..e25b796682cc 100644 > > --- a/arch/powerpc/kernel/module_32.c > > +++ b/arch/powerpc/kernel/module_32.c > > @@ -99,7 +99,7 @@ static unsigned long get_plt_size(const Elf32_Ehdr *hdr, > > > > /* Sort the relocation information based on a symbol and > > * addend key. This is a stable O(n*log n) complexity > > - * alogrithm but it will reduce the complexity of > > + * algorithm but it will reduce the complexity of > > * count_relocs() to linear complexity O(n) > > */ > > sort((void *)hdr + sechdrs[i].sh_offset, > > diff --git a/arch/powerpc/kernel/module_64.c b/arch/powerpc/kernel/module_64.c > > index 794720530442..2cce576edbc5 100644 > > --- a/arch/powerpc/kernel/module_64.c > > +++ b/arch/powerpc/kernel/module_64.c > > @@ -194,7 +194,7 @@ static unsigned long get_stubs_size(const Elf64_Ehdr *hdr, > > > > /* Sort the relocation information based on a symbol and > > * addend key. This is a stable O(n*log n) complexity > > - * alogrithm but it will reduce the complexity of > > + * algorithm but it will reduce the complexity of > > * count_relocs() to linear complexity O(n) > > */ > > sort((void *)sechdrs[i].sh_addr, > > @@ -361,7 +361,7 @@ static inline int create_ftrace_stub(struct ppc64_stub_entry *entry, > > entry->jump[1] |= PPC_HA(reladdr); > > entry->jump[2] |= PPC_LO(reladdr); > > > > - /* Eventhough we don't use funcdata in the stub, it's needed elsewhere. */ > > + /* Even though we don't use funcdata in the stub, it's needed elsewhere. */ > > entry->funcdata = func_desc(addr); > > entry->magic = STUB_MAGIC; > > > > diff --git a/arch/powerpc/kernel/pci-common.c b/arch/powerpc/kernel/pci-common.c > > index 8bc9cf62cd93..9a97a93bd11c 100644 > > --- a/arch/powerpc/kernel/pci-common.c > > +++ b/arch/powerpc/kernel/pci-common.c > > @@ -1688,7 +1688,7 @@ EXPORT_SYMBOL_GPL(pcibios_scan_phb); > > static void fixup_hide_host_resource_fsl(struct pci_dev *dev) > > { > > int i, class = dev->class >> 8; > > - /* When configured as agent, programing interface = 1 */ > > + /* When configured as agent, programming interface = 1 */ > > int prog_if = dev->class & 0xf; > > > > if ((class == PCI_CLASS_PROCESSOR_POWERPC || > > diff --git a/arch/powerpc/kernel/pci_of_scan.c b/arch/powerpc/kernel/pci_of_scan.c > > index c3024f104765..6f2b0cc1ddd6 100644 > > --- a/arch/powerpc/kernel/pci_of_scan.c > > +++ b/arch/powerpc/kernel/pci_of_scan.c > > @@ -244,7 +244,7 @@ EXPORT_SYMBOL(of_create_pci_dev); > > * @dev: pci_dev structure for the bridge > > * > > * of_scan_bus() calls this routine for each PCI bridge that it finds, and > > - * this routine in turn call of_scan_bus() recusively to scan for more child > > + * this routine in turn call of_scan_bus() recursively to scan for more child > > * devices. > > */ > > void of_scan_pci_bridge(struct pci_dev *dev) > > diff --git a/arch/powerpc/kernel/process.c b/arch/powerpc/kernel/process.c > > index 984813a4d5dc..eeb2dff95f9a 100644 > > --- a/arch/powerpc/kernel/process.c > > +++ b/arch/powerpc/kernel/process.c > > @@ -307,7 +307,7 @@ static void __giveup_vsx(struct task_struct *tsk) > > unsigned long msr = tsk->thread.regs->msr; > > > > /* > > - * We should never be ssetting MSR_VSX without also setting > > + * We should never be setting MSR_VSX without also setting > > * MSR_FP and MSR_VEC > > */ > > WARN_ON((msr & MSR_VSX) && !((msr & MSR_FP) && (msr & MSR_VEC))); > > @@ -645,7 +645,7 @@ static void do_break_handler(struct pt_regs *regs) > > return; > > } > > > > - /* Otherwise findout which DAWR caused exception and disable it. */ > > + /* Otherwise find out which DAWR caused exception and disable it. */ > > wp_get_instr_detail(regs, &instr, &type, &size, &ea); > > > > for (i = 0; i < nr_wp_slots(); i++) { > > diff --git a/arch/powerpc/kernel/prom_init.c b/arch/powerpc/kernel/prom_init.c > > index 0ac5faacc909..ace861ec4c4c 100644 > > --- a/arch/powerpc/kernel/prom_init.c > > +++ b/arch/powerpc/kernel/prom_init.c > > @@ -3416,7 +3416,7 @@ unsigned long __init prom_init(unsigned long r3, unsigned long r4, > > * > > * PowerMacs use a different mechanism to spin CPUs > > * > > - * (This must be done after instanciating RTAS) > > + * (This must be done after instantiating RTAS) > > */ > > if (of_platform != PLATFORM_POWERMAC) > > prom_hold_cpus(); > > diff --git a/arch/powerpc/kernel/ptrace/ptrace-view.c b/arch/powerpc/kernel/ptrace/ptrace-view.c > > index f15bc78caf71..076d867412c7 100644 > > --- a/arch/powerpc/kernel/ptrace/ptrace-view.c > > +++ b/arch/powerpc/kernel/ptrace/ptrace-view.c > > @@ -174,7 +174,7 @@ int ptrace_get_reg(struct task_struct *task, int regno, unsigned long *data) > > > > /* > > * softe copies paca->irq_soft_mask variable state. Since irq_soft_mask is > > - * no more used as a flag, lets force usr to alway see the softe value as 1 > > + * no more used as a flag, lets force usr to always see the softe value as 1 > > * which means interrupts are not soft disabled. > > */ > > if (IS_ENABLED(CONFIG_PPC64) && regno == PT_SOFTE) { > > diff --git a/arch/powerpc/kernel/rtas_flash.c b/arch/powerpc/kernel/rtas_flash.c > > index a99179d83538..bc817a5619d6 100644 > > --- a/arch/powerpc/kernel/rtas_flash.c > > +++ b/arch/powerpc/kernel/rtas_flash.c > > @@ -120,7 +120,7 @@ static struct kmem_cache *flash_block_cache = NULL; > > /* > > * Local copy of the flash block list. > > * > > - * The rtas_firmware_flash_list varable will be > > + * The rtas_firmware_flash_list variable will be > > * set once the data is fully read. > > * > > * For convenience as we build the list we use virtual addrs, > > diff --git a/arch/powerpc/kernel/setup-common.c b/arch/powerpc/kernel/setup-common.c > > index 518ae5aa9410..3acf2782acdf 100644 > > --- a/arch/powerpc/kernel/setup-common.c > > +++ b/arch/powerpc/kernel/setup-common.c > > @@ -279,7 +279,7 @@ static int show_cpuinfo(struct seq_file *m, void *v) > > proc_freq / 1000000, proc_freq % 1000000); > > > > /* If we are a Freescale core do a simple check so > > - * we dont have to keep adding cases in the future */ > > + * we don't have to keep adding cases in the future */ > > if (PVR_VER(pvr) & 0x8000) { > > switch (PVR_VER(pvr)) { > > case 0x8000: /* 7441/7450/7451, Voyager */ > > diff --git a/arch/powerpc/kernel/signal_64.c b/arch/powerpc/kernel/signal_64.c > > index 73d483b07ff3..858fc13b8c51 100644 > > --- a/arch/powerpc/kernel/signal_64.c > > +++ b/arch/powerpc/kernel/signal_64.c > > @@ -123,7 +123,7 @@ static long notrace __unsafe_setup_sigcontext(struct sigcontext __user *sc, > > #endif > > struct pt_regs *regs = tsk->thread.regs; > > unsigned long msr = regs->msr; > > - /* Force usr to alway see softe as 1 (interrupts enabled) */ > > + /* Force usr to always see softe as 1 (interrupts enabled) */ > > unsigned long softe = 0x1; > > > > BUG_ON(tsk != current); > > diff --git a/arch/powerpc/kernel/smp.c b/arch/powerpc/kernel/smp.c > > index 32156d98dfc8..32910af76a36 100644 > > --- a/arch/powerpc/kernel/smp.c > > +++ b/arch/powerpc/kernel/smp.c > > @@ -1102,7 +1102,7 @@ void __init smp_prepare_cpus(unsigned int max_cpus) > > DBG("smp_prepare_cpus\n"); > > > > /* > > - * setup_cpu may need to be called on the boot cpu. We havent > > + * setup_cpu may need to be called on the boot cpu. We haven't > > * spun any cpus up but lets be paranoid. > > */ > > BUG_ON(boot_cpuid != smp_processor_id()); > > diff --git a/arch/powerpc/kernel/time.c b/arch/powerpc/kernel/time.c > > index f80cce0e3899..4bf757ebe13d 100644 > > --- a/arch/powerpc/kernel/time.c > > +++ b/arch/powerpc/kernel/time.c > > @@ -829,7 +829,7 @@ static void __read_persistent_clock(struct timespec64 *ts) > > static int first = 1; > > > > ts->tv_nsec = 0; > > - /* XXX this is a litle fragile but will work okay in the short term */ > > + /* XXX this is a little fragile but will work okay in the short term */ > > if (first) { > > first = 0; > > if (ppc_md.time_init) > > @@ -974,7 +974,7 @@ void secondary_cpu_time_init(void) > > */ > > start_cpu_decrementer(); > > > > - /* FIME: Should make unrelatred change to move snapshot_timebase > > + /* FIME: Should make unrelated change to move snapshot_timebase > > * call here ! */ > > register_decrementer_clockevent(smp_processor_id()); > > } > > diff --git a/arch/powerpc/kernel/watchdog.c b/arch/powerpc/kernel/watchdog.c > > index 62937a077de7..5b9eb7640a15 100644 > > --- a/arch/powerpc/kernel/watchdog.c > > +++ b/arch/powerpc/kernel/watchdog.c > > @@ -56,7 +56,7 @@ > > * solved by also having a SMP watchdog where all CPUs check all other > > * CPUs heartbeat. > > * > > - * The SMP checker can detect lockups on other CPUs. A gobal "pending" > > + * The SMP checker can detect lockups on other CPUs. A global "pending" > > * cpumask is kept, containing all CPUs which enable the watchdog. Each > > * CPU clears their pending bit in their heartbeat timer. When the bitmask > > * becomes empty, the last CPU to clear its pending bit updates a global > > diff --git a/arch/powerpc/kexec/core_64.c b/arch/powerpc/kexec/core_64.c > > index 6cc7793b8420..c29c639551fe 100644 > > --- a/arch/powerpc/kexec/core_64.c > > +++ b/arch/powerpc/kexec/core_64.c > > @@ -406,7 +406,7 @@ static int __init export_htab_values(void) > > if (!node) > > return -ENODEV; > > > > - /* remove any stale propertys so ours can be found */ > > + /* remove any stale properties so ours can be found */ > > of_remove_property(node, of_find_property(node, htab_base_prop.name, NULL)); > > of_remove_property(node, of_find_property(node, htab_size_prop.name, NULL)); > > > > diff --git a/arch/powerpc/mm/book3s64/hash_pgtable.c b/arch/powerpc/mm/book3s64/hash_pgtable.c > > index 7ce8914992e3..2e0cad5817ba 100644 > > --- a/arch/powerpc/mm/book3s64/hash_pgtable.c > > +++ b/arch/powerpc/mm/book3s64/hash_pgtable.c > > @@ -377,7 +377,7 @@ int hash__has_transparent_hugepage(void) > > if (mmu_psize_defs[MMU_PAGE_16M].shift != PMD_SHIFT) > > return 0; > > /* > > - * We need to make sure that we support 16MB hugepage in a segement > > + * We need to make sure that we support 16MB hugepage in a segment > > * with base page size 64K or 4K. We only enable THP with a PAGE_SIZE > > * of 64K. > > */ > > diff --git a/arch/powerpc/mm/book3s64/hash_utils.c b/arch/powerpc/mm/book3s64/hash_utils.c > > index 985cabdd7f67..5b69b271707e 100644 > > --- a/arch/powerpc/mm/book3s64/hash_utils.c > > +++ b/arch/powerpc/mm/book3s64/hash_utils.c > > @@ -1343,7 +1343,7 @@ static int subpage_protection(struct mm_struct *mm, unsigned long ea) > > spp >>= 30 - 2 * ((ea >> 12) & 0xf); > > > > /* > > - * 0 -> full premission > > + * 0 -> full permission > > * 1 -> Read only > > * 2 -> no access. > > * We return the flag that need to be cleared. > > @@ -1664,7 +1664,7 @@ DEFINE_INTERRUPT_HANDLER(do_hash_fault) > > > > err = hash_page_mm(mm, ea, access, TRAP(regs), flags); > > if (unlikely(err < 0)) { > > - // failed to instert a hash PTE due to an hypervisor error > > + // failed to insert a hash PTE due to an hypervisor error > > if (user_mode(regs)) { > > if (IS_ENABLED(CONFIG_PPC_SUBPAGE_PROT) && err == -2) > > _exception(SIGSEGV, regs, SEGV_ACCERR, ea); > > diff --git a/arch/powerpc/mm/book3s64/pgtable.c b/arch/powerpc/mm/book3s64/pgtable.c > > index 8b474ab32f67..7b9966402b25 100644 > > --- a/arch/powerpc/mm/book3s64/pgtable.c > > +++ b/arch/powerpc/mm/book3s64/pgtable.c > > @@ -332,7 +332,7 @@ static pmd_t *__alloc_for_pmdcache(struct mm_struct *mm) > > spin_lock(&mm->page_table_lock); > > /* > > * If we find pgtable_page set, we return > > - * the allocated page with single fragement > > + * the allocated page with single fragment > > * count. > > */ > > if (likely(!mm->context.pmd_frag)) { > > diff --git a/arch/powerpc/mm/book3s64/radix_pgtable.c b/arch/powerpc/mm/book3s64/radix_pgtable.c > > index def04631a74d..db2f3d193448 100644 > > --- a/arch/powerpc/mm/book3s64/radix_pgtable.c > > +++ b/arch/powerpc/mm/book3s64/radix_pgtable.c > > @@ -359,7 +359,7 @@ static void __init radix_init_pgtable(void) > > if (!cpu_has_feature(CPU_FTR_HVMODE) && > > cpu_has_feature(CPU_FTR_P9_RADIX_PREFETCH_BUG)) { > > /* > > - * Older versions of KVM on these machines perfer if the > > + * Older versions of KVM on these machines prefer if the > > * guest only uses the low 19 PID bits. > > */ > > mmu_pid_bits = 19; > > diff --git a/arch/powerpc/mm/book3s64/radix_tlb.c b/arch/powerpc/mm/book3s64/radix_tlb.c > > index 7724af19ed7e..dda51fef2d2e 100644 > > --- a/arch/powerpc/mm/book3s64/radix_tlb.c > > +++ b/arch/powerpc/mm/book3s64/radix_tlb.c > > @@ -397,7 +397,7 @@ static inline void _tlbie_pid(unsigned long pid, unsigned long ric) > > > > /* > > * Workaround the fact that the "ric" argument to __tlbie_pid > > - * must be a compile-time contraint to match the "i" constraint > > + * must be a compile-time constraint to match the "i" constraint > > * in the asm statement. > > */ > > switch (ric) { > > diff --git a/arch/powerpc/mm/book3s64/slb.c b/arch/powerpc/mm/book3s64/slb.c > > index 81091b9587f6..6956f637a38c 100644 > > --- a/arch/powerpc/mm/book3s64/slb.c > > +++ b/arch/powerpc/mm/book3s64/slb.c > > @@ -347,7 +347,7 @@ void slb_setup_new_exec(void) > > /* > > * We have no good place to clear the slb preload cache on exec, > > * flush_thread is about the earliest arch hook but that happens > > - * after we switch to the mm and have aleady preloaded the SLBEs. > > + * after we switch to the mm and have already preloaded the SLBEs. > > * > > * For the most part that's probably okay to use entries from the > > * previous exec, they will age out if unused. It may turn out to > > @@ -615,7 +615,7 @@ static void slb_cache_update(unsigned long esid_data) > > } else { > > /* > > * Our cache is full and the current cache content strictly > > - * doesn't indicate the active SLB conents. Bump the ptr > > + * doesn't indicate the active SLB contents. Bump the ptr > > * so that switch_slb() will ignore the cache. > > */ > > local_paca->slb_cache_ptr = SLB_CACHE_ENTRIES + 1; > > diff --git a/arch/powerpc/mm/init_64.c b/arch/powerpc/mm/init_64.c > > index 83c0ee9fbf05..2e11952057f8 100644 > > --- a/arch/powerpc/mm/init_64.c > > +++ b/arch/powerpc/mm/init_64.c > > @@ -111,7 +111,7 @@ static int __meminit vmemmap_populated(unsigned long vmemmap_addr, int vmemmap_m > > } > > > > /* > > - * vmemmap virtual address space management does not have a traditonal page > > + * vmemmap virtual address space management does not have a traditional page > > * table to track which virtual struct pages are backed by physical mapping. > > * The virtual to physical mappings are tracked in a simple linked list > > * format. 'vmemmap_list' maintains the entire vmemmap physical mapping at > > @@ -128,7 +128,7 @@ static struct vmemmap_backing *next; > > > > /* > > * The same pointer 'next' tracks individual chunks inside the allocated > > - * full page during the boot time and again tracks the freeed nodes during > > + * full page during the boot time and again tracks the freed nodes during > > * runtime. It is racy but it does not happen as they are separated by the > > * boot process. Will create problem if some how we have memory hotplug > > * operation during boot !! > > diff --git a/arch/powerpc/mm/nohash/book3e_hugetlbpage.c b/arch/powerpc/mm/nohash/book3e_hugetlbpage.c > > index 8b88be91b622..307ca919d393 100644 > > --- a/arch/powerpc/mm/nohash/book3e_hugetlbpage.c > > +++ b/arch/powerpc/mm/nohash/book3e_hugetlbpage.c > > @@ -142,7 +142,7 @@ book3e_hugetlb_preload(struct vm_area_struct *vma, unsigned long ea, pte_t pte) > > tsize = shift - 10; > > /* > > * We can't be interrupted while we're setting up the MAS > > - * regusters or after we've confirmed that no tlb exists. > > + * registers or after we've confirmed that no tlb exists. > > */ > > local_irq_save(flags); > > > > diff --git a/arch/powerpc/mm/nohash/kaslr_booke.c b/arch/powerpc/mm/nohash/kaslr_booke.c > > index 96c38f971603..636ec583d989 100644 > > --- a/arch/powerpc/mm/nohash/kaslr_booke.c > > +++ b/arch/powerpc/mm/nohash/kaslr_booke.c > > @@ -315,7 +315,7 @@ static unsigned long __init kaslr_choose_location(void *dt_ptr, phys_addr_t size > > ram = map_mem_in_cams(ram, CONFIG_LOWMEM_CAM_NUM, true, true); > > linear_sz = min_t(unsigned long, ram, SZ_512M); > > > > - /* If the linear size is smaller than 64M, do not randmize */ > > + /* If the linear size is smaller than 64M, do not randomize */ > > if (linear_sz < SZ_64M) > > return 0; > > > > diff --git a/arch/powerpc/mm/pgtable-frag.c b/arch/powerpc/mm/pgtable-frag.c > > index 97ae4935da79..20652daa1d7e 100644 > > --- a/arch/powerpc/mm/pgtable-frag.c > > +++ b/arch/powerpc/mm/pgtable-frag.c > > @@ -83,7 +83,7 @@ static pte_t *__alloc_for_ptecache(struct mm_struct *mm, int kernel) > > spin_lock(&mm->page_table_lock); > > /* > > * If we find pgtable_page set, we return > > - * the allocated page with single fragement > > + * the allocated page with single fragment > > * count. > > */ > > if (likely(!pte_frag_get(&mm->context))) { > > diff --git a/arch/powerpc/perf/8xx-pmu.c b/arch/powerpc/perf/8xx-pmu.c > > index 4738c4dbf567..308a2e40d7be 100644 > > --- a/arch/powerpc/perf/8xx-pmu.c > > +++ b/arch/powerpc/perf/8xx-pmu.c > > @@ -157,7 +157,7 @@ static void mpc8xx_pmu_del(struct perf_event *event, int flags) > > > > mpc8xx_pmu_read(event); > > > > - /* If it was the last user, stop counting to avoid useles overhead */ > > + /* If it was the last user, stop counting to avoid useless overhead */ > > switch (event_type(event)) { > > case PERF_8xx_ID_CPU_CYCLES: > > break; > > diff --git a/arch/powerpc/perf/core-book3s.c b/arch/powerpc/perf/core-book3s.c > > index b5b42cf0a703..140502a7fdf8 100644 > > --- a/arch/powerpc/perf/core-book3s.c > > +++ b/arch/powerpc/perf/core-book3s.c > > @@ -1142,7 +1142,7 @@ static u64 check_and_compute_delta(u64 prev, u64 val) > > /* > > * POWER7 can roll back counter values, if the new value is smaller > > * than the previous value it will cause the delta and the counter to > > - * have bogus values unless we rolled a counter over. If a coutner is > > + * have bogus values unless we rolled a counter over. If a counter is > > * rolled back, it will be smaller, but within 256, which is the maximum > > * number of events to rollback at once. If we detect a rollback > > * return 0. This can lead to a small lack of precision in the > > @@ -2057,7 +2057,7 @@ static int power_pmu_event_init(struct perf_event *event) > > /* > > * PMU config registers have fields that are > > * reserved and some specific values for bit fields are reserved. > > - * For ex., MMCRA[61:62] is Randome Sampling Mode (SM) > > + * For ex., MMCRA[61:62] is Random Sampling Mode (SM) > > * and value of 0b11 to this field is reserved. > > * Check for invalid values in attr.config. > > */ > > @@ -2447,7 +2447,7 @@ static void __perf_event_interrupt(struct pt_regs *regs) > > } > > > > /* > > - * During system wide profling or while specific CPU is monitored for an > > + * During system wide profiling or while specific CPU is monitored for an > > * event, some corner cases could cause PMC to overflow in idle path. This > > * will trigger a PMI after waking up from idle. Since counter values are _not_ > > * saved/restored in idle path, can lead to below "Can't find PMC" message. > > diff --git a/arch/powerpc/perf/imc-pmu.c b/arch/powerpc/perf/imc-pmu.c > > index 526d4b767534..498f1a2f7658 100644 > > --- a/arch/powerpc/perf/imc-pmu.c > > +++ b/arch/powerpc/perf/imc-pmu.c > > @@ -521,7 +521,7 @@ static int nest_imc_event_init(struct perf_event *event) > > > > /* > > * Nest HW counter memory resides in a per-chip reserve-memory (HOMER). > > - * Get the base memory addresss for this cpu. > > + * Get the base memory address for this cpu. > > */ > > chip_id = cpu_to_chip_id(event->cpu); > > > > @@ -674,7 +674,7 @@ static int ppc_core_imc_cpu_offline(unsigned int cpu) > > /* > > * Check whether core_imc is registered. We could end up here > > * if the cpuhotplug callback registration fails. i.e, callback > > - * invokes the offline path for all sucessfully registered cpus. > > + * invokes the offline path for all successfully registered cpus. > > * At this stage, core_imc pmu will not be registered and we > > * should return here. > > * > > diff --git a/arch/powerpc/perf/isa207-common.c b/arch/powerpc/perf/isa207-common.c > > index a74d382ecbb7..839cc68212c1 100644 > > --- a/arch/powerpc/perf/isa207-common.c > > +++ b/arch/powerpc/perf/isa207-common.c > > @@ -82,11 +82,11 @@ static unsigned long sdar_mod_val(u64 event) > > static void mmcra_sdar_mode(u64 event, unsigned long *mmcra) > > { > > /* > > - * MMCRA[SDAR_MODE] specifices how the SDAR should be updated in > > - * continous sampling mode. > > + * MMCRA[SDAR_MODE] specifies how the SDAR should be updated in > > + * continuous sampling mode. > > * > > * Incase of Power8: > > - * MMCRA[SDAR_MODE] will be programmed as "0b01" for continous sampling > > + * MMCRA[SDAR_MODE] will be programmed as "0b01" for continuous sampling > > * mode and will be un-changed when setting MMCRA[63] (Marked events). > > * > > * Incase of Power9/power10: > > diff --git a/arch/powerpc/platforms/86xx/gef_ppc9a.c b/arch/powerpc/platforms/86xx/gef_ppc9a.c > > index 44bbbc535e1d..884da08806ce 100644 > > --- a/arch/powerpc/platforms/86xx/gef_ppc9a.c > > +++ b/arch/powerpc/platforms/86xx/gef_ppc9a.c > > @@ -180,7 +180,7 @@ DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NEC, PCI_DEVICE_ID_NEC_USB, > > * > > * This function is called to determine whether the BSP is compatible with the > > * supplied device-tree, which is assumed to be the correct one for the actual > > - * board. It is expected thati, in the future, a kernel may support multiple > > + * board. It is expected that, in the future, a kernel may support multiple > > * boards. > > */ > > static int __init gef_ppc9a_probe(void) > > diff --git a/arch/powerpc/platforms/86xx/gef_sbc310.c b/arch/powerpc/platforms/86xx/gef_sbc310.c > > index 46d6d3d4957a..baaf1ab07016 100644 > > --- a/arch/powerpc/platforms/86xx/gef_sbc310.c > > +++ b/arch/powerpc/platforms/86xx/gef_sbc310.c > > @@ -167,7 +167,7 @@ DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NEC, PCI_DEVICE_ID_NEC_USB, > > * > > * This function is called to determine whether the BSP is compatible with the > > * supplied device-tree, which is assumed to be the correct one for the actual > > - * board. It is expected thati, in the future, a kernel may support multiple > > + * board. It is expected that, in the future, a kernel may support multiple > > * boards. > > */ > > static int __init gef_sbc310_probe(void) > > diff --git a/arch/powerpc/platforms/86xx/gef_sbc610.c b/arch/powerpc/platforms/86xx/gef_sbc610.c > > index acf2c6c3c1eb..120caf6af71d 100644 > > --- a/arch/powerpc/platforms/86xx/gef_sbc610.c > > +++ b/arch/powerpc/platforms/86xx/gef_sbc610.c > > @@ -157,7 +157,7 @@ DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NEC, PCI_DEVICE_ID_NEC_USB, > > * > > * This function is called to determine whether the BSP is compatible with the > > * supplied device-tree, which is assumed to be the correct one for the actual > > - * board. It is expected thati, in the future, a kernel may support multiple > > + * board. It is expected that, in the future, a kernel may support multiple > > * boards. > > */ > > static int __init gef_sbc610_probe(void) > > diff --git a/arch/powerpc/platforms/book3s/vas-api.c b/arch/powerpc/platforms/book3s/vas-api.c > > index f9a1615b74da..c0799fb26b6d 100644 > > --- a/arch/powerpc/platforms/book3s/vas-api.c > > +++ b/arch/powerpc/platforms/book3s/vas-api.c > > @@ -30,7 +30,7 @@ > > * > > * where "vas_copy" and "vas_paste" are defined in copy-paste.h. > > * copy/paste returns to the user space directly. So refer NX hardware > > - * documententation for exact copy/paste usage and completion / error > > + * documentation for exact copy/paste usage and completion / error > > * conditions. > > */ > > > > diff --git a/arch/powerpc/platforms/powernv/idle.c b/arch/powerpc/platforms/powernv/idle.c > > index a6677a111aca..6f94b808dd39 100644 > > --- a/arch/powerpc/platforms/powernv/idle.c > > +++ b/arch/powerpc/platforms/powernv/idle.c > > @@ -112,7 +112,7 @@ static int __init pnv_save_sprs_for_deep_states(void) > > if (rc != 0) > > return rc; > > > > - /* Only p8 needs to set extra HID regiters */ > > + /* Only p8 needs to set extra HID registers */ > > if (!cpu_has_feature(CPU_FTR_ARCH_300)) { > > uint64_t hid1_val = mfspr(SPRN_HID1); > > uint64_t hid4_val = mfspr(SPRN_HID4); > > @@ -1204,7 +1204,7 @@ static void __init pnv_arch300_idle_init(void) > > * The idle code does not deal with TB loss occurring > > * in a shallower state than SPR loss, so force it to > > * behave like SPRs are lost if TB is lost. POWER9 would > > - * never encouter this, but a POWER8 core would if it > > + * never encounter this, but a POWER8 core would if it > > * implemented the stop instruction. So this is for forward > > * compatibility. > > */ > > diff --git a/arch/powerpc/platforms/powernv/opal-fadump.c b/arch/powerpc/platforms/powernv/opal-fadump.c > > index c8ad057c7221..a8da952610eb 100644 > > --- a/arch/powerpc/platforms/powernv/opal-fadump.c > > +++ b/arch/powerpc/platforms/powernv/opal-fadump.c > > @@ -202,7 +202,7 @@ static u64 opal_fadump_init_mem_struct(struct fw_dump *fadump_conf) > > } > > > > /* > > - * Kernel metadata is passed to f/w and retrieved in capture kerenl. > > + * Kernel metadata is passed to f/w and retrieved in capture kernel. > > * So, use it to save fadump header address instead of calculating it. > > */ > > opal_fdm->fadumphdr_addr = (opal_fdm->rgn[0].dest + > > diff --git a/arch/powerpc/platforms/powernv/opal-lpc.c b/arch/powerpc/platforms/powernv/opal-lpc.c > > index 5390c888db16..d129d6d45a50 100644 > > --- a/arch/powerpc/platforms/powernv/opal-lpc.c > > +++ b/arch/powerpc/platforms/powernv/opal-lpc.c > > @@ -197,7 +197,7 @@ static ssize_t lpc_debug_read(struct file *filp, char __user *ubuf, > > > > /* > > * Select access size based on count and alignment and > > - * access type. IO and MEM only support byte acceses, > > + * access type. IO and MEM only support byte accesses, > > * FW supports all 3. > > */ > > len = 1; > > diff --git a/arch/powerpc/platforms/powernv/opal-memory-errors.c b/arch/powerpc/platforms/powernv/opal-memory-errors.c > > index 1e8e17df9ce8..a1754a28265d 100644 > > --- a/arch/powerpc/platforms/powernv/opal-memory-errors.c > > +++ b/arch/powerpc/platforms/powernv/opal-memory-errors.c > > @@ -82,7 +82,7 @@ static DECLARE_WORK(mem_error_work, mem_error_handler); > > > > /* > > * opal_memory_err_event - notifier handler that queues up the opal message > > - * to be preocessed later. > > + * to be processed later. > > */ > > static int opal_memory_err_event(struct notifier_block *nb, > > unsigned long msg_type, void *msg) > > diff --git a/arch/powerpc/platforms/powernv/pci-sriov.c b/arch/powerpc/platforms/powernv/pci-sriov.c > > index 04155aaaadb1..fe3d111b881c 100644 > > --- a/arch/powerpc/platforms/powernv/pci-sriov.c > > +++ b/arch/powerpc/platforms/powernv/pci-sriov.c > > @@ -699,7 +699,7 @@ static int pnv_pci_sriov_enable(struct pci_dev *pdev, u16 num_vfs) > > return -ENOSPC; > > } > > > > - /* allocate a contigious block of PEs for our VFs */ > > + /* allocate a contiguous block of PEs for our VFs */ > > base_pe = pnv_ioda_alloc_pe(phb, num_vfs); > > if (!base_pe) { > > pci_err(pdev, "Unable to allocate PEs for %d VFs\n", num_vfs); > > diff --git a/arch/powerpc/platforms/pseries/iommu.c b/arch/powerpc/platforms/pseries/iommu.c > > index 4d991cf840d9..7639e7355df2 100644 > > --- a/arch/powerpc/platforms/pseries/iommu.c > > +++ b/arch/powerpc/platforms/pseries/iommu.c > > @@ -1430,7 +1430,7 @@ static bool enable_ddw(struct pci_dev *dev, struct device_node *pdn) > > > > pci->table_group->tables[1] = newtbl; > > > > - /* Keep default DMA window stuct if removed */ > > + /* Keep default DMA window struct if removed */ > > if (default_win_removed) { > > tbl->it_size = 0; > > vfree(tbl->it_map); > > diff --git a/arch/powerpc/platforms/pseries/setup.c b/arch/powerpc/platforms/pseries/setup.c > > index 0f74b2284773..1a295350c403 100644 > > --- a/arch/powerpc/platforms/pseries/setup.c > > +++ b/arch/powerpc/platforms/pseries/setup.c > > @@ -658,7 +658,7 @@ static resource_size_t pseries_get_iov_fw_value(struct pci_dev *dev, int resno, > > */ > > num_res = of_read_number(&indexes[NUM_RES_PROPERTY], 1); > > if (resno >= num_res) > > - return 0; /* or an errror */ > > + return 0; /* or an error */ > > > > i = START_OF_ENTRIES + NEXT_ENTRY * resno; > > switch (value) { > > @@ -762,7 +762,7 @@ static void pseries_pci_fixup_iov_resources(struct pci_dev *pdev) > > > > if (!pdev->is_physfn) > > return; > > - /*Firmware must support open sriov otherwise dont configure*/ > > + /*Firmware must support open sriov otherwise don't configure*/ > > indexes = of_get_property(dn, "ibm,open-sriov-vf-bar-info", NULL); > > if (indexes) > > of_pci_parse_iov_addrs(pdev, indexes); > > diff --git a/arch/powerpc/platforms/pseries/vas-sysfs.c b/arch/powerpc/platforms/pseries/vas-sysfs.c > > index 909535ca513a..59aef6062847 100644 > > --- a/arch/powerpc/platforms/pseries/vas-sysfs.c > > +++ b/arch/powerpc/platforms/pseries/vas-sysfs.c > > @@ -67,7 +67,7 @@ struct vas_sysfs_entry { > > * Create sysfs interface: > > * /sys/devices/vas/vas0/gzip/default_capabilities > > * This directory contains the following VAS GZIP capabilities > > - * for the defaule credit type. > > + * for the default credit type. > > * /sys/devices/vas/vas0/gzip/default_capabilities/nr_total_credits > > * Total number of default credits assigned to the LPAR which > > * can be changed with DLPAR operation. > > diff --git a/arch/powerpc/platforms/pseries/vas.c b/arch/powerpc/platforms/pseries/vas.c > > index 1f59d78c77a1..245e0a42981a 100644 > > --- a/arch/powerpc/platforms/pseries/vas.c > > +++ b/arch/powerpc/platforms/pseries/vas.c > > @@ -807,7 +807,7 @@ int vas_reconfig_capabilties(u8 type) > > atomic_set(&caps->nr_total_credits, new_nr_creds); > > /* > > * The total number of available credits may be decreased or > > - * inceased with DLPAR operation. Means some windows have to be > > + * increased with DLPAR operation. Means some windows have to be > > * closed / reopened. Hold the vas_pseries_mutex so that the > > * the user space can not open new windows. > > */ > > diff --git a/arch/powerpc/sysdev/fsl_lbc.c b/arch/powerpc/sysdev/fsl_lbc.c > > index 1985e067e952..18acfb4e82af 100644 > > --- a/arch/powerpc/sysdev/fsl_lbc.c > > +++ b/arch/powerpc/sysdev/fsl_lbc.c > > @@ -37,7 +37,7 @@ EXPORT_SYMBOL(fsl_lbc_ctrl_dev); > > * > > * This function converts a base address of lbc into the right format for the > > * BR register. If the SOC has eLBC then it returns 32bit physical address > > - * else it convers a 34bit local bus physical address to correct format of > > + * else it converts a 34bit local bus physical address to correct format of > > * 32bit address for BR register (Example: MPC8641). > > */ > > u32 fsl_lbc_addr(phys_addr_t addr_base) > > diff --git a/arch/powerpc/sysdev/fsl_pci.c b/arch/powerpc/sysdev/fsl_pci.c > > index a97ce602394e..e5e29033929a 100644 > > --- a/arch/powerpc/sysdev/fsl_pci.c > > +++ b/arch/powerpc/sysdev/fsl_pci.c > > @@ -218,7 +218,7 @@ static void setup_pci_atmu(struct pci_controller *hose) > > * windows have implemented the default target value as 0xf > > * for CCSR space.In all Freescale legacy devices the target > > * of 0xf is reserved for local memory space. 9132 Rev1.0 > > - * now has local mempry space mapped to target 0x0 instead of > > + * now has local memory space mapped to target 0x0 instead of > > * 0xf. Hence adding a workaround to remove the target 0xf > > * defined for memory space from Inbound window attributes. > > */ > > diff --git a/arch/powerpc/sysdev/ge/ge_pic.c b/arch/powerpc/sysdev/ge/ge_pic.c > > index 02553a8ce191..413b375c4d28 100644 > > --- a/arch/powerpc/sysdev/ge/ge_pic.c > > +++ b/arch/powerpc/sysdev/ge/ge_pic.c > > @@ -150,7 +150,7 @@ static struct irq_chip gef_pic_chip = { > > }; > > > > > > -/* When an interrupt is being configured, this call allows some flexibilty > > +/* When an interrupt is being configured, this call allows some flexibility > > * in deciding which irq_chip structure is used > > */ > > static int gef_pic_host_map(struct irq_domain *h, unsigned int virq, > > diff --git a/arch/powerpc/sysdev/mpic_msgr.c b/arch/powerpc/sysdev/mpic_msgr.c > > index 36ec0bdd8b63..a25413826b63 100644 > > --- a/arch/powerpc/sysdev/mpic_msgr.c > > +++ b/arch/powerpc/sysdev/mpic_msgr.c > > @@ -99,7 +99,7 @@ void mpic_msgr_disable(struct mpic_msgr *msgr) > > EXPORT_SYMBOL_GPL(mpic_msgr_disable); > > > > /* The following three functions are used to compute the order and number of > > - * the message register blocks. They are clearly very inefficent. However, > > + * the message register blocks. They are clearly very inefficient. However, > > * they are called *only* a few times during device initialization. > > */ > > static unsigned int mpic_msgr_number_of_blocks(void) > > diff --git a/arch/powerpc/sysdev/mpic_msi.c b/arch/powerpc/sysdev/mpic_msi.c > > index f412d6ad0b66..9936c014ac7d 100644 > > --- a/arch/powerpc/sysdev/mpic_msi.c > > +++ b/arch/powerpc/sysdev/mpic_msi.c > > @@ -37,7 +37,7 @@ static int __init mpic_msi_reserve_u3_hwirqs(struct mpic *mpic) > > /* Reserve source numbers we know are reserved in the HW. > > * > > * This is a bit of a mix of U3 and U4 reserves but that's going > > - * to work fine, we have plenty enugh numbers left so let's just > > + * to work fine, we have plenty enough numbers left so let's just > > * mark anything we don't like reserved. > > */ > > for (i = 0; i < 8; i++) > > diff --git a/arch/powerpc/sysdev/mpic_timer.c b/arch/powerpc/sysdev/mpic_timer.c > > index 444e9ce42d0a..b2f0a73e8f93 100644 > > --- a/arch/powerpc/sysdev/mpic_timer.c > > +++ b/arch/powerpc/sysdev/mpic_timer.c > > @@ -255,7 +255,7 @@ EXPORT_SYMBOL(mpic_start_timer); > > > > /** > > * mpic_stop_timer - stop hardware timer > > - * @handle: the timer to be stoped > > + * @handle: the timer to be stopped > > * > > * The timer periodically generates an interrupt. Unless user stops the timer. > > */ > > diff --git a/arch/powerpc/sysdev/mpic_u3msi.c b/arch/powerpc/sysdev/mpic_u3msi.c > > index 3f4841dfefb5..73d129594078 100644 > > --- a/arch/powerpc/sysdev/mpic_u3msi.c > > +++ b/arch/powerpc/sysdev/mpic_u3msi.c > > @@ -78,7 +78,7 @@ static u64 find_u4_magic_addr(struct pci_dev *pdev, unsigned int hwirq) > > > > /* U4 PCIe MSIs need to write to the special register in > > * the bridge that generates interrupts. There should be > > - * theorically a register at 0xf8005000 where you just write > > + * theoretically a register at 0xf8005000 where you just write > > * the MSI number and that triggers the right interrupt, but > > * unfortunately, this is busted in HW, the bridge endian swaps > > * the value and hits the wrong nibble in the register. > > diff --git a/arch/powerpc/sysdev/xive/native.c b/arch/powerpc/sysdev/xive/native.c > > index f940428ad13f..45f72fc715fc 100644 > > --- a/arch/powerpc/sysdev/xive/native.c > > +++ b/arch/powerpc/sysdev/xive/native.c > > @@ -617,7 +617,7 @@ bool __init xive_native_init(void) > > > > xive_tima_os = r.start; > > > > - /* Grab size of provisionning pages */ > > + /* Grab size of provisioning pages */ > > xive_parse_provisioning(np); > > > > /* Switch the XIVE to exploitation mode */ > > diff --git a/arch/powerpc/xmon/ppc-opc.c b/arch/powerpc/xmon/ppc-opc.c > > index dfb80810b16c..0774d711453e 100644 > > --- a/arch/powerpc/xmon/ppc-opc.c > > +++ b/arch/powerpc/xmon/ppc-opc.c > > @@ -408,7 +408,7 @@ const struct powerpc_operand powerpc_operands[] = > > #define FXM4 FXM + 1 > > { 0xff, 12, insert_fxm, extract_fxm, > > PPC_OPERAND_OPTIONAL | PPC_OPERAND_OPTIONAL_VALUE}, > > - /* If the FXM4 operand is ommitted, use the sentinel value -1. */ > > + /* If the FXM4 operand is omitted, use the sentinel value -1. */ > > { -1, -1, NULL, NULL, 0}, > > > > /* The IMM20 field in an LI instruction. */ > > diff --git a/arch/powerpc/xmon/xmon.c b/arch/powerpc/xmon/xmon.c > > index b423812e94e0..b80ab82b98e5 100644 > > --- a/arch/powerpc/xmon/xmon.c > > +++ b/arch/powerpc/xmon/xmon.c > > @@ -2024,7 +2024,7 @@ static void dump_206_sprs(void) > > if (!cpu_has_feature(CPU_FTR_ARCH_206)) > > return; > > > > - /* Actually some of these pre-date 2.06, but whatevs */ > > + /* Actually some of these pre-date 2.06, but whatever */ > > > > printf("srr0 = %.16lx srr1 = %.16lx dsisr = %.8lx\n", > > mfspr(SPRN_SRR0), mfspr(SPRN_SRR1), mfspr(SPRN_DSISR)); > > diff --git a/arch/powerpc/platforms/powernv/ocxl.c b/arch/powerpc/platforms/powernv/ocxl.c > > index 28b009b46464..27c936075031 100644 > > --- a/arch/powerpc/platforms/powernv/ocxl.c > > +++ b/arch/powerpc/platforms/powernv/ocxl.c > > @@ -289,7 +289,7 @@ int pnv_ocxl_get_pasid_count(struct pci_dev *dev, int *count) > > * be used by a function depends on how many functions exist > > * on the device. The NPU needs to be configured to know how > > * many bits are available to PASIDs and how many are to be > > - * used by the function BDF indentifier. > > + * used by the function BDF identifier. > > * > > * We only support one AFU-carrying function for now. > > */ > > >