Correct spelling typo in lustre/ldlm Signed-off-by: Masanari Iida <standby24x7@xxxxxxxxx> --- drivers/staging/lustre/lustre/ldlm/ldlm_lock.c | 8 ++++---- drivers/staging/lustre/lustre/ldlm/ldlm_pool.c | 2 +- drivers/staging/lustre/lustre/ldlm/ldlm_request.c | 10 +++++----- drivers/staging/lustre/lustre/ldlm/ldlm_resource.c | 4 ++-- 4 files changed, 12 insertions(+), 12 deletions(-) diff --git a/drivers/staging/lustre/lustre/ldlm/ldlm_lock.c b/drivers/staging/lustre/lustre/ldlm/ldlm_lock.c index 95eff79..09526e6 100644 --- a/drivers/staging/lustre/lustre/ldlm/ldlm_lock.c +++ b/drivers/staging/lustre/lustre/ldlm/ldlm_lock.c @@ -801,7 +801,7 @@ void ldlm_lock_addref_internal(struct ldlm_lock *lock, __u32 mode) * Removes reader/writer reference for LDLM lock \a lock. * Assumes LDLM lock is already locked. * only called in ldlm_flock_destroy and for local locks. - * Does NOT add lock to LRU if no r/w references left to accomodate flock locks + * Does NOT add lock to LRU if no r/w references left to accommodate flock locks * that cannot be placed in LRU. */ void ldlm_lock_decref_internal_nolock(struct ldlm_lock *lock, __u32 mode) @@ -1254,7 +1254,7 @@ EXPORT_SYMBOL(ldlm_lock_allow_match); * list will be considered * If 'flags' contains LDLM_FL_CBPENDING, then locks that have been marked * to be canceled can still be matched as long as they still have reader - * or writer refernces + * or writer references * If 'flags' contains LDLM_FL_TEST_LOCK, then don't actually reference a lock, * just tell us if we would have matched. * @@ -2097,8 +2097,8 @@ void ldlm_cancel_locks_for_export(struct obd_export *exp) /** * Downgrade an exclusive lock. * - * A fast variant of ldlm_lock_convert for convertion of exclusive - * locks. The convertion is always successful. + * A fast variant of ldlm_lock_convert for conversion of exclusive + * locks. The conversion is always successful. * Used by Commit on Sharing (COS) code. * * \param lock A lock to convert diff --git a/drivers/staging/lustre/lustre/ldlm/ldlm_pool.c b/drivers/staging/lustre/lustre/ldlm/ldlm_pool.c index 6758646..7283994 100644 --- a/drivers/staging/lustre/lustre/ldlm/ldlm_pool.c +++ b/drivers/staging/lustre/lustre/ldlm/ldlm_pool.c @@ -1322,7 +1322,7 @@ static int ldlm_pools_thread_main(void *arg) struct l_wait_info lwi; /* - * Recal all pools on this tick. + * Recalc all pools on this tick. */ s_time = ldlm_pools_recalc(LDLM_NAMESPACE_SERVER); c_time = ldlm_pools_recalc(LDLM_NAMESPACE_CLIENT); diff --git a/drivers/staging/lustre/lustre/ldlm/ldlm_request.c b/drivers/staging/lustre/lustre/ldlm/ldlm_request.c index a8f8c1c..ab73467 100644 --- a/drivers/staging/lustre/lustre/ldlm/ldlm_request.c +++ b/drivers/staging/lustre/lustre/ldlm/ldlm_request.c @@ -299,7 +299,7 @@ EXPORT_SYMBOL(ldlm_completion_ast); * A helper to build a blocking AST function * * Perform a common operation for blocking ASTs: - * defferred lock cancellation. + * deferred lock cancellation. * * \param lock the lock blocking or canceling AST was called on * \retval 0 @@ -781,7 +781,7 @@ int ldlm_prep_elc_req(struct obd_export *exp, struct ptlrpc_request *req, dlm = req_capsule_client_get(pill, &RMF_DLM_REQ); LASSERT(dlm); /* Skip first lock handler in ldlm_request_pack(), - * this method will incrment @lock_count according + * this method will increment @lock_count according * to the lock handle amount actually written to * the buffer. */ dlm->lock_count = canceloff; @@ -1329,7 +1329,7 @@ int ldlm_cli_cancel(struct lustre_handle *lockh, return 0; } /* Even if the lock is marked as LDLM_FL_BL_AST, this is a LDLM_CANCEL - * RPC which goes to canceld portal, so we can cancel other LRU locks + * RPC which goes to canceled portal, so we can cancel other LRU locks * here and send them all as one LDLM_CANCEL RPC. */ LASSERT(list_empty(&lock->l_bl_ast)); list_add(&lock->l_bl_ast, &cancels); @@ -1584,7 +1584,7 @@ ldlm_cancel_lru_policy(struct ldlm_namespace *ns, int flags) * the beginning of LRU list); * * flags & LDLM_CANCEL_SHRINK - cancel not more than \a count locks according to - * memory pressre policy function; + * memory pressure policy function; * * flags & LDLM_CANCEL_AGED - cancel \a count locks according to "aged policy". * @@ -2076,7 +2076,7 @@ static int ldlm_chain_lock_for_replay(struct ldlm_lock *lock, void *closure) lock, &lock->l_pending_chain.next,&lock->l_pending_chain.prev); /* bug 9573: don't replay locks left after eviction, or * bug 17614: locks being actively cancelled. Get a reference - * on a lock so that it does not disapear under us (e.g. due to cancel) + * on a lock so that it does not disappear under us (e.g. due to cancel) */ if (!(lock->l_flags & (LDLM_FL_FAILED|LDLM_FL_CANCELING))) { list_add(&lock->l_pending_chain, list); diff --git a/drivers/staging/lustre/lustre/ldlm/ldlm_resource.c b/drivers/staging/lustre/lustre/ldlm/ldlm_resource.c index 25e14e1..5f89864 100644 --- a/drivers/staging/lustre/lustre/ldlm/ldlm_resource.c +++ b/drivers/staging/lustre/lustre/ldlm/ldlm_resource.c @@ -162,7 +162,7 @@ static int lprocfs_ns_resources_seq_show(struct seq_file *m, void *v) struct cfs_hash_bd bd; int i; - /* result is not strictly consistant */ + /* result is not strictly consistent */ cfs_hash_for_each_bucket(ns->ns_rs_hash, &bd, i) res += cfs_hash_bd_count_get(&bd); return lprocfs_rd_u64(m, &res); @@ -874,7 +874,7 @@ void ldlm_namespace_free_prior(struct ldlm_namespace *ns, /* * With all requests dropped and the import inactive - * we are gaurenteed all reference will be dropped. + * we are guaranteed all reference will be dropped. */ rc = __ldlm_namespace_free(ns, 1); LASSERT(rc == 0); -- 1.8.5.2.192.g7794a68 _______________________________________________ devel mailing list devel@xxxxxxxxxxxxxxxxxxxxxx http://driverdev.linuxdriverproject.org/mailman/listinfo/driverdev-devel