From: Randy Dunlap <rdunlap@xxxxxxxxxxxxx> Fix some kernel-doc formatting in fs/eventpoll.c for "Returns:": - use "Returns:" without another "Returns" after it; - make multi-line Returns: indentation consistent; - add a ':' after "Returns" in a few places; Signed-off-by: Randy Dunlap <rdunlap@xxxxxxxxxxxxx> Cc: Alexander Viro <viro@xxxxxxxxxxxxxxxxxx> Cc: linux-fsdevel@xxxxxxxxxxxxxxx Cc: Jiri Kosina <trivial@xxxxxxxxxx> --- fs/eventpoll.c | 18 +++++++++--------- 1 file changed, 9 insertions(+), 9 deletions(-) --- linux-next-20200917.orig/fs/eventpoll.c +++ linux-next-20200917/fs/eventpoll.c @@ -371,7 +371,7 @@ static void ep_nested_calls_init(struct * * @ep: Pointer to the eventpoll context. * - * Returns: Returns a value different than zero if ready events are available, + * Returns: a value different than zero if ready events are available, * or zero otherwise. */ static inline int ep_events_available(struct eventpoll *ep) @@ -472,7 +472,7 @@ static inline void ep_set_busy_poll_napi * @cookie: Cookie to be used to identify this nested call. * @ctx: This instance context. * - * Returns: Returns the code returned by the @nproc callback, or -1 if + * Returns: the code returned by the @nproc callback, or -1 if * the maximum recursion limit has been exceeded. */ static int ep_call_nested(struct nested_calls *ncalls, @@ -1123,8 +1123,8 @@ struct file *get_epoll_tfile_raw_ptr(str * direction i.e. either to the tail either to the head, otherwise * concurrent access will corrupt the list. * - * Returns %false if element has been already added to the list, %true - * otherwise. + * Returns: %false if element has been already added to the list, %true + * otherwise. */ static inline bool list_add_tail_lockless(struct list_head *new, struct list_head *head) @@ -1165,7 +1165,7 @@ static inline bool list_add_tail_lockles * Chains a new epi entry to the tail of the ep->ovflist in a lockless way, * i.e. multiple CPUs are allowed to call this function concurrently. * - * Returns %false if epi element has been already chained, %true otherwise. + * Returns: %false if epi element has been already chained, %true otherwise. */ static inline bool chain_epi_lockless(struct epitem *epi) { @@ -1428,7 +1428,7 @@ static int reverse_path_check_proc(void * paths such that we will spend all our time waking up * eventpoll objects. * - * Returns: Returns zero if the proposed links don't create too many paths, + * Returns: zero if the proposed links don't create too many paths, * -1 otherwise. */ static int reverse_path_check(void) @@ -1814,7 +1814,7 @@ static inline struct timespec64 ep_set_m * until at least one event has been retrieved (or an error * occurred). * - * Returns: Returns the number of ready events which have been fetched, or an + * Returns: the number of ready events which have been fetched, or an * error code, in case of error. */ static int ep_poll(struct eventpoll *ep, struct epoll_event __user *events, @@ -1959,7 +1959,7 @@ send_events: * data structure pointer. * @call_nests: Current dept of the @ep_call_nested() call stack. * - * Returns: Returns zero if adding the epoll @file inside current epoll + * Returns: zero if adding the epoll @file inside current epoll * structure @ep does not violate the constraints, or -1 otherwise. */ static int ep_loop_check_proc(void *priv, void *cookie, int call_nests) @@ -2014,7 +2014,7 @@ static int ep_loop_check_proc(void *priv * @ep: Pointer to the epoll private data structure. * @file: Pointer to the epoll file to be checked. * - * Returns: Returns zero if adding the epoll @file inside current epoll + * Returns: zero if adding the epoll @file inside current epoll * structure @ep does not violate the constraints, or -1 otherwise. */ static int ep_loop_check(struct eventpoll *ep, struct file *file)