git@xxxxxxxxxxxxxxxxx writes: > +/* > + * We use n mutexes to guard n partitions of the "istate->dir_hash" > + * hashtable. Since "find" and "insert" operations will hash to a > + * particular bucket and modify/search a single chain, we can say > + * that "all chains mod n" are guarded by the same mutex -- rather > + * than having a single mutex to guard the entire table. (This does > + * require that we disable "rehashing" on the hashtable.) > + * > + * So, a larger value here decreases the probability of a collision > + * and the time that each thread must wait for the mutex. > + */ > +#define LAZY_MAX_MUTEX (32) Good thinking is very well explained in the in-code comment. > +static int handle_range_dir( > + struct index_state *istate, > + int k_start, > + int k_end, > + struct dir_entry *parent, > + struct strbuf *prefix, > + struct lazy_entry *lazy_entries, > + struct dir_entry **dir_new_out) > +{ > + int rc, k; > + int input_prefix_len = prefix->len; > + struct dir_entry *dir_new; > + > + dir_new = hash_dir_entry_with_parent_and_prefix(istate, parent, prefix); > + > + strbuf_addch(prefix, '/'); > + > + /* > + * Scan forward in the index array for index entries having the same > + * path prefix (that are also in this directory). > + */ > + if (strncmp(istate->cache[k_start + 1]->name, prefix->buf, prefix->len) > 0) > + k = k_start + 1; > + else if (strncmp(istate->cache[k_end - 1]->name, prefix->buf, prefix->len) == 0) > + k = k_end; > + else { > + int begin = k_start; > + int end = k_end; > + while (begin < end) { > + int mid = (begin + end) >> 1; > + int cmp = strncmp(istate->cache[mid]->name, prefix->buf, prefix->len); > + if (cmp == 0) /* mid has same prefix; look in second part */ > + begin = mid + 1; > + else if (cmp > 0) /* mid is past group; look in first part */ > + end = mid; > + else > + die("cache entry out of order"); Dying at this low level in the callchain in a worker thread made me feel a little bit nervous, but even if we arrange to return to the caller without doing extra computation, we would want to detect and abort when the cache entries are not sorted anyway, so this hsould be OK. > + } > + k = begin; > + } > + > + /* > + * Recurse and process what we can of this subset [k_start, k). > + */ > + rc = handle_range_1(istate, k_start, k, dir_new, prefix, lazy_entries); > + > + strbuf_setlen(prefix, input_prefix_len); > + > + *dir_new_out = dir_new; > + return rc; > +} The varilable "rc" (return code?) is a lot more than return code; it tells how many entries we processed and signals the caller that it still needs to sweep the remainder if we didn't reach k_end. The caller of this function calls the variable to receive this "processed", so I didn't find it too confusing while reading the code top-down, though. > +static int handle_range_1( > + struct index_state *istate, > + int k_start, > + int k_end, > + struct dir_entry *parent, > + struct strbuf *prefix, > + struct lazy_entry *lazy_entries) > +{ > + int input_prefix_len = prefix->len; > + int k = k_start; > + > + while (k < k_end) { > + struct cache_entry *ce_k = istate->cache[k]; > + const char *name, *slash; > + > + if (prefix->len && strncmp(ce_k->name, prefix->buf, prefix->len)) > + break; At first I was worried by this early return (i.e. we chop the entire active_nr entries into lazy_nr_dir_threads and hand each of them a range [k_start, k_end)---stopping before a thread reaches the end of the range it is responsible for will leave gaps), but then realized that this early return "we are done with the entries in the same directory" kicks in only for recursive invocation of this function, which makes it correct and perfectly fine. I also briefly wondered if it is worth wiggling the boundary of ranges for adjacent workers to align with directory boundaries, as the last round of hashes done by a worker and the first round of hashes done by another worker adjacent to it will be hashing for the same parent directory, but I think that would be counter-productive and think your "almost even" distribution would be a lot more sensible. After all, when distribution of paths is skewed, a single directory may have disproportionally more (leaf) entries than the remainder of the index and in such a case, we would want multiple workers to share the load of hashing them, even if that means they may have to hash the same leading path independently. Nicely done. Let's have this in 'next' and then in 'master' soonish. Thanks.