Eric Wong <e@xxxxxxxxx> writes: > For objects already in the delta_base_cache, we can safely use > one entry at-a-time directly to avoid the malloc+memcpy+free > overhead. For a 1MB delta base object, this eliminates the > speed penalty of duplicating large objects into memory and > speeds up those 1MB delta base cached content retrievals by > roughly 30%. > > While only 2-7% of objects are delta bases in repos I've looked > at, this avoids up to 96MB of duplicated memory in the worst > case with the default git config. > > The new delta_base_cache_lock is a simple single-threaded > assertion to ensure cat-file (and similar) is the exclusive user > of the delta_base_cache. In other words, we cannot have diff > or similar commands using two or more entries directly from the > delta base cache. The new lock has nothing to do with parallel > access via multiple threads at the moment. > > Signed-off-by: Eric Wong <e@xxxxxxxxx> > --- > builtin/cat-file.c | 16 +++++++++++++++- > object-file.c | 5 +++++ > object-store-ll.h | 8 ++++++++ > packfile.c | 33 ++++++++++++++++++++++++++++++--- > packfile.h | 4 ++++ > 5 files changed, 62 insertions(+), 4 deletions(-) > > diff --git a/builtin/cat-file.c b/builtin/cat-file.c > index bc4bb89610..8debcdca3e 100644 > --- a/builtin/cat-file.c > +++ b/builtin/cat-file.c > @@ -386,7 +386,20 @@ static void print_object_or_die(struct batch_options *opt, struct expand_data *d > > if (data->content) { > batch_write(opt, data->content, data->size); > - FREE_AND_NULL(data->content); > + switch (data->info.whence) { > + case OI_CACHED: > + /* > + * only blame uses OI_CACHED atm, so it's unlikely > + * we'll ever hit this path > + */ > + BUG("TODO OI_CACHED support not done"); > + case OI_LOOSE: > + case OI_PACKED: > + FREE_AND_NULL(data->content); > + break; > + case OI_DBCACHED: > + unlock_delta_base_cache(); > + } > } else if (data->type == OBJ_BLOB) { > if (opt->buffer_output) > fflush(stdout); > @@ -815,6 +828,7 @@ static int batch_objects(struct batch_options *opt) > data.info.sizep = &data.size; > data.info.contentp = &data.content; > data.info.content_limit = big_file_threshold; > + data.info.direct_cache = 1; > } > } > > diff --git a/object-file.c b/object-file.c > index 1cc29c3c58..19100e823d 100644 > --- a/object-file.c > +++ b/object-file.c > @@ -1586,6 +1586,11 @@ static int do_oid_object_info_extended(struct repository *r, > oidclr(oi->delta_base_oid, the_repository->hash_algo); > if (oi->type_name) > strbuf_addstr(oi->type_name, type_name(co->type)); > + /* > + * Currently `blame' is the only command which creates > + * OI_CACHED, and direct_cache is only used by `cat-file'. > + */ > + assert(!oi->direct_cache); If "git cat-file" ever gets enhanced to also use pretend_object_file(), this assumption gets violated. What happens then? Would we return inconsistent answer to the caller that may result in garbage output, breaking the downstream process somehow? If so, I'd prefer to see this guarded more explicitly with "if (...) BUG()" than assert() here. > if (oi->contentp) > *oi->contentp = xmemdupz(co->buf, co->size); Regardless of the answer of the question above about assert(), shouldn't the step [02/10] of this series also have been made to pay attention to oi->content_limit mechanism? It looks inconsistent. > diff --git a/object-store-ll.h b/object-store-ll.h > index b71a15f590..669bb93784 100644 > --- a/object-store-ll.h > +++ b/object-store-ll.h > @@ -298,6 +298,14 @@ struct object_info { > OI_PACKED, > OI_DBCACHED > } whence; > + > + /* > + * Set if caller is able to use OI_DBCACHED entries without copying. > + * This only applies to OI_DBCACHED entries at the moment, > + * not OI_CACHED or any other type of entry. > + */ > + unsigned direct_cache:1; This is a "response" from the API to the caller, but "Set if ..." sounds as if you are telling the caller what to do. Would the caller set this bit when making a request to say "I am cat-file and I use entry before making any other requests to the object layer so I am sure object layer will not evict the entry I am using"? No, that is not a "response". You seem to set this bit in batch_objects(), so it does sound like that the bit is expected to be set by the caller to tell the API something. If that is the case, then (1) move it to "request" part of the object_info structure, and (2) define what it means to be "able to use ... without copying". Mechanically, it may mean "contentp directly points into the delta base cache", but what implication does it have to callers? If the caller obtains such a pointer in .contentp, what is required for its access pattern to make accessing the pointer safe? The caller cannot use the pointed memory after it accesses another object? What is the definition of "access" in the context of this discussion? Does "checking for existence of an object" count as an access? > +static void lock_delta_base_cache(void) > +{ > + delta_base_cache_lock++; > + assert(delta_base_cache_lock == 1); > +} > + > +void unlock_delta_base_cache(void) > +{ > + delta_base_cache_lock--; > + assert(delta_base_cache_lock == 0); > +} I view assert() validating precondition for the control flow to pass the point in the code path, so the above looks somewhat surprising. Shouldn't we instead checking the current value first and then only if the caller is _allowed_ to lock/unlock from that state (i.e. the lock is not held/the lock is held), increment/decrement the variable? > + /* ignore content_limit if avoiding copy from cache */ > + if (oi->direct_cache) { > + lock_delta_base_cache(); > + *oi->contentp = ent->data; > + } else if (!oi->content_limit || > + ent->size <= oi->content_limit) { > *oi->contentp = xmemdupz(ent->data, ent->size); This somehow looks making the memory ownership rules confusing. How does the caller know when it can free(oi->contentp) after a call returns? If it sets direct_cache and got a non-NULL *oi->contentp back, is it always pointing at a borrowed piece memory, so that it does not have to worry about freeing, but if it didn't set direct_cache, it can be holding an allocated piece of memory and you are responsible for freeing it? The following is a tangent that is not necessary to be addressed in this series, but since I've spent time to think about it already, let me record it as #leftoverbits here. Whatever the answers to the earlier questions on the comment on the direct_cache member are, I wonder if the access pattern of the caller that satisfies such requirement allows a lot simpler solution. Would it give us similar performance boost by (morally) reverting 85fe35ab (cache_or_unpack_entry: drop keep_cache parameter, 2016-08-22) to resurrect the "pass ownership of a delta base cache entry to the caller" feature? We give the content to the caller of object_info, and evict the cached delta base. Ideally, we shouldn't have to ask the caller to do anything special, other than just setting contentp to non-NULL (and optionally setting content_limit), and the choice of copying or passing the ownership should be made in the object access layer, which knows how big the object is (e.g., small things may be easier to recreate), how deep the delta chain the entry in the delta base cache is (e.g., a delta base object that is simply deflated can be recreated cheaply, while one based on top of 100-delta deep chain is very expensive to recreate), etc.