mirror of
https://github.com/git/git
synced 2024-11-05 01:58:18 +00:00
61133e6ebb
When unregistering a shallow root we shrink the array of grafts by one and move remaining grafts one to the left. This can of course only happen when there are any grafts left, because otherwise there is nothing to move. As such, this code is guarded by a condition that only performs the move in case there are grafts after the position of the graft to be unregistered. By mistake we also put the call to free the unregistered graft into that condition. But that doesn't make any sense, as we want to always free the graft when it exists. Fix the resulting memory leak by doing so. This leak is exposed by t5500, but plugging it does not make the whole test suite pass. Signed-off-by: Patrick Steinhardt <ps@pks.im> Signed-off-by: Jeff King <peff@peff.net> Signed-off-by: Junio C Hamano <gitster@pobox.com>
859 lines
23 KiB
C
859 lines
23 KiB
C
#define USE_THE_REPOSITORY_VARIABLE
|
|
|
|
#include "git-compat-util.h"
|
|
#include "hex.h"
|
|
#include "repository.h"
|
|
#include "tempfile.h"
|
|
#include "lockfile.h"
|
|
#include "object-store-ll.h"
|
|
#include "commit.h"
|
|
#include "tag.h"
|
|
#include "pkt-line.h"
|
|
#include "refs.h"
|
|
#include "oid-array.h"
|
|
#include "path.h"
|
|
#include "diff.h"
|
|
#include "revision.h"
|
|
#include "commit-slab.h"
|
|
#include "list-objects.h"
|
|
#include "commit-reach.h"
|
|
#include "shallow.h"
|
|
#include "statinfo.h"
|
|
#include "trace.h"
|
|
|
|
void set_alternate_shallow_file(struct repository *r, const char *path, int override)
|
|
{
|
|
if (r->parsed_objects->is_shallow != -1)
|
|
BUG("is_repository_shallow must not be called before set_alternate_shallow_file");
|
|
if (r->parsed_objects->alternate_shallow_file && !override)
|
|
return;
|
|
free(r->parsed_objects->alternate_shallow_file);
|
|
r->parsed_objects->alternate_shallow_file = xstrdup_or_null(path);
|
|
}
|
|
|
|
int register_shallow(struct repository *r, const struct object_id *oid)
|
|
{
|
|
struct commit_graft *graft =
|
|
xmalloc(sizeof(struct commit_graft));
|
|
struct commit *commit = lookup_commit(r, oid);
|
|
|
|
oidcpy(&graft->oid, oid);
|
|
graft->nr_parent = -1;
|
|
if (commit && commit->object.parsed) {
|
|
free_commit_list(commit->parents);
|
|
commit->parents = NULL;
|
|
}
|
|
return register_commit_graft(r, graft, 0);
|
|
}
|
|
|
|
int unregister_shallow(const struct object_id *oid)
|
|
{
|
|
int pos = commit_graft_pos(the_repository, oid);
|
|
if (pos < 0)
|
|
return -1;
|
|
free(the_repository->parsed_objects->grafts[pos]);
|
|
if (pos + 1 < the_repository->parsed_objects->grafts_nr)
|
|
MOVE_ARRAY(the_repository->parsed_objects->grafts + pos,
|
|
the_repository->parsed_objects->grafts + pos + 1,
|
|
the_repository->parsed_objects->grafts_nr - pos - 1);
|
|
the_repository->parsed_objects->grafts_nr--;
|
|
return 0;
|
|
}
|
|
|
|
int is_repository_shallow(struct repository *r)
|
|
{
|
|
FILE *fp;
|
|
char buf[1024];
|
|
const char *path = r->parsed_objects->alternate_shallow_file;
|
|
|
|
if (r->parsed_objects->is_shallow >= 0)
|
|
return r->parsed_objects->is_shallow;
|
|
|
|
if (!path)
|
|
path = git_path_shallow(r);
|
|
/*
|
|
* fetch-pack sets '--shallow-file ""' as an indicator that no
|
|
* shallow file should be used. We could just open it and it
|
|
* will likely fail. But let's do an explicit check instead.
|
|
*/
|
|
if (!*path || (fp = fopen(path, "r")) == NULL) {
|
|
stat_validity_clear(r->parsed_objects->shallow_stat);
|
|
r->parsed_objects->is_shallow = 0;
|
|
return r->parsed_objects->is_shallow;
|
|
}
|
|
stat_validity_update(r->parsed_objects->shallow_stat, fileno(fp));
|
|
r->parsed_objects->is_shallow = 1;
|
|
|
|
while (fgets(buf, sizeof(buf), fp)) {
|
|
struct object_id oid;
|
|
if (get_oid_hex(buf, &oid))
|
|
die("bad shallow line: %s", buf);
|
|
register_shallow(r, &oid);
|
|
}
|
|
fclose(fp);
|
|
return r->parsed_objects->is_shallow;
|
|
}
|
|
|
|
static void reset_repository_shallow(struct repository *r)
|
|
{
|
|
r->parsed_objects->is_shallow = -1;
|
|
stat_validity_clear(r->parsed_objects->shallow_stat);
|
|
parsed_object_pool_reset_commit_grafts(r->parsed_objects);
|
|
}
|
|
|
|
int commit_shallow_file(struct repository *r, struct shallow_lock *lk)
|
|
{
|
|
int res = commit_lock_file(&lk->lock);
|
|
reset_repository_shallow(r);
|
|
|
|
/*
|
|
* Update in-memory data structures with the new shallow information,
|
|
* including unparsing all commits that now have grafts.
|
|
*/
|
|
is_repository_shallow(r);
|
|
|
|
return res;
|
|
}
|
|
|
|
void rollback_shallow_file(struct repository *r, struct shallow_lock *lk)
|
|
{
|
|
rollback_lock_file(&lk->lock);
|
|
reset_repository_shallow(r);
|
|
}
|
|
|
|
/*
|
|
* TODO: use "int" elemtype instead of "int *" when/if commit-slab
|
|
* supports a "valid" flag.
|
|
*/
|
|
define_commit_slab(commit_depth, int *);
|
|
static void free_depth_in_slab(int **ptr)
|
|
{
|
|
FREE_AND_NULL(*ptr);
|
|
}
|
|
struct commit_list *get_shallow_commits(struct object_array *heads, int depth,
|
|
int shallow_flag, int not_shallow_flag)
|
|
{
|
|
int i = 0, cur_depth = 0;
|
|
struct commit_list *result = NULL;
|
|
struct object_array stack = OBJECT_ARRAY_INIT;
|
|
struct commit *commit = NULL;
|
|
struct commit_graft *graft;
|
|
struct commit_depth depths;
|
|
|
|
init_commit_depth(&depths);
|
|
while (commit || i < heads->nr || stack.nr) {
|
|
struct commit_list *p;
|
|
if (!commit) {
|
|
if (i < heads->nr) {
|
|
int **depth_slot;
|
|
commit = (struct commit *)
|
|
deref_tag(the_repository,
|
|
heads->objects[i++].item,
|
|
NULL, 0);
|
|
if (!commit || commit->object.type != OBJ_COMMIT) {
|
|
commit = NULL;
|
|
continue;
|
|
}
|
|
depth_slot = commit_depth_at(&depths, commit);
|
|
if (!*depth_slot)
|
|
*depth_slot = xmalloc(sizeof(int));
|
|
**depth_slot = 0;
|
|
cur_depth = 0;
|
|
} else {
|
|
commit = (struct commit *)
|
|
object_array_pop(&stack);
|
|
cur_depth = **commit_depth_at(&depths, commit);
|
|
}
|
|
}
|
|
parse_commit_or_die(commit);
|
|
cur_depth++;
|
|
if ((depth != INFINITE_DEPTH && cur_depth >= depth) ||
|
|
(is_repository_shallow(the_repository) && !commit->parents &&
|
|
(graft = lookup_commit_graft(the_repository, &commit->object.oid)) != NULL &&
|
|
graft->nr_parent < 0)) {
|
|
commit_list_insert(commit, &result);
|
|
commit->object.flags |= shallow_flag;
|
|
commit = NULL;
|
|
continue;
|
|
}
|
|
commit->object.flags |= not_shallow_flag;
|
|
for (p = commit->parents, commit = NULL; p; p = p->next) {
|
|
int **depth_slot = commit_depth_at(&depths, p->item);
|
|
if (!*depth_slot) {
|
|
*depth_slot = xmalloc(sizeof(int));
|
|
**depth_slot = cur_depth;
|
|
} else {
|
|
if (cur_depth >= **depth_slot)
|
|
continue;
|
|
**depth_slot = cur_depth;
|
|
}
|
|
if (p->next)
|
|
add_object_array(&p->item->object,
|
|
NULL, &stack);
|
|
else {
|
|
commit = p->item;
|
|
cur_depth = **commit_depth_at(&depths, commit);
|
|
}
|
|
}
|
|
}
|
|
deep_clear_commit_depth(&depths, free_depth_in_slab);
|
|
|
|
return result;
|
|
}
|
|
|
|
static void show_commit(struct commit *commit, void *data)
|
|
{
|
|
commit_list_insert(commit, data);
|
|
}
|
|
|
|
/*
|
|
* Given rev-list arguments, run rev-list. All reachable commits
|
|
* except border ones are marked with not_shallow_flag. Border commits
|
|
* are marked with shallow_flag. The list of border/shallow commits
|
|
* are also returned.
|
|
*/
|
|
struct commit_list *get_shallow_commits_by_rev_list(int ac, const char **av,
|
|
int shallow_flag,
|
|
int not_shallow_flag)
|
|
{
|
|
struct commit_list *result = NULL, *p;
|
|
struct commit_list *not_shallow_list = NULL;
|
|
struct rev_info revs;
|
|
int both_flags = shallow_flag | not_shallow_flag;
|
|
|
|
/*
|
|
* SHALLOW (excluded) and NOT_SHALLOW (included) should not be
|
|
* set at this point. But better be safe than sorry.
|
|
*/
|
|
clear_object_flags(both_flags);
|
|
|
|
is_repository_shallow(the_repository); /* make sure shallows are read */
|
|
|
|
repo_init_revisions(the_repository, &revs, NULL);
|
|
save_commit_buffer = 0;
|
|
setup_revisions(ac, av, &revs, NULL);
|
|
|
|
if (prepare_revision_walk(&revs))
|
|
die("revision walk setup failed");
|
|
traverse_commit_list(&revs, show_commit, NULL, ¬_shallow_list);
|
|
|
|
if (!not_shallow_list)
|
|
die("no commits selected for shallow requests");
|
|
|
|
/* Mark all reachable commits as NOT_SHALLOW */
|
|
for (p = not_shallow_list; p; p = p->next)
|
|
p->item->object.flags |= not_shallow_flag;
|
|
|
|
/*
|
|
* mark border commits SHALLOW + NOT_SHALLOW.
|
|
* We cannot clear NOT_SHALLOW right now. Imagine border
|
|
* commit A is processed first, then commit B, whose parent is
|
|
* A, later. If NOT_SHALLOW on A is cleared at step 1, B
|
|
* itself is considered border at step 2, which is incorrect.
|
|
*/
|
|
for (p = not_shallow_list; p; p = p->next) {
|
|
struct commit *c = p->item;
|
|
struct commit_list *parent;
|
|
|
|
if (repo_parse_commit(the_repository, c))
|
|
die("unable to parse commit %s",
|
|
oid_to_hex(&c->object.oid));
|
|
|
|
for (parent = c->parents; parent; parent = parent->next)
|
|
if (!(parent->item->object.flags & not_shallow_flag)) {
|
|
c->object.flags |= shallow_flag;
|
|
commit_list_insert(c, &result);
|
|
break;
|
|
}
|
|
}
|
|
free_commit_list(not_shallow_list);
|
|
|
|
/*
|
|
* Now we can clean up NOT_SHALLOW on border commits. Having
|
|
* both flags set can confuse the caller.
|
|
*/
|
|
for (p = result; p; p = p->next) {
|
|
struct object *o = &p->item->object;
|
|
if ((o->flags & both_flags) == both_flags)
|
|
o->flags &= ~not_shallow_flag;
|
|
}
|
|
release_revisions(&revs);
|
|
return result;
|
|
}
|
|
|
|
static void check_shallow_file_for_update(struct repository *r)
|
|
{
|
|
if (r->parsed_objects->is_shallow == -1)
|
|
BUG("shallow must be initialized by now");
|
|
|
|
if (!stat_validity_check(r->parsed_objects->shallow_stat,
|
|
git_path_shallow(r)))
|
|
die("shallow file has changed since we read it");
|
|
}
|
|
|
|
#define SEEN_ONLY 1
|
|
#define VERBOSE 2
|
|
#define QUICK 4
|
|
|
|
struct write_shallow_data {
|
|
struct strbuf *out;
|
|
int use_pack_protocol;
|
|
int count;
|
|
unsigned flags;
|
|
};
|
|
|
|
static int write_one_shallow(const struct commit_graft *graft, void *cb_data)
|
|
{
|
|
struct write_shallow_data *data = cb_data;
|
|
const char *hex = oid_to_hex(&graft->oid);
|
|
if (graft->nr_parent != -1)
|
|
return 0;
|
|
if (data->flags & QUICK) {
|
|
if (!repo_has_object_file(the_repository, &graft->oid))
|
|
return 0;
|
|
} else if (data->flags & SEEN_ONLY) {
|
|
struct commit *c = lookup_commit(the_repository, &graft->oid);
|
|
if (!c || !(c->object.flags & SEEN)) {
|
|
if (data->flags & VERBOSE)
|
|
printf("Removing %s from .git/shallow\n",
|
|
oid_to_hex(&c->object.oid));
|
|
return 0;
|
|
}
|
|
}
|
|
data->count++;
|
|
if (data->use_pack_protocol)
|
|
packet_buf_write(data->out, "shallow %s", hex);
|
|
else {
|
|
strbuf_addstr(data->out, hex);
|
|
strbuf_addch(data->out, '\n');
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int write_shallow_commits_1(struct strbuf *out, int use_pack_protocol,
|
|
const struct oid_array *extra,
|
|
unsigned flags)
|
|
{
|
|
struct write_shallow_data data;
|
|
int i;
|
|
data.out = out;
|
|
data.use_pack_protocol = use_pack_protocol;
|
|
data.count = 0;
|
|
data.flags = flags;
|
|
for_each_commit_graft(write_one_shallow, &data);
|
|
if (!extra)
|
|
return data.count;
|
|
for (i = 0; i < extra->nr; i++) {
|
|
strbuf_addstr(out, oid_to_hex(extra->oid + i));
|
|
strbuf_addch(out, '\n');
|
|
data.count++;
|
|
}
|
|
return data.count;
|
|
}
|
|
|
|
int write_shallow_commits(struct strbuf *out, int use_pack_protocol,
|
|
const struct oid_array *extra)
|
|
{
|
|
return write_shallow_commits_1(out, use_pack_protocol, extra, 0);
|
|
}
|
|
|
|
const char *setup_temporary_shallow(const struct oid_array *extra)
|
|
{
|
|
struct tempfile *temp;
|
|
struct strbuf sb = STRBUF_INIT;
|
|
|
|
if (write_shallow_commits(&sb, 0, extra)) {
|
|
temp = xmks_tempfile(git_path("shallow_XXXXXX"));
|
|
|
|
if (write_in_full(temp->fd, sb.buf, sb.len) < 0 ||
|
|
close_tempfile_gently(temp) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_tempfile_path(temp));
|
|
strbuf_release(&sb);
|
|
return get_tempfile_path(temp);
|
|
}
|
|
/*
|
|
* is_repository_shallow() sees empty string as "no shallow
|
|
* file".
|
|
*/
|
|
return "";
|
|
}
|
|
|
|
void setup_alternate_shallow(struct shallow_lock *shallow_lock,
|
|
const char **alternate_shallow_file,
|
|
const struct oid_array *extra)
|
|
{
|
|
struct strbuf sb = STRBUF_INIT;
|
|
int fd;
|
|
|
|
fd = hold_lock_file_for_update(&shallow_lock->lock,
|
|
git_path_shallow(the_repository),
|
|
LOCK_DIE_ON_ERROR);
|
|
check_shallow_file_for_update(the_repository);
|
|
if (write_shallow_commits(&sb, 0, extra)) {
|
|
if (write_in_full(fd, sb.buf, sb.len) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_lock_file_path(&shallow_lock->lock));
|
|
*alternate_shallow_file = get_lock_file_path(&shallow_lock->lock);
|
|
} else
|
|
/*
|
|
* is_repository_shallow() sees empty string as "no
|
|
* shallow file".
|
|
*/
|
|
*alternate_shallow_file = "";
|
|
strbuf_release(&sb);
|
|
}
|
|
|
|
static int advertise_shallow_grafts_cb(const struct commit_graft *graft, void *cb)
|
|
{
|
|
int fd = *(int *)cb;
|
|
if (graft->nr_parent == -1)
|
|
packet_write_fmt(fd, "shallow %s\n", oid_to_hex(&graft->oid));
|
|
return 0;
|
|
}
|
|
|
|
void advertise_shallow_grafts(int fd)
|
|
{
|
|
if (!is_repository_shallow(the_repository))
|
|
return;
|
|
for_each_commit_graft(advertise_shallow_grafts_cb, &fd);
|
|
}
|
|
|
|
/*
|
|
* mark_reachable_objects() should have been run prior to this and all
|
|
* reachable commits marked as "SEEN", except when quick_prune is non-zero,
|
|
* in which case lines are excised from the shallow file if they refer to
|
|
* commits that do not exist (any longer).
|
|
*/
|
|
void prune_shallow(unsigned options)
|
|
{
|
|
struct shallow_lock shallow_lock = SHALLOW_LOCK_INIT;
|
|
struct strbuf sb = STRBUF_INIT;
|
|
unsigned flags = SEEN_ONLY;
|
|
int fd;
|
|
|
|
if (options & PRUNE_QUICK)
|
|
flags |= QUICK;
|
|
|
|
if (options & PRUNE_SHOW_ONLY) {
|
|
flags |= VERBOSE;
|
|
write_shallow_commits_1(&sb, 0, NULL, flags);
|
|
strbuf_release(&sb);
|
|
return;
|
|
}
|
|
fd = hold_lock_file_for_update(&shallow_lock.lock,
|
|
git_path_shallow(the_repository),
|
|
LOCK_DIE_ON_ERROR);
|
|
check_shallow_file_for_update(the_repository);
|
|
if (write_shallow_commits_1(&sb, 0, NULL, flags)) {
|
|
if (write_in_full(fd, sb.buf, sb.len) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_lock_file_path(&shallow_lock.lock));
|
|
commit_shallow_file(the_repository, &shallow_lock);
|
|
} else {
|
|
unlink(git_path_shallow(the_repository));
|
|
rollback_shallow_file(the_repository, &shallow_lock);
|
|
}
|
|
strbuf_release(&sb);
|
|
}
|
|
|
|
struct trace_key trace_shallow = TRACE_KEY_INIT(SHALLOW);
|
|
|
|
/*
|
|
* Step 1, split sender shallow commits into "ours" and "theirs"
|
|
* Step 2, clean "ours" based on .git/shallow
|
|
*/
|
|
void prepare_shallow_info(struct shallow_info *info, struct oid_array *sa)
|
|
{
|
|
int i;
|
|
trace_printf_key(&trace_shallow, "shallow: prepare_shallow_info\n");
|
|
memset(info, 0, sizeof(*info));
|
|
info->shallow = sa;
|
|
if (!sa)
|
|
return;
|
|
ALLOC_ARRAY(info->ours, sa->nr);
|
|
ALLOC_ARRAY(info->theirs, sa->nr);
|
|
for (i = 0; i < sa->nr; i++) {
|
|
if (repo_has_object_file(the_repository, sa->oid + i)) {
|
|
struct commit_graft *graft;
|
|
graft = lookup_commit_graft(the_repository,
|
|
&sa->oid[i]);
|
|
if (graft && graft->nr_parent < 0)
|
|
continue;
|
|
info->ours[info->nr_ours++] = i;
|
|
} else
|
|
info->theirs[info->nr_theirs++] = i;
|
|
}
|
|
}
|
|
|
|
void clear_shallow_info(struct shallow_info *info)
|
|
{
|
|
if (info->used_shallow) {
|
|
for (size_t i = 0; i < info->shallow->nr; i++)
|
|
free(info->used_shallow[i]);
|
|
free(info->used_shallow);
|
|
}
|
|
|
|
free(info->need_reachability_test);
|
|
free(info->reachable);
|
|
free(info->shallow_ref);
|
|
free(info->ours);
|
|
free(info->theirs);
|
|
}
|
|
|
|
/* Step 4, remove non-existent ones in "theirs" after getting the pack */
|
|
|
|
void remove_nonexistent_theirs_shallow(struct shallow_info *info)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
int i, dst;
|
|
trace_printf_key(&trace_shallow, "shallow: remove_nonexistent_theirs_shallow\n");
|
|
for (i = dst = 0; i < info->nr_theirs; i++) {
|
|
if (i != dst)
|
|
info->theirs[dst] = info->theirs[i];
|
|
if (repo_has_object_file(the_repository, oid + info->theirs[i]))
|
|
dst++;
|
|
}
|
|
info->nr_theirs = dst;
|
|
}
|
|
|
|
define_commit_slab(ref_bitmap, uint32_t *);
|
|
|
|
#define POOL_SIZE (512 * 1024)
|
|
|
|
struct paint_info {
|
|
struct ref_bitmap ref_bitmap;
|
|
unsigned nr_bits;
|
|
char **pools;
|
|
char *free, *end;
|
|
unsigned pool_count;
|
|
};
|
|
|
|
static uint32_t *paint_alloc(struct paint_info *info)
|
|
{
|
|
unsigned nr = DIV_ROUND_UP(info->nr_bits, 32);
|
|
unsigned size = nr * sizeof(uint32_t);
|
|
void *p;
|
|
if (!info->pool_count || size > info->end - info->free) {
|
|
if (size > POOL_SIZE)
|
|
BUG("pool size too small for %d in paint_alloc()",
|
|
size);
|
|
info->pool_count++;
|
|
REALLOC_ARRAY(info->pools, info->pool_count);
|
|
info->free = xmalloc(POOL_SIZE);
|
|
info->pools[info->pool_count - 1] = info->free;
|
|
info->end = info->free + POOL_SIZE;
|
|
}
|
|
p = info->free;
|
|
info->free += size;
|
|
return p;
|
|
}
|
|
|
|
/*
|
|
* Given a commit SHA-1, walk down to parents until either SEEN,
|
|
* UNINTERESTING or BOTTOM is hit. Set the id-th bit in ref_bitmap for
|
|
* all walked commits.
|
|
*/
|
|
static void paint_down(struct paint_info *info, const struct object_id *oid,
|
|
unsigned int id)
|
|
{
|
|
unsigned int i, nr;
|
|
struct commit_list *head = NULL;
|
|
int bitmap_nr = DIV_ROUND_UP(info->nr_bits, 32);
|
|
size_t bitmap_size = st_mult(sizeof(uint32_t), bitmap_nr);
|
|
struct commit *c = lookup_commit_reference_gently(the_repository, oid,
|
|
1);
|
|
uint32_t *tmp; /* to be freed before return */
|
|
uint32_t *bitmap;
|
|
|
|
if (!c)
|
|
return;
|
|
|
|
tmp = xmalloc(bitmap_size);
|
|
bitmap = paint_alloc(info);
|
|
memset(bitmap, 0, bitmap_size);
|
|
bitmap[id / 32] |= (1U << (id % 32));
|
|
commit_list_insert(c, &head);
|
|
while (head) {
|
|
struct commit_list *p;
|
|
struct commit *c = pop_commit(&head);
|
|
uint32_t **refs = ref_bitmap_at(&info->ref_bitmap, c);
|
|
|
|
/* XXX check "UNINTERESTING" from pack bitmaps if available */
|
|
if (c->object.flags & (SEEN | UNINTERESTING))
|
|
continue;
|
|
else
|
|
c->object.flags |= SEEN;
|
|
|
|
if (!*refs)
|
|
*refs = bitmap;
|
|
else {
|
|
memcpy(tmp, *refs, bitmap_size);
|
|
for (i = 0; i < bitmap_nr; i++)
|
|
tmp[i] |= bitmap[i];
|
|
if (memcmp(tmp, *refs, bitmap_size)) {
|
|
*refs = paint_alloc(info);
|
|
memcpy(*refs, tmp, bitmap_size);
|
|
}
|
|
}
|
|
|
|
if (c->object.flags & BOTTOM)
|
|
continue;
|
|
|
|
if (repo_parse_commit(the_repository, c))
|
|
die("unable to parse commit %s",
|
|
oid_to_hex(&c->object.oid));
|
|
|
|
for (p = c->parents; p; p = p->next) {
|
|
if (p->item->object.flags & SEEN)
|
|
continue;
|
|
commit_list_insert(p->item, &head);
|
|
}
|
|
}
|
|
|
|
nr = get_max_object_index();
|
|
for (i = 0; i < nr; i++) {
|
|
struct object *o = get_indexed_object(i);
|
|
if (o && o->type == OBJ_COMMIT)
|
|
o->flags &= ~SEEN;
|
|
}
|
|
|
|
free(tmp);
|
|
}
|
|
|
|
static int mark_uninteresting(const char *refname UNUSED,
|
|
const char *referent UNUSED,
|
|
const struct object_id *oid,
|
|
int flags UNUSED,
|
|
void *cb_data UNUSED)
|
|
{
|
|
struct commit *commit = lookup_commit_reference_gently(the_repository,
|
|
oid, 1);
|
|
if (!commit)
|
|
return 0;
|
|
commit->object.flags |= UNINTERESTING;
|
|
mark_parents_uninteresting(NULL, commit);
|
|
return 0;
|
|
}
|
|
|
|
static void post_assign_shallow(struct shallow_info *info,
|
|
struct ref_bitmap *ref_bitmap,
|
|
int *ref_status);
|
|
/*
|
|
* Step 6(+7), associate shallow commits with new refs
|
|
*
|
|
* info->ref must be initialized before calling this function.
|
|
*
|
|
* If used is not NULL, it's an array of info->shallow->nr
|
|
* bitmaps. The n-th bit set in the m-th bitmap if ref[n] needs the
|
|
* m-th shallow commit from info->shallow.
|
|
*
|
|
* If used is NULL, "ours" and "theirs" are updated. And if ref_status
|
|
* is not NULL it's an array of ref->nr ints. ref_status[i] is true if
|
|
* the ref needs some shallow commits from either info->ours or
|
|
* info->theirs.
|
|
*/
|
|
void assign_shallow_commits_to_refs(struct shallow_info *info,
|
|
uint32_t **used, int *ref_status)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
struct oid_array *ref = info->ref;
|
|
unsigned int i, nr;
|
|
int *shallow, nr_shallow = 0;
|
|
struct paint_info pi;
|
|
|
|
trace_printf_key(&trace_shallow, "shallow: assign_shallow_commits_to_refs\n");
|
|
ALLOC_ARRAY(shallow, info->nr_ours + info->nr_theirs);
|
|
for (i = 0; i < info->nr_ours; i++)
|
|
shallow[nr_shallow++] = info->ours[i];
|
|
for (i = 0; i < info->nr_theirs; i++)
|
|
shallow[nr_shallow++] = info->theirs[i];
|
|
|
|
/*
|
|
* Prepare the commit graph to track what refs can reach what
|
|
* (new) shallow commits.
|
|
*/
|
|
nr = get_max_object_index();
|
|
for (i = 0; i < nr; i++) {
|
|
struct object *o = get_indexed_object(i);
|
|
if (!o || o->type != OBJ_COMMIT)
|
|
continue;
|
|
|
|
o->flags &= ~(UNINTERESTING | BOTTOM | SEEN);
|
|
}
|
|
|
|
memset(&pi, 0, sizeof(pi));
|
|
init_ref_bitmap(&pi.ref_bitmap);
|
|
pi.nr_bits = ref->nr;
|
|
|
|
/*
|
|
* "--not --all" to cut short the traversal if new refs
|
|
* connect to old refs. If not (e.g. force ref updates) it'll
|
|
* have to go down to the current shallow commits.
|
|
*/
|
|
refs_head_ref(get_main_ref_store(the_repository), mark_uninteresting,
|
|
NULL);
|
|
refs_for_each_ref(get_main_ref_store(the_repository),
|
|
mark_uninteresting, NULL);
|
|
|
|
/* Mark potential bottoms so we won't go out of bound */
|
|
for (i = 0; i < nr_shallow; i++) {
|
|
struct commit *c = lookup_commit(the_repository,
|
|
&oid[shallow[i]]);
|
|
c->object.flags |= BOTTOM;
|
|
}
|
|
|
|
for (i = 0; i < ref->nr; i++)
|
|
paint_down(&pi, ref->oid + i, i);
|
|
|
|
if (used) {
|
|
int bitmap_size = DIV_ROUND_UP(pi.nr_bits, 32) * sizeof(uint32_t);
|
|
memset(used, 0, sizeof(*used) * info->shallow->nr);
|
|
for (i = 0; i < nr_shallow; i++) {
|
|
const struct commit *c = lookup_commit(the_repository,
|
|
&oid[shallow[i]]);
|
|
uint32_t **map = ref_bitmap_at(&pi.ref_bitmap, c);
|
|
if (*map)
|
|
used[shallow[i]] = xmemdupz(*map, bitmap_size);
|
|
}
|
|
/*
|
|
* unreachable shallow commits are not removed from
|
|
* "ours" and "theirs". The user is supposed to run
|
|
* step 7 on every ref separately and not trust "ours"
|
|
* and "theirs" any more.
|
|
*/
|
|
} else
|
|
post_assign_shallow(info, &pi.ref_bitmap, ref_status);
|
|
|
|
clear_ref_bitmap(&pi.ref_bitmap);
|
|
for (i = 0; i < pi.pool_count; i++)
|
|
free(pi.pools[i]);
|
|
free(pi.pools);
|
|
free(shallow);
|
|
}
|
|
|
|
struct commit_array {
|
|
struct commit **commits;
|
|
int nr, alloc;
|
|
};
|
|
|
|
static int add_ref(const char *refname UNUSED,
|
|
const char *referent UNUSED,
|
|
const struct object_id *oid,
|
|
int flags UNUSED,
|
|
void *cb_data)
|
|
{
|
|
struct commit_array *ca = cb_data;
|
|
ALLOC_GROW(ca->commits, ca->nr + 1, ca->alloc);
|
|
ca->commits[ca->nr] = lookup_commit_reference_gently(the_repository,
|
|
oid, 1);
|
|
if (ca->commits[ca->nr])
|
|
ca->nr++;
|
|
return 0;
|
|
}
|
|
|
|
static void update_refstatus(int *ref_status, int nr, uint32_t *bitmap)
|
|
{
|
|
unsigned int i;
|
|
if (!ref_status)
|
|
return;
|
|
for (i = 0; i < nr; i++)
|
|
if (bitmap[i / 32] & (1U << (i % 32)))
|
|
ref_status[i]++;
|
|
}
|
|
|
|
/*
|
|
* Step 7, reachability test on "ours" at commit level
|
|
*/
|
|
static void post_assign_shallow(struct shallow_info *info,
|
|
struct ref_bitmap *ref_bitmap,
|
|
int *ref_status)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
struct commit *c;
|
|
uint32_t **bitmap;
|
|
int dst, i, j;
|
|
int bitmap_nr = DIV_ROUND_UP(info->ref->nr, 32);
|
|
struct commit_array ca;
|
|
|
|
trace_printf_key(&trace_shallow, "shallow: post_assign_shallow\n");
|
|
if (ref_status)
|
|
memset(ref_status, 0, sizeof(*ref_status) * info->ref->nr);
|
|
|
|
/* Remove unreachable shallow commits from "theirs" */
|
|
for (i = dst = 0; i < info->nr_theirs; i++) {
|
|
if (i != dst)
|
|
info->theirs[dst] = info->theirs[i];
|
|
c = lookup_commit(the_repository, &oid[info->theirs[i]]);
|
|
bitmap = ref_bitmap_at(ref_bitmap, c);
|
|
if (!*bitmap)
|
|
continue;
|
|
for (j = 0; j < bitmap_nr; j++)
|
|
if (bitmap[0][j]) {
|
|
update_refstatus(ref_status, info->ref->nr, *bitmap);
|
|
dst++;
|
|
break;
|
|
}
|
|
}
|
|
info->nr_theirs = dst;
|
|
|
|
memset(&ca, 0, sizeof(ca));
|
|
refs_head_ref(get_main_ref_store(the_repository), add_ref, &ca);
|
|
refs_for_each_ref(get_main_ref_store(the_repository), add_ref, &ca);
|
|
|
|
/* Remove unreachable shallow commits from "ours" */
|
|
for (i = dst = 0; i < info->nr_ours; i++) {
|
|
if (i != dst)
|
|
info->ours[dst] = info->ours[i];
|
|
c = lookup_commit(the_repository, &oid[info->ours[i]]);
|
|
bitmap = ref_bitmap_at(ref_bitmap, c);
|
|
if (!*bitmap)
|
|
continue;
|
|
for (j = 0; j < bitmap_nr; j++)
|
|
if (bitmap[0][j]) {
|
|
/* Step 7, reachability test at commit level */
|
|
int ret = repo_in_merge_bases_many(the_repository, c, ca.nr, ca.commits, 1);
|
|
if (ret < 0)
|
|
exit(128);
|
|
if (!ret) {
|
|
update_refstatus(ref_status, info->ref->nr, *bitmap);
|
|
dst++;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
info->nr_ours = dst;
|
|
|
|
free(ca.commits);
|
|
}
|
|
|
|
/* (Delayed) step 7, reachability test at commit level */
|
|
int delayed_reachability_test(struct shallow_info *si, int c)
|
|
{
|
|
if (si->need_reachability_test[c]) {
|
|
struct commit *commit = lookup_commit(the_repository,
|
|
&si->shallow->oid[c]);
|
|
|
|
if (!si->commits) {
|
|
struct commit_array ca;
|
|
|
|
memset(&ca, 0, sizeof(ca));
|
|
refs_head_ref(get_main_ref_store(the_repository),
|
|
add_ref, &ca);
|
|
refs_for_each_ref(get_main_ref_store(the_repository),
|
|
add_ref, &ca);
|
|
si->commits = ca.commits;
|
|
si->nr_commits = ca.nr;
|
|
}
|
|
|
|
si->reachable[c] = repo_in_merge_bases_many(the_repository,
|
|
commit,
|
|
si->nr_commits,
|
|
si->commits,
|
|
1);
|
|
if (si->reachable[c] < 0)
|
|
exit(128);
|
|
si->need_reachability_test[c] = 0;
|
|
}
|
|
return si->reachable[c];
|
|
}
|