mirror of
https://github.com/git/git
synced 2024-10-30 04:01:21 +00:00
bd0b42aed3
When fetching using protocol v2, the remote may send a "shallow-info" section if the client is shallow. If so, Git as the client currently takes the shallow file lock, even if the "shallow-info" section is empty. This is not a problem except that Git does not support taking the shallow file lock after modifying the shallow file, because is_repository_shallow() stores information that is never cleared. And this take-after-modify occurs when Git does a tag-following fetch from a shallow repository on a transport that does not support tag following (since in this case, 2 fetches are performed). To solve this issue, take the shallow file lock (and perform all other shallow processing) only if the "shallow-info" section is non-empty; otherwise, behave as if it were empty. A full solution (probably, ensuring that any action of committing shallow file locks also includes clearing the information stored by is_repository_shallow()) would solve the issue without need for this patch, but this patch is independently useful (as an optimization to prevent writing a file in an unnecessary case), hence why I wrote it. I have included a NEEDSWORK outlining the full solution. Signed-off-by: Jonathan Tan <jonathantanmy@google.com> Signed-off-by: Junio C Hamano <gitster@pobox.com>
793 lines
21 KiB
C
793 lines
21 KiB
C
#include "cache.h"
|
|
#include "repository.h"
|
|
#include "tempfile.h"
|
|
#include "lockfile.h"
|
|
#include "object-store.h"
|
|
#include "commit.h"
|
|
#include "tag.h"
|
|
#include "pkt-line.h"
|
|
#include "remote.h"
|
|
#include "refs.h"
|
|
#include "sha1-array.h"
|
|
#include "diff.h"
|
|
#include "revision.h"
|
|
#include "commit-slab.h"
|
|
#include "revision.h"
|
|
#include "list-objects.h"
|
|
#include "commit-slab.h"
|
|
#include "repository.h"
|
|
#include "commit-reach.h"
|
|
|
|
void set_alternate_shallow_file(struct repository *r, const char *path, int override)
|
|
{
|
|
if (r->parsed_objects->is_shallow != -1)
|
|
BUG("is_repository_shallow must not be called before set_alternate_shallow_file");
|
|
if (r->parsed_objects->alternate_shallow_file && !override)
|
|
return;
|
|
free(r->parsed_objects->alternate_shallow_file);
|
|
r->parsed_objects->alternate_shallow_file = xstrdup_or_null(path);
|
|
}
|
|
|
|
int register_shallow(struct repository *r, const struct object_id *oid)
|
|
{
|
|
struct commit_graft *graft =
|
|
xmalloc(sizeof(struct commit_graft));
|
|
struct commit *commit = lookup_commit(the_repository, oid);
|
|
|
|
oidcpy(&graft->oid, oid);
|
|
graft->nr_parent = -1;
|
|
if (commit && commit->object.parsed)
|
|
commit->parents = NULL;
|
|
return register_commit_graft(r, graft, 0);
|
|
}
|
|
|
|
int is_repository_shallow(struct repository *r)
|
|
{
|
|
/*
|
|
* NEEDSWORK: This function updates
|
|
* r->parsed_objects->{is_shallow,shallow_stat} as a side effect but
|
|
* there is no corresponding function to clear them when the shallow
|
|
* file is updated.
|
|
*/
|
|
|
|
FILE *fp;
|
|
char buf[1024];
|
|
const char *path = r->parsed_objects->alternate_shallow_file;
|
|
|
|
if (r->parsed_objects->is_shallow >= 0)
|
|
return r->parsed_objects->is_shallow;
|
|
|
|
if (!path)
|
|
path = git_path_shallow(r);
|
|
/*
|
|
* fetch-pack sets '--shallow-file ""' as an indicator that no
|
|
* shallow file should be used. We could just open it and it
|
|
* will likely fail. But let's do an explicit check instead.
|
|
*/
|
|
if (!*path || (fp = fopen(path, "r")) == NULL) {
|
|
stat_validity_clear(r->parsed_objects->shallow_stat);
|
|
r->parsed_objects->is_shallow = 0;
|
|
return r->parsed_objects->is_shallow;
|
|
}
|
|
stat_validity_update(r->parsed_objects->shallow_stat, fileno(fp));
|
|
r->parsed_objects->is_shallow = 1;
|
|
|
|
while (fgets(buf, sizeof(buf), fp)) {
|
|
struct object_id oid;
|
|
if (get_oid_hex(buf, &oid))
|
|
die("bad shallow line: %s", buf);
|
|
register_shallow(r, &oid);
|
|
}
|
|
fclose(fp);
|
|
return r->parsed_objects->is_shallow;
|
|
}
|
|
|
|
/*
|
|
* TODO: use "int" elemtype instead of "int *" when/if commit-slab
|
|
* supports a "valid" flag.
|
|
*/
|
|
define_commit_slab(commit_depth, int *);
|
|
struct commit_list *get_shallow_commits(struct object_array *heads, int depth,
|
|
int shallow_flag, int not_shallow_flag)
|
|
{
|
|
int i = 0, cur_depth = 0;
|
|
struct commit_list *result = NULL;
|
|
struct object_array stack = OBJECT_ARRAY_INIT;
|
|
struct commit *commit = NULL;
|
|
struct commit_graft *graft;
|
|
struct commit_depth depths;
|
|
|
|
init_commit_depth(&depths);
|
|
while (commit || i < heads->nr || stack.nr) {
|
|
struct commit_list *p;
|
|
if (!commit) {
|
|
if (i < heads->nr) {
|
|
int **depth_slot;
|
|
commit = (struct commit *)
|
|
deref_tag(the_repository,
|
|
heads->objects[i++].item,
|
|
NULL, 0);
|
|
if (!commit || commit->object.type != OBJ_COMMIT) {
|
|
commit = NULL;
|
|
continue;
|
|
}
|
|
depth_slot = commit_depth_at(&depths, commit);
|
|
if (!*depth_slot)
|
|
*depth_slot = xmalloc(sizeof(int));
|
|
**depth_slot = 0;
|
|
cur_depth = 0;
|
|
} else {
|
|
commit = (struct commit *)
|
|
object_array_pop(&stack);
|
|
cur_depth = **commit_depth_at(&depths, commit);
|
|
}
|
|
}
|
|
parse_commit_or_die(commit);
|
|
cur_depth++;
|
|
if ((depth != INFINITE_DEPTH && cur_depth >= depth) ||
|
|
(is_repository_shallow(the_repository) && !commit->parents &&
|
|
(graft = lookup_commit_graft(the_repository, &commit->object.oid)) != NULL &&
|
|
graft->nr_parent < 0)) {
|
|
commit_list_insert(commit, &result);
|
|
commit->object.flags |= shallow_flag;
|
|
commit = NULL;
|
|
continue;
|
|
}
|
|
commit->object.flags |= not_shallow_flag;
|
|
for (p = commit->parents, commit = NULL; p; p = p->next) {
|
|
int **depth_slot = commit_depth_at(&depths, p->item);
|
|
if (!*depth_slot) {
|
|
*depth_slot = xmalloc(sizeof(int));
|
|
**depth_slot = cur_depth;
|
|
} else {
|
|
if (cur_depth >= **depth_slot)
|
|
continue;
|
|
**depth_slot = cur_depth;
|
|
}
|
|
if (p->next)
|
|
add_object_array(&p->item->object,
|
|
NULL, &stack);
|
|
else {
|
|
commit = p->item;
|
|
cur_depth = **commit_depth_at(&depths, commit);
|
|
}
|
|
}
|
|
}
|
|
for (i = 0; i < depths.slab_count; i++) {
|
|
int j;
|
|
|
|
for (j = 0; j < depths.slab_size; j++)
|
|
free(depths.slab[i][j]);
|
|
}
|
|
clear_commit_depth(&depths);
|
|
|
|
return result;
|
|
}
|
|
|
|
static void show_commit(struct commit *commit, void *data)
|
|
{
|
|
commit_list_insert(commit, data);
|
|
}
|
|
|
|
/*
|
|
* Given rev-list arguments, run rev-list. All reachable commits
|
|
* except border ones are marked with not_shallow_flag. Border commits
|
|
* are marked with shallow_flag. The list of border/shallow commits
|
|
* are also returned.
|
|
*/
|
|
struct commit_list *get_shallow_commits_by_rev_list(int ac, const char **av,
|
|
int shallow_flag,
|
|
int not_shallow_flag)
|
|
{
|
|
struct commit_list *result = NULL, *p;
|
|
struct commit_list *not_shallow_list = NULL;
|
|
struct rev_info revs;
|
|
int both_flags = shallow_flag | not_shallow_flag;
|
|
|
|
/*
|
|
* SHALLOW (excluded) and NOT_SHALLOW (included) should not be
|
|
* set at this point. But better be safe than sorry.
|
|
*/
|
|
clear_object_flags(both_flags);
|
|
|
|
is_repository_shallow(the_repository); /* make sure shallows are read */
|
|
|
|
repo_init_revisions(the_repository, &revs, NULL);
|
|
save_commit_buffer = 0;
|
|
setup_revisions(ac, av, &revs, NULL);
|
|
|
|
if (prepare_revision_walk(&revs))
|
|
die("revision walk setup failed");
|
|
traverse_commit_list(&revs, show_commit, NULL, ¬_shallow_list);
|
|
|
|
if (!not_shallow_list)
|
|
die("no commits selected for shallow requests");
|
|
|
|
/* Mark all reachable commits as NOT_SHALLOW */
|
|
for (p = not_shallow_list; p; p = p->next)
|
|
p->item->object.flags |= not_shallow_flag;
|
|
|
|
/*
|
|
* mark border commits SHALLOW + NOT_SHALLOW.
|
|
* We cannot clear NOT_SHALLOW right now. Imagine border
|
|
* commit A is processed first, then commit B, whose parent is
|
|
* A, later. If NOT_SHALLOW on A is cleared at step 1, B
|
|
* itself is considered border at step 2, which is incorrect.
|
|
*/
|
|
for (p = not_shallow_list; p; p = p->next) {
|
|
struct commit *c = p->item;
|
|
struct commit_list *parent;
|
|
|
|
if (parse_commit(c))
|
|
die("unable to parse commit %s",
|
|
oid_to_hex(&c->object.oid));
|
|
|
|
for (parent = c->parents; parent; parent = parent->next)
|
|
if (!(parent->item->object.flags & not_shallow_flag)) {
|
|
c->object.flags |= shallow_flag;
|
|
commit_list_insert(c, &result);
|
|
break;
|
|
}
|
|
}
|
|
free_commit_list(not_shallow_list);
|
|
|
|
/*
|
|
* Now we can clean up NOT_SHALLOW on border commits. Having
|
|
* both flags set can confuse the caller.
|
|
*/
|
|
for (p = result; p; p = p->next) {
|
|
struct object *o = &p->item->object;
|
|
if ((o->flags & both_flags) == both_flags)
|
|
o->flags &= ~not_shallow_flag;
|
|
}
|
|
return result;
|
|
}
|
|
|
|
static void check_shallow_file_for_update(struct repository *r)
|
|
{
|
|
if (r->parsed_objects->is_shallow == -1)
|
|
BUG("shallow must be initialized by now");
|
|
|
|
if (!stat_validity_check(r->parsed_objects->shallow_stat, git_path_shallow(the_repository)))
|
|
die("shallow file has changed since we read it");
|
|
}
|
|
|
|
#define SEEN_ONLY 1
|
|
#define VERBOSE 2
|
|
#define QUICK 4
|
|
|
|
struct write_shallow_data {
|
|
struct strbuf *out;
|
|
int use_pack_protocol;
|
|
int count;
|
|
unsigned flags;
|
|
};
|
|
|
|
static int write_one_shallow(const struct commit_graft *graft, void *cb_data)
|
|
{
|
|
struct write_shallow_data *data = cb_data;
|
|
const char *hex = oid_to_hex(&graft->oid);
|
|
if (graft->nr_parent != -1)
|
|
return 0;
|
|
if (data->flags & QUICK) {
|
|
if (!has_object_file(&graft->oid))
|
|
return 0;
|
|
} else if (data->flags & SEEN_ONLY) {
|
|
struct commit *c = lookup_commit(the_repository, &graft->oid);
|
|
if (!c || !(c->object.flags & SEEN)) {
|
|
if (data->flags & VERBOSE)
|
|
printf("Removing %s from .git/shallow\n",
|
|
oid_to_hex(&c->object.oid));
|
|
return 0;
|
|
}
|
|
}
|
|
data->count++;
|
|
if (data->use_pack_protocol)
|
|
packet_buf_write(data->out, "shallow %s", hex);
|
|
else {
|
|
strbuf_addstr(data->out, hex);
|
|
strbuf_addch(data->out, '\n');
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int write_shallow_commits_1(struct strbuf *out, int use_pack_protocol,
|
|
const struct oid_array *extra,
|
|
unsigned flags)
|
|
{
|
|
struct write_shallow_data data;
|
|
int i;
|
|
data.out = out;
|
|
data.use_pack_protocol = use_pack_protocol;
|
|
data.count = 0;
|
|
data.flags = flags;
|
|
for_each_commit_graft(write_one_shallow, &data);
|
|
if (!extra)
|
|
return data.count;
|
|
for (i = 0; i < extra->nr; i++) {
|
|
strbuf_addstr(out, oid_to_hex(extra->oid + i));
|
|
strbuf_addch(out, '\n');
|
|
data.count++;
|
|
}
|
|
return data.count;
|
|
}
|
|
|
|
int write_shallow_commits(struct strbuf *out, int use_pack_protocol,
|
|
const struct oid_array *extra)
|
|
{
|
|
return write_shallow_commits_1(out, use_pack_protocol, extra, 0);
|
|
}
|
|
|
|
const char *setup_temporary_shallow(const struct oid_array *extra)
|
|
{
|
|
struct tempfile *temp;
|
|
struct strbuf sb = STRBUF_INIT;
|
|
|
|
if (write_shallow_commits(&sb, 0, extra)) {
|
|
temp = xmks_tempfile(git_path("shallow_XXXXXX"));
|
|
|
|
if (write_in_full(temp->fd, sb.buf, sb.len) < 0 ||
|
|
close_tempfile_gently(temp) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_tempfile_path(temp));
|
|
strbuf_release(&sb);
|
|
return get_tempfile_path(temp);
|
|
}
|
|
/*
|
|
* is_repository_shallow() sees empty string as "no shallow
|
|
* file".
|
|
*/
|
|
return "";
|
|
}
|
|
|
|
void setup_alternate_shallow(struct lock_file *shallow_lock,
|
|
const char **alternate_shallow_file,
|
|
const struct oid_array *extra)
|
|
{
|
|
struct strbuf sb = STRBUF_INIT;
|
|
int fd;
|
|
|
|
fd = hold_lock_file_for_update(shallow_lock,
|
|
git_path_shallow(the_repository),
|
|
LOCK_DIE_ON_ERROR);
|
|
check_shallow_file_for_update(the_repository);
|
|
if (write_shallow_commits(&sb, 0, extra)) {
|
|
if (write_in_full(fd, sb.buf, sb.len) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_lock_file_path(shallow_lock));
|
|
*alternate_shallow_file = get_lock_file_path(shallow_lock);
|
|
} else
|
|
/*
|
|
* is_repository_shallow() sees empty string as "no
|
|
* shallow file".
|
|
*/
|
|
*alternate_shallow_file = "";
|
|
strbuf_release(&sb);
|
|
}
|
|
|
|
static int advertise_shallow_grafts_cb(const struct commit_graft *graft, void *cb)
|
|
{
|
|
int fd = *(int *)cb;
|
|
if (graft->nr_parent == -1)
|
|
packet_write_fmt(fd, "shallow %s\n", oid_to_hex(&graft->oid));
|
|
return 0;
|
|
}
|
|
|
|
void advertise_shallow_grafts(int fd)
|
|
{
|
|
if (!is_repository_shallow(the_repository))
|
|
return;
|
|
for_each_commit_graft(advertise_shallow_grafts_cb, &fd);
|
|
}
|
|
|
|
/*
|
|
* mark_reachable_objects() should have been run prior to this and all
|
|
* reachable commits marked as "SEEN", except when quick_prune is non-zero,
|
|
* in which case lines are excised from the shallow file if they refer to
|
|
* commits that do not exist (any longer).
|
|
*/
|
|
void prune_shallow(unsigned options)
|
|
{
|
|
struct lock_file shallow_lock = LOCK_INIT;
|
|
struct strbuf sb = STRBUF_INIT;
|
|
unsigned flags = SEEN_ONLY;
|
|
int fd;
|
|
|
|
if (options & PRUNE_QUICK)
|
|
flags |= QUICK;
|
|
|
|
if (options & PRUNE_SHOW_ONLY) {
|
|
flags |= VERBOSE;
|
|
write_shallow_commits_1(&sb, 0, NULL, flags);
|
|
strbuf_release(&sb);
|
|
return;
|
|
}
|
|
fd = hold_lock_file_for_update(&shallow_lock,
|
|
git_path_shallow(the_repository),
|
|
LOCK_DIE_ON_ERROR);
|
|
check_shallow_file_for_update(the_repository);
|
|
if (write_shallow_commits_1(&sb, 0, NULL, flags)) {
|
|
if (write_in_full(fd, sb.buf, sb.len) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_lock_file_path(&shallow_lock));
|
|
commit_lock_file(&shallow_lock);
|
|
} else {
|
|
unlink(git_path_shallow(the_repository));
|
|
rollback_lock_file(&shallow_lock);
|
|
}
|
|
strbuf_release(&sb);
|
|
}
|
|
|
|
struct trace_key trace_shallow = TRACE_KEY_INIT(SHALLOW);
|
|
|
|
/*
|
|
* Step 1, split sender shallow commits into "ours" and "theirs"
|
|
* Step 2, clean "ours" based on .git/shallow
|
|
*/
|
|
void prepare_shallow_info(struct shallow_info *info, struct oid_array *sa)
|
|
{
|
|
int i;
|
|
trace_printf_key(&trace_shallow, "shallow: prepare_shallow_info\n");
|
|
memset(info, 0, sizeof(*info));
|
|
info->shallow = sa;
|
|
if (!sa)
|
|
return;
|
|
ALLOC_ARRAY(info->ours, sa->nr);
|
|
ALLOC_ARRAY(info->theirs, sa->nr);
|
|
for (i = 0; i < sa->nr; i++) {
|
|
if (has_object_file(sa->oid + i)) {
|
|
struct commit_graft *graft;
|
|
graft = lookup_commit_graft(the_repository,
|
|
&sa->oid[i]);
|
|
if (graft && graft->nr_parent < 0)
|
|
continue;
|
|
info->ours[info->nr_ours++] = i;
|
|
} else
|
|
info->theirs[info->nr_theirs++] = i;
|
|
}
|
|
}
|
|
|
|
void clear_shallow_info(struct shallow_info *info)
|
|
{
|
|
free(info->ours);
|
|
free(info->theirs);
|
|
}
|
|
|
|
/* Step 4, remove non-existent ones in "theirs" after getting the pack */
|
|
|
|
void remove_nonexistent_theirs_shallow(struct shallow_info *info)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
int i, dst;
|
|
trace_printf_key(&trace_shallow, "shallow: remove_nonexistent_theirs_shallow\n");
|
|
for (i = dst = 0; i < info->nr_theirs; i++) {
|
|
if (i != dst)
|
|
info->theirs[dst] = info->theirs[i];
|
|
if (has_object_file(oid + info->theirs[i]))
|
|
dst++;
|
|
}
|
|
info->nr_theirs = dst;
|
|
}
|
|
|
|
define_commit_slab(ref_bitmap, uint32_t *);
|
|
|
|
#define POOL_SIZE (512 * 1024)
|
|
|
|
struct paint_info {
|
|
struct ref_bitmap ref_bitmap;
|
|
unsigned nr_bits;
|
|
char **pools;
|
|
char *free, *end;
|
|
unsigned pool_count;
|
|
};
|
|
|
|
static uint32_t *paint_alloc(struct paint_info *info)
|
|
{
|
|
unsigned nr = DIV_ROUND_UP(info->nr_bits, 32);
|
|
unsigned size = nr * sizeof(uint32_t);
|
|
void *p;
|
|
if (!info->pool_count || size > info->end - info->free) {
|
|
if (size > POOL_SIZE)
|
|
BUG("pool size too small for %d in paint_alloc()",
|
|
size);
|
|
info->pool_count++;
|
|
REALLOC_ARRAY(info->pools, info->pool_count);
|
|
info->free = xmalloc(POOL_SIZE);
|
|
info->pools[info->pool_count - 1] = info->free;
|
|
info->end = info->free + POOL_SIZE;
|
|
}
|
|
p = info->free;
|
|
info->free += size;
|
|
return p;
|
|
}
|
|
|
|
/*
|
|
* Given a commit SHA-1, walk down to parents until either SEEN,
|
|
* UNINTERESTING or BOTTOM is hit. Set the id-th bit in ref_bitmap for
|
|
* all walked commits.
|
|
*/
|
|
static void paint_down(struct paint_info *info, const struct object_id *oid,
|
|
unsigned int id)
|
|
{
|
|
unsigned int i, nr;
|
|
struct commit_list *head = NULL;
|
|
int bitmap_nr = DIV_ROUND_UP(info->nr_bits, 32);
|
|
size_t bitmap_size = st_mult(sizeof(uint32_t), bitmap_nr);
|
|
struct commit *c = lookup_commit_reference_gently(the_repository, oid,
|
|
1);
|
|
uint32_t *tmp; /* to be freed before return */
|
|
uint32_t *bitmap;
|
|
|
|
if (!c)
|
|
return;
|
|
|
|
tmp = xmalloc(bitmap_size);
|
|
bitmap = paint_alloc(info);
|
|
memset(bitmap, 0, bitmap_size);
|
|
bitmap[id / 32] |= (1U << (id % 32));
|
|
commit_list_insert(c, &head);
|
|
while (head) {
|
|
struct commit_list *p;
|
|
struct commit *c = pop_commit(&head);
|
|
uint32_t **refs = ref_bitmap_at(&info->ref_bitmap, c);
|
|
|
|
/* XXX check "UNINTERESTING" from pack bitmaps if available */
|
|
if (c->object.flags & (SEEN | UNINTERESTING))
|
|
continue;
|
|
else
|
|
c->object.flags |= SEEN;
|
|
|
|
if (*refs == NULL)
|
|
*refs = bitmap;
|
|
else {
|
|
memcpy(tmp, *refs, bitmap_size);
|
|
for (i = 0; i < bitmap_nr; i++)
|
|
tmp[i] |= bitmap[i];
|
|
if (memcmp(tmp, *refs, bitmap_size)) {
|
|
*refs = paint_alloc(info);
|
|
memcpy(*refs, tmp, bitmap_size);
|
|
}
|
|
}
|
|
|
|
if (c->object.flags & BOTTOM)
|
|
continue;
|
|
|
|
if (parse_commit(c))
|
|
die("unable to parse commit %s",
|
|
oid_to_hex(&c->object.oid));
|
|
|
|
for (p = c->parents; p; p = p->next) {
|
|
if (p->item->object.flags & SEEN)
|
|
continue;
|
|
commit_list_insert(p->item, &head);
|
|
}
|
|
}
|
|
|
|
nr = get_max_object_index();
|
|
for (i = 0; i < nr; i++) {
|
|
struct object *o = get_indexed_object(i);
|
|
if (o && o->type == OBJ_COMMIT)
|
|
o->flags &= ~SEEN;
|
|
}
|
|
|
|
free(tmp);
|
|
}
|
|
|
|
static int mark_uninteresting(const char *refname, const struct object_id *oid,
|
|
int flags, void *cb_data)
|
|
{
|
|
struct commit *commit = lookup_commit_reference_gently(the_repository,
|
|
oid, 1);
|
|
if (!commit)
|
|
return 0;
|
|
commit->object.flags |= UNINTERESTING;
|
|
mark_parents_uninteresting(commit);
|
|
return 0;
|
|
}
|
|
|
|
static void post_assign_shallow(struct shallow_info *info,
|
|
struct ref_bitmap *ref_bitmap,
|
|
int *ref_status);
|
|
/*
|
|
* Step 6(+7), associate shallow commits with new refs
|
|
*
|
|
* info->ref must be initialized before calling this function.
|
|
*
|
|
* If used is not NULL, it's an array of info->shallow->nr
|
|
* bitmaps. The n-th bit set in the m-th bitmap if ref[n] needs the
|
|
* m-th shallow commit from info->shallow.
|
|
*
|
|
* If used is NULL, "ours" and "theirs" are updated. And if ref_status
|
|
* is not NULL it's an array of ref->nr ints. ref_status[i] is true if
|
|
* the ref needs some shallow commits from either info->ours or
|
|
* info->theirs.
|
|
*/
|
|
void assign_shallow_commits_to_refs(struct shallow_info *info,
|
|
uint32_t **used, int *ref_status)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
struct oid_array *ref = info->ref;
|
|
unsigned int i, nr;
|
|
int *shallow, nr_shallow = 0;
|
|
struct paint_info pi;
|
|
|
|
trace_printf_key(&trace_shallow, "shallow: assign_shallow_commits_to_refs\n");
|
|
ALLOC_ARRAY(shallow, info->nr_ours + info->nr_theirs);
|
|
for (i = 0; i < info->nr_ours; i++)
|
|
shallow[nr_shallow++] = info->ours[i];
|
|
for (i = 0; i < info->nr_theirs; i++)
|
|
shallow[nr_shallow++] = info->theirs[i];
|
|
|
|
/*
|
|
* Prepare the commit graph to track what refs can reach what
|
|
* (new) shallow commits.
|
|
*/
|
|
nr = get_max_object_index();
|
|
for (i = 0; i < nr; i++) {
|
|
struct object *o = get_indexed_object(i);
|
|
if (!o || o->type != OBJ_COMMIT)
|
|
continue;
|
|
|
|
o->flags &= ~(UNINTERESTING | BOTTOM | SEEN);
|
|
}
|
|
|
|
memset(&pi, 0, sizeof(pi));
|
|
init_ref_bitmap(&pi.ref_bitmap);
|
|
pi.nr_bits = ref->nr;
|
|
|
|
/*
|
|
* "--not --all" to cut short the traversal if new refs
|
|
* connect to old refs. If not (e.g. force ref updates) it'll
|
|
* have to go down to the current shallow commits.
|
|
*/
|
|
head_ref(mark_uninteresting, NULL);
|
|
for_each_ref(mark_uninteresting, NULL);
|
|
|
|
/* Mark potential bottoms so we won't go out of bound */
|
|
for (i = 0; i < nr_shallow; i++) {
|
|
struct commit *c = lookup_commit(the_repository,
|
|
&oid[shallow[i]]);
|
|
c->object.flags |= BOTTOM;
|
|
}
|
|
|
|
for (i = 0; i < ref->nr; i++)
|
|
paint_down(&pi, ref->oid + i, i);
|
|
|
|
if (used) {
|
|
int bitmap_size = DIV_ROUND_UP(pi.nr_bits, 32) * sizeof(uint32_t);
|
|
memset(used, 0, sizeof(*used) * info->shallow->nr);
|
|
for (i = 0; i < nr_shallow; i++) {
|
|
const struct commit *c = lookup_commit(the_repository,
|
|
&oid[shallow[i]]);
|
|
uint32_t **map = ref_bitmap_at(&pi.ref_bitmap, c);
|
|
if (*map)
|
|
used[shallow[i]] = xmemdupz(*map, bitmap_size);
|
|
}
|
|
/*
|
|
* unreachable shallow commits are not removed from
|
|
* "ours" and "theirs". The user is supposed to run
|
|
* step 7 on every ref separately and not trust "ours"
|
|
* and "theirs" any more.
|
|
*/
|
|
} else
|
|
post_assign_shallow(info, &pi.ref_bitmap, ref_status);
|
|
|
|
clear_ref_bitmap(&pi.ref_bitmap);
|
|
for (i = 0; i < pi.pool_count; i++)
|
|
free(pi.pools[i]);
|
|
free(pi.pools);
|
|
free(shallow);
|
|
}
|
|
|
|
struct commit_array {
|
|
struct commit **commits;
|
|
int nr, alloc;
|
|
};
|
|
|
|
static int add_ref(const char *refname, const struct object_id *oid,
|
|
int flags, void *cb_data)
|
|
{
|
|
struct commit_array *ca = cb_data;
|
|
ALLOC_GROW(ca->commits, ca->nr + 1, ca->alloc);
|
|
ca->commits[ca->nr] = lookup_commit_reference_gently(the_repository,
|
|
oid, 1);
|
|
if (ca->commits[ca->nr])
|
|
ca->nr++;
|
|
return 0;
|
|
}
|
|
|
|
static void update_refstatus(int *ref_status, int nr, uint32_t *bitmap)
|
|
{
|
|
unsigned int i;
|
|
if (!ref_status)
|
|
return;
|
|
for (i = 0; i < nr; i++)
|
|
if (bitmap[i / 32] & (1U << (i % 32)))
|
|
ref_status[i]++;
|
|
}
|
|
|
|
/*
|
|
* Step 7, reachability test on "ours" at commit level
|
|
*/
|
|
static void post_assign_shallow(struct shallow_info *info,
|
|
struct ref_bitmap *ref_bitmap,
|
|
int *ref_status)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
struct commit *c;
|
|
uint32_t **bitmap;
|
|
int dst, i, j;
|
|
int bitmap_nr = DIV_ROUND_UP(info->ref->nr, 32);
|
|
struct commit_array ca;
|
|
|
|
trace_printf_key(&trace_shallow, "shallow: post_assign_shallow\n");
|
|
if (ref_status)
|
|
memset(ref_status, 0, sizeof(*ref_status) * info->ref->nr);
|
|
|
|
/* Remove unreachable shallow commits from "theirs" */
|
|
for (i = dst = 0; i < info->nr_theirs; i++) {
|
|
if (i != dst)
|
|
info->theirs[dst] = info->theirs[i];
|
|
c = lookup_commit(the_repository, &oid[info->theirs[i]]);
|
|
bitmap = ref_bitmap_at(ref_bitmap, c);
|
|
if (!*bitmap)
|
|
continue;
|
|
for (j = 0; j < bitmap_nr; j++)
|
|
if (bitmap[0][j]) {
|
|
update_refstatus(ref_status, info->ref->nr, *bitmap);
|
|
dst++;
|
|
break;
|
|
}
|
|
}
|
|
info->nr_theirs = dst;
|
|
|
|
memset(&ca, 0, sizeof(ca));
|
|
head_ref(add_ref, &ca);
|
|
for_each_ref(add_ref, &ca);
|
|
|
|
/* Remove unreachable shallow commits from "ours" */
|
|
for (i = dst = 0; i < info->nr_ours; i++) {
|
|
if (i != dst)
|
|
info->ours[dst] = info->ours[i];
|
|
c = lookup_commit(the_repository, &oid[info->ours[i]]);
|
|
bitmap = ref_bitmap_at(ref_bitmap, c);
|
|
if (!*bitmap)
|
|
continue;
|
|
for (j = 0; j < bitmap_nr; j++)
|
|
if (bitmap[0][j] &&
|
|
/* Step 7, reachability test at commit level */
|
|
!in_merge_bases_many(c, ca.nr, ca.commits)) {
|
|
update_refstatus(ref_status, info->ref->nr, *bitmap);
|
|
dst++;
|
|
break;
|
|
}
|
|
}
|
|
info->nr_ours = dst;
|
|
|
|
free(ca.commits);
|
|
}
|
|
|
|
/* (Delayed) step 7, reachability test at commit level */
|
|
int delayed_reachability_test(struct shallow_info *si, int c)
|
|
{
|
|
if (si->need_reachability_test[c]) {
|
|
struct commit *commit = lookup_commit(the_repository,
|
|
&si->shallow->oid[c]);
|
|
|
|
if (!si->commits) {
|
|
struct commit_array ca;
|
|
|
|
memset(&ca, 0, sizeof(ca));
|
|
head_ref(add_ref, &ca);
|
|
for_each_ref(add_ref, &ca);
|
|
si->commits = ca.commits;
|
|
si->nr_commits = ca.nr;
|
|
}
|
|
|
|
si->reachable[c] = in_merge_bases_many(commit,
|
|
si->nr_commits,
|
|
si->commits);
|
|
si->need_reachability_test[c] = 0;
|
|
}
|
|
return si->reachable[c];
|
|
}
|