mirror of
https://github.com/git-for-windows/git.git
synced 2026-04-29 22:49:20 -05:00
When a shallowed repository gets deepened beyond the beginning of a merged branch, we may end up with some shallows that are hidden behind the reachable shallow commits. Added test 'fetching deepen beyond merged branch' exposes that behaviour. An example showing the problem based on added test: 0. Whole initial git repo to be cloned from Graph: * 033585d (HEAD -> main) Merge branch 'branch' |\ | * 984f8b1 (branch) five | * ecb578a four |/ * 0cb5d20 three * 2b4e70d two * 61ba98b one 1. Initial shallow clone --depth=3 (all good) Shallows: 2b4e70da2a10e1d3231a0ae2df396024735601f1 ecb578a3cf37198d122ae5df7efed9abaca17144 Graph: * 033585d (HEAD -> main) Merge branch 'branch' |\ | * 984f8b1 five | * ecb578a (grafted) four * 0cb5d20 three * 2b4e70d (grafted) two 2. Deepen shallow clone with fetch --deepen=1 (NOT OK) Shallows: 0cb5d204f4ef96ed241feb0f2088c9f4794ba758 61ba98be443fd51c542eb66585a1f6d7e15fcdae Graph: * 033585d (HEAD -> main) Merge branch 'branch' |\ | * 984f8b1 five | * ecb578a four |/ * 0cb5d20 (grafted) three --- Note that second shallow commit 61ba98be443fd51c542eb66585a1f6d7e15fcdae is not reachable. On the other hand, it seems that equivalent absolute depth driven fetches result in all the correct shallows. That led to this proposal, which unifies absolute and relative deepening in a way that the same get_shallow_commits() call is used in both cases. The difference is only that depth is adapted for relative deepening by measuring equivalent depth of current local shallow commits in the current remote repo. Thus a new function get_shallows_depth() has been added and the function get_reachable_list() became redundant / removed. Same example showing the corrected second step: 2. Deepen shallow clone with fetch --deepen=1 (all good) Shallow: 61ba98be443fd51c542eb66585a1f6d7e15fcdae Graph: * 033585d (HEAD -> main) Merge branch 'branch' |\ | * 984f8b1 five | * ecb578a four |/ * 0cb5d20 three * 2b4e70d two * 61ba98b (grafted) one The get_shallows_depth() function also shares the logic of the get_shallow_commits() function, but it focuses on counting depth of each existing shallow commit. The minimum result is stored as 'data->deepen_relative', which is set not to be zero for relative deepening anyway. That way we can always sum 'data->deepen_relative' and 'depth' values, because 'data->deepen_relative' is always 0 in absolute deepening. To avoid duplicating logic between get_shallows_depth() and get_shallow_commits(), get_shallow_commits() was modified so that it is used by get_shallows_depth(). Signed-off-by: Samo Pogačnik <samo_pogacnik@t-2.net> Signed-off-by: Junio C Hamano <gitster@pobox.com>
905 lines
25 KiB
C
905 lines
25 KiB
C
#define USE_THE_REPOSITORY_VARIABLE
|
|
|
|
#include "git-compat-util.h"
|
|
#include "hex.h"
|
|
#include "repository.h"
|
|
#include "tempfile.h"
|
|
#include "lockfile.h"
|
|
#include "odb.h"
|
|
#include "commit.h"
|
|
#include "tag.h"
|
|
#include "pkt-line.h"
|
|
#include "refs.h"
|
|
#include "oid-array.h"
|
|
#include "path.h"
|
|
#include "diff.h"
|
|
#include "revision.h"
|
|
#include "commit-slab.h"
|
|
#include "list-objects.h"
|
|
#include "commit-reach.h"
|
|
#include "shallow.h"
|
|
#include "statinfo.h"
|
|
#include "trace.h"
|
|
|
|
void set_alternate_shallow_file(struct repository *r, const char *path, int override)
|
|
{
|
|
if (r->parsed_objects->is_shallow != -1)
|
|
BUG("is_repository_shallow must not be called before set_alternate_shallow_file");
|
|
if (r->parsed_objects->alternate_shallow_file && !override)
|
|
return;
|
|
free(r->parsed_objects->alternate_shallow_file);
|
|
r->parsed_objects->alternate_shallow_file = xstrdup_or_null(path);
|
|
}
|
|
|
|
int register_shallow(struct repository *r, const struct object_id *oid)
|
|
{
|
|
struct commit_graft *graft =
|
|
xmalloc(sizeof(struct commit_graft));
|
|
struct commit *commit = lookup_commit(r, oid);
|
|
|
|
oidcpy(&graft->oid, oid);
|
|
graft->nr_parent = -1;
|
|
if (commit && commit->object.parsed) {
|
|
free_commit_list(commit->parents);
|
|
commit->parents = NULL;
|
|
}
|
|
return register_commit_graft(r, graft, 0);
|
|
}
|
|
|
|
int unregister_shallow(const struct object_id *oid)
|
|
{
|
|
int pos = commit_graft_pos(the_repository, oid);
|
|
if (pos < 0)
|
|
return -1;
|
|
free(the_repository->parsed_objects->grafts[pos]);
|
|
if (pos + 1 < the_repository->parsed_objects->grafts_nr)
|
|
MOVE_ARRAY(the_repository->parsed_objects->grafts + pos,
|
|
the_repository->parsed_objects->grafts + pos + 1,
|
|
the_repository->parsed_objects->grafts_nr - pos - 1);
|
|
the_repository->parsed_objects->grafts_nr--;
|
|
return 0;
|
|
}
|
|
|
|
int is_repository_shallow(struct repository *r)
|
|
{
|
|
FILE *fp;
|
|
char buf[1024];
|
|
const char *path = r->parsed_objects->alternate_shallow_file;
|
|
|
|
if (r->parsed_objects->is_shallow >= 0)
|
|
return r->parsed_objects->is_shallow;
|
|
|
|
if (!path)
|
|
path = git_path_shallow(r);
|
|
/*
|
|
* fetch-pack sets '--shallow-file ""' as an indicator that no
|
|
* shallow file should be used. We could just open it and it
|
|
* will likely fail. But let's do an explicit check instead.
|
|
*/
|
|
if (!*path || (fp = fopen(path, "r")) == NULL) {
|
|
stat_validity_clear(r->parsed_objects->shallow_stat);
|
|
r->parsed_objects->is_shallow = 0;
|
|
return r->parsed_objects->is_shallow;
|
|
}
|
|
stat_validity_update(r->parsed_objects->shallow_stat, fileno(fp));
|
|
r->parsed_objects->is_shallow = 1;
|
|
|
|
while (fgets(buf, sizeof(buf), fp)) {
|
|
struct object_id oid;
|
|
if (get_oid_hex(buf, &oid))
|
|
die("bad shallow line: %s", buf);
|
|
register_shallow(r, &oid);
|
|
}
|
|
fclose(fp);
|
|
return r->parsed_objects->is_shallow;
|
|
}
|
|
|
|
static void reset_repository_shallow(struct repository *r)
|
|
{
|
|
r->parsed_objects->is_shallow = -1;
|
|
stat_validity_clear(r->parsed_objects->shallow_stat);
|
|
parsed_object_pool_reset_commit_grafts(r->parsed_objects);
|
|
}
|
|
|
|
int commit_shallow_file(struct repository *r, struct shallow_lock *lk)
|
|
{
|
|
int res = commit_lock_file(&lk->lock);
|
|
reset_repository_shallow(r);
|
|
|
|
/*
|
|
* Update in-memory data structures with the new shallow information,
|
|
* including unparsing all commits that now have grafts.
|
|
*/
|
|
is_repository_shallow(r);
|
|
|
|
return res;
|
|
}
|
|
|
|
void rollback_shallow_file(struct repository *r, struct shallow_lock *lk)
|
|
{
|
|
rollback_lock_file(&lk->lock);
|
|
reset_repository_shallow(r);
|
|
}
|
|
|
|
/*
|
|
* TODO: use "int" elemtype instead of "int *" when/if commit-slab
|
|
* supports a "valid" flag.
|
|
*/
|
|
define_commit_slab(commit_depth, int *);
|
|
static void free_depth_in_slab(int **ptr)
|
|
{
|
|
FREE_AND_NULL(*ptr);
|
|
}
|
|
/*
|
|
* This is a common internal function that can either return a list of
|
|
* shallow commits or calculate the current maximum depth of a shallow
|
|
* repository, depending on the input parameters.
|
|
*
|
|
* Depth calculation is triggered by passing the `shallows` parameter.
|
|
* In this case, the computed depth is stored in `max_cur_depth` (if it is
|
|
* provided), and the function returns NULL.
|
|
*
|
|
* Otherwise, `max_cur_depth` remains unchanged and the function returns
|
|
* a list of shallow commits.
|
|
*/
|
|
static struct commit_list *get_shallows_or_depth(struct object_array *heads,
|
|
struct object_array *shallows, int *max_cur_depth,
|
|
int depth, int shallow_flag, int not_shallow_flag)
|
|
{
|
|
size_t i = 0;
|
|
int cur_depth = 0, cur_depth_shallow = 0;
|
|
struct commit_list *result = NULL;
|
|
struct object_array stack = OBJECT_ARRAY_INIT;
|
|
struct commit *commit = NULL;
|
|
struct commit_graft *graft;
|
|
struct commit_depth depths;
|
|
|
|
init_commit_depth(&depths);
|
|
while (commit || i < heads->nr || stack.nr) {
|
|
struct commit_list *p;
|
|
if (!commit) {
|
|
if (i < heads->nr) {
|
|
int **depth_slot;
|
|
commit = (struct commit *)
|
|
deref_tag(the_repository,
|
|
heads->objects[i++].item,
|
|
NULL, 0);
|
|
if (!commit || commit->object.type != OBJ_COMMIT) {
|
|
commit = NULL;
|
|
continue;
|
|
}
|
|
depth_slot = commit_depth_at(&depths, commit);
|
|
if (!*depth_slot)
|
|
*depth_slot = xmalloc(sizeof(int));
|
|
**depth_slot = 0;
|
|
cur_depth = 0;
|
|
} else {
|
|
commit = (struct commit *)
|
|
object_array_pop(&stack);
|
|
cur_depth = **commit_depth_at(&depths, commit);
|
|
}
|
|
}
|
|
parse_commit_or_die(commit);
|
|
cur_depth++;
|
|
if (shallows) {
|
|
for (size_t j = 0; j < shallows->nr; j++)
|
|
if (oideq(&commit->object.oid, &shallows->objects[j].item->oid))
|
|
if (!cur_depth_shallow || cur_depth < cur_depth_shallow)
|
|
cur_depth_shallow = cur_depth;
|
|
|
|
if ((is_repository_shallow(the_repository) && !commit->parents &&
|
|
(graft = lookup_commit_graft(the_repository, &commit->object.oid)) != NULL &&
|
|
graft->nr_parent < 0)) {
|
|
commit = NULL;
|
|
continue;
|
|
}
|
|
} else {
|
|
if ((depth != INFINITE_DEPTH && cur_depth >= depth) ||
|
|
(is_repository_shallow(the_repository) && !commit->parents &&
|
|
(graft = lookup_commit_graft(the_repository, &commit->object.oid)) != NULL &&
|
|
graft->nr_parent < 0)) {
|
|
commit_list_insert(commit, &result);
|
|
commit->object.flags |= shallow_flag;
|
|
commit = NULL;
|
|
continue;
|
|
}
|
|
commit->object.flags |= not_shallow_flag;
|
|
}
|
|
for (p = commit->parents, commit = NULL; p; p = p->next) {
|
|
int **depth_slot = commit_depth_at(&depths, p->item);
|
|
if (!*depth_slot) {
|
|
*depth_slot = xmalloc(sizeof(int));
|
|
**depth_slot = cur_depth;
|
|
} else {
|
|
if (cur_depth >= **depth_slot)
|
|
continue;
|
|
**depth_slot = cur_depth;
|
|
}
|
|
if (p->next)
|
|
add_object_array(&p->item->object,
|
|
NULL, &stack);
|
|
else {
|
|
commit = p->item;
|
|
cur_depth = **commit_depth_at(&depths, commit);
|
|
}
|
|
}
|
|
}
|
|
deep_clear_commit_depth(&depths, free_depth_in_slab);
|
|
object_array_clear(&stack);
|
|
|
|
if (shallows && max_cur_depth)
|
|
*max_cur_depth = cur_depth_shallow;
|
|
return result;
|
|
}
|
|
|
|
int get_shallows_depth(struct object_array *heads, struct object_array *shallows)
|
|
{
|
|
int max_cur_depth = 0;
|
|
get_shallows_or_depth(heads, shallows, &max_cur_depth, 0, 0, 0);
|
|
return max_cur_depth;
|
|
|
|
}
|
|
|
|
struct commit_list *get_shallow_commits(struct object_array *heads,
|
|
struct object_array *shallows, int deepen_relative,
|
|
int depth, int shallow_flag, int not_shallow_flag)
|
|
{
|
|
if (shallows && deepen_relative) {
|
|
depth += get_shallows_depth(heads, shallows);
|
|
}
|
|
return get_shallows_or_depth(heads, NULL, NULL,
|
|
depth, shallow_flag, not_shallow_flag);
|
|
}
|
|
|
|
static void show_commit(struct commit *commit, void *data)
|
|
{
|
|
commit_list_insert(commit, data);
|
|
}
|
|
|
|
/*
|
|
* Given rev-list arguments, run rev-list. All reachable commits
|
|
* except border ones are marked with not_shallow_flag. Border commits
|
|
* are marked with shallow_flag. The list of border/shallow commits
|
|
* are also returned.
|
|
*/
|
|
struct commit_list *get_shallow_commits_by_rev_list(struct strvec *argv,
|
|
int shallow_flag,
|
|
int not_shallow_flag)
|
|
{
|
|
struct commit_list *result = NULL, *p;
|
|
struct commit_list *not_shallow_list = NULL;
|
|
struct rev_info revs;
|
|
int both_flags = shallow_flag | not_shallow_flag;
|
|
|
|
/*
|
|
* SHALLOW (excluded) and NOT_SHALLOW (included) should not be
|
|
* set at this point. But better be safe than sorry.
|
|
*/
|
|
clear_object_flags(the_repository, both_flags);
|
|
|
|
is_repository_shallow(the_repository); /* make sure shallows are read */
|
|
|
|
repo_init_revisions(the_repository, &revs, NULL);
|
|
save_commit_buffer = 0;
|
|
setup_revisions_from_strvec(argv, &revs, NULL);
|
|
|
|
if (prepare_revision_walk(&revs))
|
|
die("revision walk setup failed");
|
|
traverse_commit_list(&revs, show_commit, NULL, ¬_shallow_list);
|
|
|
|
if (!not_shallow_list)
|
|
die("no commits selected for shallow requests");
|
|
|
|
/* Mark all reachable commits as NOT_SHALLOW */
|
|
for (p = not_shallow_list; p; p = p->next)
|
|
p->item->object.flags |= not_shallow_flag;
|
|
|
|
/*
|
|
* mark border commits SHALLOW + NOT_SHALLOW.
|
|
* We cannot clear NOT_SHALLOW right now. Imagine border
|
|
* commit A is processed first, then commit B, whose parent is
|
|
* A, later. If NOT_SHALLOW on A is cleared at step 1, B
|
|
* itself is considered border at step 2, which is incorrect.
|
|
*/
|
|
for (p = not_shallow_list; p; p = p->next) {
|
|
struct commit *c = p->item;
|
|
struct commit_list *parent;
|
|
|
|
if (repo_parse_commit(the_repository, c))
|
|
die("unable to parse commit %s",
|
|
oid_to_hex(&c->object.oid));
|
|
|
|
for (parent = c->parents; parent; parent = parent->next)
|
|
if (!(parent->item->object.flags & not_shallow_flag)) {
|
|
c->object.flags |= shallow_flag;
|
|
commit_list_insert(c, &result);
|
|
break;
|
|
}
|
|
}
|
|
free_commit_list(not_shallow_list);
|
|
|
|
/*
|
|
* Now we can clean up NOT_SHALLOW on border commits. Having
|
|
* both flags set can confuse the caller.
|
|
*/
|
|
for (p = result; p; p = p->next) {
|
|
struct object *o = &p->item->object;
|
|
if ((o->flags & both_flags) == both_flags)
|
|
o->flags &= ~not_shallow_flag;
|
|
}
|
|
release_revisions(&revs);
|
|
return result;
|
|
}
|
|
|
|
static void check_shallow_file_for_update(struct repository *r)
|
|
{
|
|
if (r->parsed_objects->is_shallow == -1)
|
|
BUG("shallow must be initialized by now");
|
|
|
|
if (!stat_validity_check(r->parsed_objects->shallow_stat,
|
|
git_path_shallow(r)))
|
|
die("shallow file has changed since we read it");
|
|
}
|
|
|
|
#define SEEN_ONLY 1
|
|
#define VERBOSE 2
|
|
#define QUICK 4
|
|
|
|
struct write_shallow_data {
|
|
struct strbuf *out;
|
|
int use_pack_protocol;
|
|
int count;
|
|
unsigned flags;
|
|
};
|
|
|
|
static int write_one_shallow(const struct commit_graft *graft, void *cb_data)
|
|
{
|
|
struct write_shallow_data *data = cb_data;
|
|
const char *hex = oid_to_hex(&graft->oid);
|
|
if (graft->nr_parent != -1)
|
|
return 0;
|
|
if (data->flags & QUICK) {
|
|
if (!odb_has_object(the_repository->objects, &graft->oid,
|
|
HAS_OBJECT_RECHECK_PACKED | HAS_OBJECT_FETCH_PROMISOR))
|
|
return 0;
|
|
} else if (data->flags & SEEN_ONLY) {
|
|
struct commit *c = lookup_commit(the_repository, &graft->oid);
|
|
if (!c || !(c->object.flags & SEEN)) {
|
|
if (data->flags & VERBOSE)
|
|
printf("Removing %s from .git/shallow\n",
|
|
oid_to_hex(&c->object.oid));
|
|
return 0;
|
|
}
|
|
}
|
|
data->count++;
|
|
if (data->use_pack_protocol)
|
|
packet_buf_write(data->out, "shallow %s", hex);
|
|
else {
|
|
strbuf_addstr(data->out, hex);
|
|
strbuf_addch(data->out, '\n');
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int write_shallow_commits_1(struct strbuf *out, int use_pack_protocol,
|
|
const struct oid_array *extra,
|
|
unsigned flags)
|
|
{
|
|
struct write_shallow_data data = {
|
|
.out = out,
|
|
.use_pack_protocol = use_pack_protocol,
|
|
.flags = flags,
|
|
};
|
|
|
|
for_each_commit_graft(write_one_shallow, &data);
|
|
if (!extra)
|
|
return data.count;
|
|
for (size_t i = 0; i < extra->nr; i++) {
|
|
strbuf_addstr(out, oid_to_hex(extra->oid + i));
|
|
strbuf_addch(out, '\n');
|
|
data.count++;
|
|
}
|
|
return data.count;
|
|
}
|
|
|
|
int write_shallow_commits(struct strbuf *out, int use_pack_protocol,
|
|
const struct oid_array *extra)
|
|
{
|
|
return write_shallow_commits_1(out, use_pack_protocol, extra, 0);
|
|
}
|
|
|
|
const char *setup_temporary_shallow(const struct oid_array *extra)
|
|
{
|
|
struct tempfile *temp;
|
|
struct strbuf sb = STRBUF_INIT;
|
|
|
|
if (write_shallow_commits(&sb, 0, extra)) {
|
|
char *path = repo_git_path(the_repository, "shallow_XXXXXX");
|
|
temp = xmks_tempfile(path);
|
|
free(path);
|
|
|
|
if (write_in_full(temp->fd, sb.buf, sb.len) < 0 ||
|
|
close_tempfile_gently(temp) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_tempfile_path(temp));
|
|
strbuf_release(&sb);
|
|
return get_tempfile_path(temp);
|
|
}
|
|
/*
|
|
* is_repository_shallow() sees empty string as "no shallow
|
|
* file".
|
|
*/
|
|
return "";
|
|
}
|
|
|
|
void setup_alternate_shallow(struct shallow_lock *shallow_lock,
|
|
const char **alternate_shallow_file,
|
|
const struct oid_array *extra)
|
|
{
|
|
struct strbuf sb = STRBUF_INIT;
|
|
int fd;
|
|
|
|
fd = hold_lock_file_for_update(&shallow_lock->lock,
|
|
git_path_shallow(the_repository),
|
|
LOCK_DIE_ON_ERROR);
|
|
check_shallow_file_for_update(the_repository);
|
|
if (write_shallow_commits(&sb, 0, extra)) {
|
|
if (write_in_full(fd, sb.buf, sb.len) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_lock_file_path(&shallow_lock->lock));
|
|
*alternate_shallow_file = get_lock_file_path(&shallow_lock->lock);
|
|
} else
|
|
/*
|
|
* is_repository_shallow() sees empty string as "no
|
|
* shallow file".
|
|
*/
|
|
*alternate_shallow_file = "";
|
|
strbuf_release(&sb);
|
|
}
|
|
|
|
static int advertise_shallow_grafts_cb(const struct commit_graft *graft, void *cb)
|
|
{
|
|
int fd = *(int *)cb;
|
|
if (graft->nr_parent == -1)
|
|
packet_write_fmt(fd, "shallow %s\n", oid_to_hex(&graft->oid));
|
|
return 0;
|
|
}
|
|
|
|
void advertise_shallow_grafts(int fd)
|
|
{
|
|
if (!is_repository_shallow(the_repository))
|
|
return;
|
|
for_each_commit_graft(advertise_shallow_grafts_cb, &fd);
|
|
}
|
|
|
|
/*
|
|
* mark_reachable_objects() should have been run prior to this and all
|
|
* reachable commits marked as "SEEN", except when quick_prune is non-zero,
|
|
* in which case lines are excised from the shallow file if they refer to
|
|
* commits that do not exist (any longer).
|
|
*/
|
|
void prune_shallow(unsigned options)
|
|
{
|
|
struct shallow_lock shallow_lock = SHALLOW_LOCK_INIT;
|
|
struct strbuf sb = STRBUF_INIT;
|
|
unsigned flags = SEEN_ONLY;
|
|
int fd;
|
|
|
|
if (options & PRUNE_QUICK)
|
|
flags |= QUICK;
|
|
|
|
if (options & PRUNE_SHOW_ONLY) {
|
|
flags |= VERBOSE;
|
|
write_shallow_commits_1(&sb, 0, NULL, flags);
|
|
strbuf_release(&sb);
|
|
return;
|
|
}
|
|
fd = hold_lock_file_for_update(&shallow_lock.lock,
|
|
git_path_shallow(the_repository),
|
|
LOCK_DIE_ON_ERROR);
|
|
check_shallow_file_for_update(the_repository);
|
|
if (write_shallow_commits_1(&sb, 0, NULL, flags)) {
|
|
if (write_in_full(fd, sb.buf, sb.len) < 0)
|
|
die_errno("failed to write to %s",
|
|
get_lock_file_path(&shallow_lock.lock));
|
|
commit_shallow_file(the_repository, &shallow_lock);
|
|
} else {
|
|
unlink(git_path_shallow(the_repository));
|
|
rollback_shallow_file(the_repository, &shallow_lock);
|
|
}
|
|
strbuf_release(&sb);
|
|
}
|
|
|
|
struct trace_key trace_shallow = TRACE_KEY_INIT(SHALLOW);
|
|
|
|
/*
|
|
* Step 1, split sender shallow commits into "ours" and "theirs"
|
|
* Step 2, clean "ours" based on .git/shallow
|
|
*/
|
|
void prepare_shallow_info(struct shallow_info *info, struct oid_array *sa)
|
|
{
|
|
trace_printf_key(&trace_shallow, "shallow: prepare_shallow_info\n");
|
|
memset(info, 0, sizeof(*info));
|
|
info->shallow = sa;
|
|
if (!sa)
|
|
return;
|
|
ALLOC_ARRAY(info->ours, sa->nr);
|
|
ALLOC_ARRAY(info->theirs, sa->nr);
|
|
for (size_t i = 0; i < sa->nr; i++) {
|
|
if (odb_has_object(the_repository->objects, sa->oid + i,
|
|
HAS_OBJECT_RECHECK_PACKED | HAS_OBJECT_FETCH_PROMISOR)) {
|
|
struct commit_graft *graft;
|
|
graft = lookup_commit_graft(the_repository,
|
|
&sa->oid[i]);
|
|
if (graft && graft->nr_parent < 0)
|
|
continue;
|
|
info->ours[info->nr_ours++] = i;
|
|
} else
|
|
info->theirs[info->nr_theirs++] = i;
|
|
}
|
|
}
|
|
|
|
void clear_shallow_info(struct shallow_info *info)
|
|
{
|
|
if (info->used_shallow) {
|
|
for (size_t i = 0; i < info->shallow->nr; i++)
|
|
free(info->used_shallow[i]);
|
|
free(info->used_shallow);
|
|
}
|
|
|
|
free(info->need_reachability_test);
|
|
free(info->reachable);
|
|
free(info->shallow_ref);
|
|
free(info->ours);
|
|
free(info->theirs);
|
|
}
|
|
|
|
/* Step 4, remove non-existent ones in "theirs" after getting the pack */
|
|
|
|
void remove_nonexistent_theirs_shallow(struct shallow_info *info)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
size_t i, dst;
|
|
trace_printf_key(&trace_shallow, "shallow: remove_nonexistent_theirs_shallow\n");
|
|
for (i = dst = 0; i < info->nr_theirs; i++) {
|
|
if (i != dst)
|
|
info->theirs[dst] = info->theirs[i];
|
|
if (odb_has_object(the_repository->objects, oid + info->theirs[i],
|
|
HAS_OBJECT_RECHECK_PACKED | HAS_OBJECT_FETCH_PROMISOR))
|
|
dst++;
|
|
}
|
|
info->nr_theirs = dst;
|
|
}
|
|
|
|
define_commit_slab(ref_bitmap, uint32_t *);
|
|
|
|
#define POOL_SIZE (512 * 1024)
|
|
|
|
struct paint_info {
|
|
struct ref_bitmap ref_bitmap;
|
|
unsigned nr_bits;
|
|
char **pools;
|
|
char *free, *end;
|
|
unsigned pool_count;
|
|
};
|
|
|
|
static uint32_t *paint_alloc(struct paint_info *info)
|
|
{
|
|
unsigned nr = DIV_ROUND_UP(info->nr_bits, 32);
|
|
unsigned size = nr * sizeof(uint32_t);
|
|
void *p;
|
|
if (!info->pool_count || info->end < info->free + size) {
|
|
if (size > POOL_SIZE)
|
|
BUG("pool size too small for %d in paint_alloc()",
|
|
size);
|
|
info->pool_count++;
|
|
REALLOC_ARRAY(info->pools, info->pool_count);
|
|
info->free = xmalloc(POOL_SIZE);
|
|
info->pools[info->pool_count - 1] = info->free;
|
|
info->end = info->free + POOL_SIZE;
|
|
}
|
|
p = info->free;
|
|
info->free += size;
|
|
return p;
|
|
}
|
|
|
|
/*
|
|
* Given a commit SHA-1, walk down to parents until either SEEN,
|
|
* UNINTERESTING or BOTTOM is hit. Set the id-th bit in ref_bitmap for
|
|
* all walked commits.
|
|
*/
|
|
static void paint_down(struct paint_info *info, const struct object_id *oid,
|
|
unsigned int id)
|
|
{
|
|
unsigned int i, nr;
|
|
struct commit_list *head = NULL;
|
|
size_t bitmap_nr = DIV_ROUND_UP(info->nr_bits, 32);
|
|
size_t bitmap_size = st_mult(sizeof(uint32_t), bitmap_nr);
|
|
struct commit *c = lookup_commit_reference_gently(the_repository, oid,
|
|
1);
|
|
uint32_t *tmp; /* to be freed before return */
|
|
uint32_t *bitmap;
|
|
|
|
if (!c)
|
|
return;
|
|
|
|
tmp = xmalloc(bitmap_size);
|
|
bitmap = paint_alloc(info);
|
|
memset(bitmap, 0, bitmap_size);
|
|
bitmap[id / 32] |= (1U << (id % 32));
|
|
commit_list_insert(c, &head);
|
|
while (head) {
|
|
struct commit_list *p;
|
|
struct commit *c = pop_commit(&head);
|
|
uint32_t **refs = ref_bitmap_at(&info->ref_bitmap, c);
|
|
|
|
/* XXX check "UNINTERESTING" from pack bitmaps if available */
|
|
if (c->object.flags & (SEEN | UNINTERESTING))
|
|
continue;
|
|
else
|
|
c->object.flags |= SEEN;
|
|
|
|
if (!*refs)
|
|
*refs = bitmap;
|
|
else {
|
|
memcpy(tmp, *refs, bitmap_size);
|
|
for (i = 0; i < bitmap_nr; i++)
|
|
tmp[i] |= bitmap[i];
|
|
if (memcmp(tmp, *refs, bitmap_size)) {
|
|
*refs = paint_alloc(info);
|
|
memcpy(*refs, tmp, bitmap_size);
|
|
}
|
|
}
|
|
|
|
if (c->object.flags & BOTTOM)
|
|
continue;
|
|
|
|
if (repo_parse_commit(the_repository, c))
|
|
die("unable to parse commit %s",
|
|
oid_to_hex(&c->object.oid));
|
|
|
|
for (p = c->parents; p; p = p->next) {
|
|
if (p->item->object.flags & SEEN)
|
|
continue;
|
|
commit_list_insert(p->item, &head);
|
|
}
|
|
}
|
|
|
|
nr = get_max_object_index(the_repository);
|
|
for (i = 0; i < nr; i++) {
|
|
struct object *o = get_indexed_object(the_repository, i);
|
|
if (o && o->type == OBJ_COMMIT)
|
|
o->flags &= ~SEEN;
|
|
}
|
|
|
|
free(tmp);
|
|
}
|
|
|
|
static int mark_uninteresting(const struct reference *ref, void *cb_data UNUSED)
|
|
{
|
|
struct commit *commit = lookup_commit_reference_gently(the_repository,
|
|
ref->oid, 1);
|
|
if (!commit)
|
|
return 0;
|
|
commit->object.flags |= UNINTERESTING;
|
|
mark_parents_uninteresting(NULL, commit);
|
|
return 0;
|
|
}
|
|
|
|
static void post_assign_shallow(struct shallow_info *info,
|
|
struct ref_bitmap *ref_bitmap,
|
|
int *ref_status);
|
|
/*
|
|
* Step 6(+7), associate shallow commits with new refs
|
|
*
|
|
* info->ref must be initialized before calling this function.
|
|
*
|
|
* If used is not NULL, it's an array of info->shallow->nr
|
|
* bitmaps. The n-th bit set in the m-th bitmap if ref[n] needs the
|
|
* m-th shallow commit from info->shallow.
|
|
*
|
|
* If used is NULL, "ours" and "theirs" are updated. And if ref_status
|
|
* is not NULL it's an array of ref->nr ints. ref_status[i] is true if
|
|
* the ref needs some shallow commits from either info->ours or
|
|
* info->theirs.
|
|
*/
|
|
void assign_shallow_commits_to_refs(struct shallow_info *info,
|
|
uint32_t **used, int *ref_status)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
struct oid_array *ref = info->ref;
|
|
unsigned int i, nr;
|
|
size_t *shallow, nr_shallow = 0;
|
|
struct paint_info pi;
|
|
|
|
trace_printf_key(&trace_shallow, "shallow: assign_shallow_commits_to_refs\n");
|
|
ALLOC_ARRAY(shallow, info->nr_ours + info->nr_theirs);
|
|
for (i = 0; i < info->nr_ours; i++)
|
|
shallow[nr_shallow++] = info->ours[i];
|
|
for (i = 0; i < info->nr_theirs; i++)
|
|
shallow[nr_shallow++] = info->theirs[i];
|
|
|
|
/*
|
|
* Prepare the commit graph to track what refs can reach what
|
|
* (new) shallow commits.
|
|
*/
|
|
nr = get_max_object_index(the_repository);
|
|
for (i = 0; i < nr; i++) {
|
|
struct object *o = get_indexed_object(the_repository, i);
|
|
if (!o || o->type != OBJ_COMMIT)
|
|
continue;
|
|
|
|
o->flags &= ~(UNINTERESTING | BOTTOM | SEEN);
|
|
}
|
|
|
|
memset(&pi, 0, sizeof(pi));
|
|
init_ref_bitmap(&pi.ref_bitmap);
|
|
pi.nr_bits = ref->nr;
|
|
|
|
/*
|
|
* "--not --all" to cut short the traversal if new refs
|
|
* connect to old refs. If not (e.g. force ref updates) it'll
|
|
* have to go down to the current shallow commits.
|
|
*/
|
|
refs_head_ref(get_main_ref_store(the_repository), mark_uninteresting,
|
|
NULL);
|
|
refs_for_each_ref(get_main_ref_store(the_repository),
|
|
mark_uninteresting, NULL);
|
|
|
|
/* Mark potential bottoms so we won't go out of bound */
|
|
for (i = 0; i < nr_shallow; i++) {
|
|
struct commit *c = lookup_commit(the_repository,
|
|
&oid[shallow[i]]);
|
|
c->object.flags |= BOTTOM;
|
|
}
|
|
|
|
for (i = 0; i < ref->nr; i++)
|
|
paint_down(&pi, ref->oid + i, i);
|
|
|
|
if (used) {
|
|
int bitmap_size = DIV_ROUND_UP(pi.nr_bits, 32) * sizeof(uint32_t);
|
|
MEMZERO_ARRAY(used, info->shallow->nr);
|
|
for (i = 0; i < nr_shallow; i++) {
|
|
const struct commit *c = lookup_commit(the_repository,
|
|
&oid[shallow[i]]);
|
|
uint32_t **map = ref_bitmap_at(&pi.ref_bitmap, c);
|
|
if (*map)
|
|
used[shallow[i]] = xmemdupz(*map, bitmap_size);
|
|
}
|
|
/*
|
|
* unreachable shallow commits are not removed from
|
|
* "ours" and "theirs". The user is supposed to run
|
|
* step 7 on every ref separately and not trust "ours"
|
|
* and "theirs" any more.
|
|
*/
|
|
} else
|
|
post_assign_shallow(info, &pi.ref_bitmap, ref_status);
|
|
|
|
clear_ref_bitmap(&pi.ref_bitmap);
|
|
for (i = 0; i < pi.pool_count; i++)
|
|
free(pi.pools[i]);
|
|
free(pi.pools);
|
|
free(shallow);
|
|
}
|
|
|
|
struct commit_array {
|
|
struct commit **commits;
|
|
size_t nr, alloc;
|
|
};
|
|
|
|
static int add_ref(const struct reference *ref, void *cb_data)
|
|
{
|
|
struct commit_array *ca = cb_data;
|
|
ALLOC_GROW(ca->commits, ca->nr + 1, ca->alloc);
|
|
ca->commits[ca->nr] = lookup_commit_reference_gently(the_repository,
|
|
ref->oid, 1);
|
|
if (ca->commits[ca->nr])
|
|
ca->nr++;
|
|
return 0;
|
|
}
|
|
|
|
static void update_refstatus(int *ref_status, size_t nr, uint32_t *bitmap)
|
|
{
|
|
if (!ref_status)
|
|
return;
|
|
for (size_t i = 0; i < nr; i++)
|
|
if (bitmap[i / 32] & (1U << (i % 32)))
|
|
ref_status[i]++;
|
|
}
|
|
|
|
/*
|
|
* Step 7, reachability test on "ours" at commit level
|
|
*/
|
|
static void post_assign_shallow(struct shallow_info *info,
|
|
struct ref_bitmap *ref_bitmap,
|
|
int *ref_status)
|
|
{
|
|
struct object_id *oid = info->shallow->oid;
|
|
struct commit *c;
|
|
uint32_t **bitmap;
|
|
size_t dst, i, j;
|
|
size_t bitmap_nr = DIV_ROUND_UP(info->ref->nr, 32);
|
|
struct commit_array ca;
|
|
|
|
trace_printf_key(&trace_shallow, "shallow: post_assign_shallow\n");
|
|
if (ref_status)
|
|
MEMZERO_ARRAY(ref_status, info->ref->nr);
|
|
|
|
/* Remove unreachable shallow commits from "theirs" */
|
|
for (i = dst = 0; i < info->nr_theirs; i++) {
|
|
if (i != dst)
|
|
info->theirs[dst] = info->theirs[i];
|
|
c = lookup_commit(the_repository, &oid[info->theirs[i]]);
|
|
bitmap = ref_bitmap_at(ref_bitmap, c);
|
|
if (!*bitmap)
|
|
continue;
|
|
for (j = 0; j < bitmap_nr; j++)
|
|
if (bitmap[0][j]) {
|
|
update_refstatus(ref_status, info->ref->nr, *bitmap);
|
|
dst++;
|
|
break;
|
|
}
|
|
}
|
|
info->nr_theirs = dst;
|
|
|
|
memset(&ca, 0, sizeof(ca));
|
|
refs_head_ref(get_main_ref_store(the_repository), add_ref, &ca);
|
|
refs_for_each_ref(get_main_ref_store(the_repository), add_ref, &ca);
|
|
|
|
/* Remove unreachable shallow commits from "ours" */
|
|
for (i = dst = 0; i < info->nr_ours; i++) {
|
|
if (i != dst)
|
|
info->ours[dst] = info->ours[i];
|
|
c = lookup_commit(the_repository, &oid[info->ours[i]]);
|
|
bitmap = ref_bitmap_at(ref_bitmap, c);
|
|
if (!*bitmap)
|
|
continue;
|
|
for (j = 0; j < bitmap_nr; j++)
|
|
if (bitmap[0][j]) {
|
|
/* Step 7, reachability test at commit level */
|
|
int ret = repo_in_merge_bases_many(the_repository, c, ca.nr, ca.commits, 1);
|
|
if (ret < 0)
|
|
exit(128);
|
|
if (!ret) {
|
|
update_refstatus(ref_status, info->ref->nr, *bitmap);
|
|
dst++;
|
|
break;
|
|
}
|
|
}
|
|
}
|
|
info->nr_ours = dst;
|
|
|
|
free(ca.commits);
|
|
}
|
|
|
|
/* (Delayed) step 7, reachability test at commit level */
|
|
int delayed_reachability_test(struct shallow_info *si, int c)
|
|
{
|
|
if (si->need_reachability_test[c]) {
|
|
struct commit *commit = lookup_commit(the_repository,
|
|
&si->shallow->oid[c]);
|
|
|
|
if (!si->commits) {
|
|
struct commit_array ca;
|
|
|
|
memset(&ca, 0, sizeof(ca));
|
|
refs_head_ref(get_main_ref_store(the_repository),
|
|
add_ref, &ca);
|
|
refs_for_each_ref(get_main_ref_store(the_repository),
|
|
add_ref, &ca);
|
|
si->commits = ca.commits;
|
|
si->nr_commits = ca.nr;
|
|
}
|
|
|
|
si->reachable[c] = repo_in_merge_bases_many(the_repository,
|
|
commit,
|
|
si->nr_commits,
|
|
si->commits,
|
|
1);
|
|
if (si->reachable[c] < 0)
|
|
exit(128);
|
|
si->need_reachability_test[c] = 0;
|
|
}
|
|
return si->reachable[c];
|
|
}
|