2 #include "repository.h"
5 #include "object-store.h"
11 #include "sha1-array.h"
14 #include "commit-slab.h"
16 #include "list-objects.h"
17 #include "commit-slab.h"
18 #include "repository.h"
20 void set_alternate_shallow_file(struct repository *r, const char *path, int override)
22 if (r->parsed_objects->is_shallow != -1)
23 BUG("is_repository_shallow must not be called before set_alternate_shallow_file");
24 if (r->parsed_objects->alternate_shallow_file && !override)
26 free(r->parsed_objects->alternate_shallow_file);
27 r->parsed_objects->alternate_shallow_file = xstrdup_or_null(path);
30 int register_shallow(struct repository *r, const struct object_id *oid)
32 struct commit_graft *graft =
33 xmalloc(sizeof(struct commit_graft));
34 struct commit *commit = lookup_commit(oid);
36 oidcpy(&graft->oid, oid);
37 graft->nr_parent = -1;
38 if (commit && commit->object.parsed)
39 commit->parents = NULL;
40 return register_commit_graft(r, graft, 0);
43 int is_repository_shallow(struct repository *r)
47 const char *path = r->parsed_objects->alternate_shallow_file;
49 if (r->parsed_objects->is_shallow >= 0)
50 return r->parsed_objects->is_shallow;
53 path = git_path_shallow(r);
55 * fetch-pack sets '--shallow-file ""' as an indicator that no
56 * shallow file should be used. We could just open it and it
57 * will likely fail. But let's do an explicit check instead.
59 if (!*path || (fp = fopen(path, "r")) == NULL) {
60 stat_validity_clear(r->parsed_objects->shallow_stat);
61 r->parsed_objects->is_shallow = 0;
62 return r->parsed_objects->is_shallow;
64 stat_validity_update(r->parsed_objects->shallow_stat, fileno(fp));
65 r->parsed_objects->is_shallow = 1;
67 while (fgets(buf, sizeof(buf), fp)) {
69 if (get_oid_hex(buf, &oid))
70 die("bad shallow line: %s", buf);
71 register_shallow(r, &oid);
74 return r->parsed_objects->is_shallow;
78 * TODO: use "int" elemtype instead of "int *" when/if commit-slab
79 * supports a "valid" flag.
81 define_commit_slab(commit_depth, int *);
82 struct commit_list *get_shallow_commits(struct object_array *heads, int depth,
83 int shallow_flag, int not_shallow_flag)
85 int i = 0, cur_depth = 0;
86 struct commit_list *result = NULL;
87 struct object_array stack = OBJECT_ARRAY_INIT;
88 struct commit *commit = NULL;
89 struct commit_graft *graft;
90 struct commit_depth depths;
92 init_commit_depth(&depths);
93 while (commit || i < heads->nr || stack.nr) {
94 struct commit_list *p;
98 commit = (struct commit *)
99 deref_tag(heads->objects[i++].item, NULL, 0);
100 if (!commit || commit->object.type != OBJ_COMMIT) {
104 depth_slot = commit_depth_at(&depths, commit);
106 *depth_slot = xmalloc(sizeof(int));
110 commit = (struct commit *)
111 object_array_pop(&stack);
112 cur_depth = **commit_depth_at(&depths, commit);
115 parse_commit_or_die(commit);
117 if ((depth != INFINITE_DEPTH && cur_depth >= depth) ||
118 (is_repository_shallow(the_repository) && !commit->parents &&
119 (graft = lookup_commit_graft(the_repository, &commit->object.oid)) != NULL &&
120 graft->nr_parent < 0)) {
121 commit_list_insert(commit, &result);
122 commit->object.flags |= shallow_flag;
126 commit->object.flags |= not_shallow_flag;
127 for (p = commit->parents, commit = NULL; p; p = p->next) {
128 int **depth_slot = commit_depth_at(&depths, p->item);
130 *depth_slot = xmalloc(sizeof(int));
131 **depth_slot = cur_depth;
133 if (cur_depth >= **depth_slot)
135 **depth_slot = cur_depth;
138 add_object_array(&p->item->object,
142 cur_depth = **commit_depth_at(&depths, commit);
146 for (i = 0; i < depths.slab_count; i++) {
149 for (j = 0; j < depths.slab_size; j++)
150 free(depths.slab[i][j]);
152 clear_commit_depth(&depths);
157 static void show_commit(struct commit *commit, void *data)
159 commit_list_insert(commit, data);
163 * Given rev-list arguments, run rev-list. All reachable commits
164 * except border ones are marked with not_shallow_flag. Border commits
165 * are marked with shallow_flag. The list of border/shallow commits
168 struct commit_list *get_shallow_commits_by_rev_list(int ac, const char **av,
170 int not_shallow_flag)
172 struct commit_list *result = NULL, *p;
173 struct commit_list *not_shallow_list = NULL;
174 struct rev_info revs;
175 int both_flags = shallow_flag | not_shallow_flag;
178 * SHALLOW (excluded) and NOT_SHALLOW (included) should not be
179 * set at this point. But better be safe than sorry.
181 clear_object_flags(both_flags);
183 is_repository_shallow(the_repository); /* make sure shallows are read */
185 init_revisions(&revs, NULL);
186 save_commit_buffer = 0;
187 setup_revisions(ac, av, &revs, NULL);
189 if (prepare_revision_walk(&revs))
190 die("revision walk setup failed");
191 traverse_commit_list(&revs, show_commit, NULL, ¬_shallow_list);
193 if (!not_shallow_list)
194 die("no commits selected for shallow requests");
196 /* Mark all reachable commits as NOT_SHALLOW */
197 for (p = not_shallow_list; p; p = p->next)
198 p->item->object.flags |= not_shallow_flag;
201 * mark border commits SHALLOW + NOT_SHALLOW.
202 * We cannot clear NOT_SHALLOW right now. Imagine border
203 * commit A is processed first, then commit B, whose parent is
204 * A, later. If NOT_SHALLOW on A is cleared at step 1, B
205 * itself is considered border at step 2, which is incorrect.
207 for (p = not_shallow_list; p; p = p->next) {
208 struct commit *c = p->item;
209 struct commit_list *parent;
212 die("unable to parse commit %s",
213 oid_to_hex(&c->object.oid));
215 for (parent = c->parents; parent; parent = parent->next)
216 if (!(parent->item->object.flags & not_shallow_flag)) {
217 c->object.flags |= shallow_flag;
218 commit_list_insert(c, &result);
222 free_commit_list(not_shallow_list);
225 * Now we can clean up NOT_SHALLOW on border commits. Having
226 * both flags set can confuse the caller.
228 for (p = result; p; p = p->next) {
229 struct object *o = &p->item->object;
230 if ((o->flags & both_flags) == both_flags)
231 o->flags &= ~not_shallow_flag;
236 static void check_shallow_file_for_update(struct repository *r)
238 if (r->parsed_objects->is_shallow == -1)
239 BUG("shallow must be initialized by now");
241 if (!stat_validity_check(r->parsed_objects->shallow_stat, git_path_shallow(the_repository)))
242 die("shallow file has changed since we read it");
248 struct write_shallow_data {
250 int use_pack_protocol;
255 static int write_one_shallow(const struct commit_graft *graft, void *cb_data)
257 struct write_shallow_data *data = cb_data;
258 const char *hex = oid_to_hex(&graft->oid);
259 if (graft->nr_parent != -1)
261 if (data->flags & SEEN_ONLY) {
262 struct commit *c = lookup_commit(&graft->oid);
263 if (!c || !(c->object.flags & SEEN)) {
264 if (data->flags & VERBOSE)
265 printf("Removing %s from .git/shallow\n",
266 oid_to_hex(&c->object.oid));
271 if (data->use_pack_protocol)
272 packet_buf_write(data->out, "shallow %s", hex);
274 strbuf_addstr(data->out, hex);
275 strbuf_addch(data->out, '\n');
280 static int write_shallow_commits_1(struct strbuf *out, int use_pack_protocol,
281 const struct oid_array *extra,
284 struct write_shallow_data data;
287 data.use_pack_protocol = use_pack_protocol;
290 for_each_commit_graft(write_one_shallow, &data);
293 for (i = 0; i < extra->nr; i++) {
294 strbuf_addstr(out, oid_to_hex(extra->oid + i));
295 strbuf_addch(out, '\n');
301 int write_shallow_commits(struct strbuf *out, int use_pack_protocol,
302 const struct oid_array *extra)
304 return write_shallow_commits_1(out, use_pack_protocol, extra, 0);
307 const char *setup_temporary_shallow(const struct oid_array *extra)
309 struct tempfile *temp;
310 struct strbuf sb = STRBUF_INIT;
312 if (write_shallow_commits(&sb, 0, extra)) {
313 temp = xmks_tempfile(git_path("shallow_XXXXXX"));
315 if (write_in_full(temp->fd, sb.buf, sb.len) < 0 ||
316 close_tempfile_gently(temp) < 0)
317 die_errno("failed to write to %s",
318 get_tempfile_path(temp));
320 return get_tempfile_path(temp);
323 * is_repository_shallow() sees empty string as "no shallow
329 void setup_alternate_shallow(struct lock_file *shallow_lock,
330 const char **alternate_shallow_file,
331 const struct oid_array *extra)
333 struct strbuf sb = STRBUF_INIT;
336 fd = hold_lock_file_for_update(shallow_lock,
337 git_path_shallow(the_repository),
339 check_shallow_file_for_update(the_repository);
340 if (write_shallow_commits(&sb, 0, extra)) {
341 if (write_in_full(fd, sb.buf, sb.len) < 0)
342 die_errno("failed to write to %s",
343 get_lock_file_path(shallow_lock));
344 *alternate_shallow_file = get_lock_file_path(shallow_lock);
347 * is_repository_shallow() sees empty string as "no
350 *alternate_shallow_file = "";
354 static int advertise_shallow_grafts_cb(const struct commit_graft *graft, void *cb)
357 if (graft->nr_parent == -1)
358 packet_write_fmt(fd, "shallow %s\n", oid_to_hex(&graft->oid));
362 void advertise_shallow_grafts(int fd)
364 if (!is_repository_shallow(the_repository))
366 for_each_commit_graft(advertise_shallow_grafts_cb, &fd);
370 * mark_reachable_objects() should have been run prior to this and all
371 * reachable commits marked as "SEEN".
373 void prune_shallow(int show_only)
375 struct lock_file shallow_lock = LOCK_INIT;
376 struct strbuf sb = STRBUF_INIT;
380 write_shallow_commits_1(&sb, 0, NULL, SEEN_ONLY | VERBOSE);
384 fd = hold_lock_file_for_update(&shallow_lock,
385 git_path_shallow(the_repository),
387 check_shallow_file_for_update(the_repository);
388 if (write_shallow_commits_1(&sb, 0, NULL, SEEN_ONLY)) {
389 if (write_in_full(fd, sb.buf, sb.len) < 0)
390 die_errno("failed to write to %s",
391 get_lock_file_path(&shallow_lock));
392 commit_lock_file(&shallow_lock);
394 unlink(git_path_shallow(the_repository));
395 rollback_lock_file(&shallow_lock);
400 struct trace_key trace_shallow = TRACE_KEY_INIT(SHALLOW);
403 * Step 1, split sender shallow commits into "ours" and "theirs"
404 * Step 2, clean "ours" based on .git/shallow
406 void prepare_shallow_info(struct shallow_info *info, struct oid_array *sa)
409 trace_printf_key(&trace_shallow, "shallow: prepare_shallow_info\n");
410 memset(info, 0, sizeof(*info));
414 ALLOC_ARRAY(info->ours, sa->nr);
415 ALLOC_ARRAY(info->theirs, sa->nr);
416 for (i = 0; i < sa->nr; i++) {
417 if (has_object_file(sa->oid + i)) {
418 struct commit_graft *graft;
419 graft = lookup_commit_graft(the_repository,
421 if (graft && graft->nr_parent < 0)
423 info->ours[info->nr_ours++] = i;
425 info->theirs[info->nr_theirs++] = i;
429 void clear_shallow_info(struct shallow_info *info)
435 /* Step 4, remove non-existent ones in "theirs" after getting the pack */
437 void remove_nonexistent_theirs_shallow(struct shallow_info *info)
439 struct object_id *oid = info->shallow->oid;
441 trace_printf_key(&trace_shallow, "shallow: remove_nonexistent_theirs_shallow\n");
442 for (i = dst = 0; i < info->nr_theirs; i++) {
444 info->theirs[dst] = info->theirs[i];
445 if (has_object_file(oid + info->theirs[i]))
448 info->nr_theirs = dst;
451 define_commit_slab(ref_bitmap, uint32_t *);
453 #define POOL_SIZE (512 * 1024)
456 struct ref_bitmap ref_bitmap;
463 static uint32_t *paint_alloc(struct paint_info *info)
465 unsigned nr = DIV_ROUND_UP(info->nr_bits, 32);
466 unsigned size = nr * sizeof(uint32_t);
468 if (!info->pool_count || size > info->end - info->free) {
469 if (size > POOL_SIZE)
470 BUG("pool size too small for %d in paint_alloc()",
473 REALLOC_ARRAY(info->pools, info->pool_count);
474 info->free = xmalloc(POOL_SIZE);
475 info->pools[info->pool_count - 1] = info->free;
476 info->end = info->free + POOL_SIZE;
484 * Given a commit SHA-1, walk down to parents until either SEEN,
485 * UNINTERESTING or BOTTOM is hit. Set the id-th bit in ref_bitmap for
486 * all walked commits.
488 static void paint_down(struct paint_info *info, const struct object_id *oid,
492 struct commit_list *head = NULL;
493 int bitmap_nr = DIV_ROUND_UP(info->nr_bits, 32);
494 size_t bitmap_size = st_mult(sizeof(uint32_t), bitmap_nr);
495 struct commit *c = lookup_commit_reference_gently(oid, 1);
496 uint32_t *tmp; /* to be freed before return */
502 tmp = xmalloc(bitmap_size);
503 bitmap = paint_alloc(info);
504 memset(bitmap, 0, bitmap_size);
505 bitmap[id / 32] |= (1U << (id % 32));
506 commit_list_insert(c, &head);
508 struct commit_list *p;
509 struct commit *c = pop_commit(&head);
510 uint32_t **refs = ref_bitmap_at(&info->ref_bitmap, c);
512 /* XXX check "UNINTERESTING" from pack bitmaps if available */
513 if (c->object.flags & (SEEN | UNINTERESTING))
516 c->object.flags |= SEEN;
521 memcpy(tmp, *refs, bitmap_size);
522 for (i = 0; i < bitmap_nr; i++)
524 if (memcmp(tmp, *refs, bitmap_size)) {
525 *refs = paint_alloc(info);
526 memcpy(*refs, tmp, bitmap_size);
530 if (c->object.flags & BOTTOM)
534 die("unable to parse commit %s",
535 oid_to_hex(&c->object.oid));
537 for (p = c->parents; p; p = p->next) {
538 if (p->item->object.flags & SEEN)
540 commit_list_insert(p->item, &head);
544 nr = get_max_object_index();
545 for (i = 0; i < nr; i++) {
546 struct object *o = get_indexed_object(i);
547 if (o && o->type == OBJ_COMMIT)
554 static int mark_uninteresting(const char *refname, const struct object_id *oid,
555 int flags, void *cb_data)
557 struct commit *commit = lookup_commit_reference_gently(oid, 1);
560 commit->object.flags |= UNINTERESTING;
561 mark_parents_uninteresting(commit);
565 static void post_assign_shallow(struct shallow_info *info,
566 struct ref_bitmap *ref_bitmap,
569 * Step 6(+7), associate shallow commits with new refs
571 * info->ref must be initialized before calling this function.
573 * If used is not NULL, it's an array of info->shallow->nr
574 * bitmaps. The n-th bit set in the m-th bitmap if ref[n] needs the
575 * m-th shallow commit from info->shallow.
577 * If used is NULL, "ours" and "theirs" are updated. And if ref_status
578 * is not NULL it's an array of ref->nr ints. ref_status[i] is true if
579 * the ref needs some shallow commits from either info->ours or
582 void assign_shallow_commits_to_refs(struct shallow_info *info,
583 uint32_t **used, int *ref_status)
585 struct object_id *oid = info->shallow->oid;
586 struct oid_array *ref = info->ref;
588 int *shallow, nr_shallow = 0;
589 struct paint_info pi;
591 trace_printf_key(&trace_shallow, "shallow: assign_shallow_commits_to_refs\n");
592 ALLOC_ARRAY(shallow, info->nr_ours + info->nr_theirs);
593 for (i = 0; i < info->nr_ours; i++)
594 shallow[nr_shallow++] = info->ours[i];
595 for (i = 0; i < info->nr_theirs; i++)
596 shallow[nr_shallow++] = info->theirs[i];
599 * Prepare the commit graph to track what refs can reach what
600 * (new) shallow commits.
602 nr = get_max_object_index();
603 for (i = 0; i < nr; i++) {
604 struct object *o = get_indexed_object(i);
605 if (!o || o->type != OBJ_COMMIT)
608 o->flags &= ~(UNINTERESTING | BOTTOM | SEEN);
611 memset(&pi, 0, sizeof(pi));
612 init_ref_bitmap(&pi.ref_bitmap);
613 pi.nr_bits = ref->nr;
616 * "--not --all" to cut short the traversal if new refs
617 * connect to old refs. If not (e.g. force ref updates) it'll
618 * have to go down to the current shallow commits.
620 head_ref(mark_uninteresting, NULL);
621 for_each_ref(mark_uninteresting, NULL);
623 /* Mark potential bottoms so we won't go out of bound */
624 for (i = 0; i < nr_shallow; i++) {
625 struct commit *c = lookup_commit(&oid[shallow[i]]);
626 c->object.flags |= BOTTOM;
629 for (i = 0; i < ref->nr; i++)
630 paint_down(&pi, ref->oid + i, i);
633 int bitmap_size = DIV_ROUND_UP(pi.nr_bits, 32) * sizeof(uint32_t);
634 memset(used, 0, sizeof(*used) * info->shallow->nr);
635 for (i = 0; i < nr_shallow; i++) {
636 const struct commit *c = lookup_commit(&oid[shallow[i]]);
637 uint32_t **map = ref_bitmap_at(&pi.ref_bitmap, c);
639 used[shallow[i]] = xmemdupz(*map, bitmap_size);
642 * unreachable shallow commits are not removed from
643 * "ours" and "theirs". The user is supposed to run
644 * step 7 on every ref separately and not trust "ours"
645 * and "theirs" any more.
648 post_assign_shallow(info, &pi.ref_bitmap, ref_status);
650 clear_ref_bitmap(&pi.ref_bitmap);
651 for (i = 0; i < pi.pool_count; i++)
657 struct commit_array {
658 struct commit **commits;
662 static int add_ref(const char *refname, const struct object_id *oid,
663 int flags, void *cb_data)
665 struct commit_array *ca = cb_data;
666 ALLOC_GROW(ca->commits, ca->nr + 1, ca->alloc);
667 ca->commits[ca->nr] = lookup_commit_reference_gently(oid, 1);
668 if (ca->commits[ca->nr])
673 static void update_refstatus(int *ref_status, int nr, uint32_t *bitmap)
678 for (i = 0; i < nr; i++)
679 if (bitmap[i / 32] & (1U << (i % 32)))
684 * Step 7, reachability test on "ours" at commit level
686 static void post_assign_shallow(struct shallow_info *info,
687 struct ref_bitmap *ref_bitmap,
690 struct object_id *oid = info->shallow->oid;
694 int bitmap_nr = DIV_ROUND_UP(info->ref->nr, 32);
695 struct commit_array ca;
697 trace_printf_key(&trace_shallow, "shallow: post_assign_shallow\n");
699 memset(ref_status, 0, sizeof(*ref_status) * info->ref->nr);
701 /* Remove unreachable shallow commits from "theirs" */
702 for (i = dst = 0; i < info->nr_theirs; i++) {
704 info->theirs[dst] = info->theirs[i];
705 c = lookup_commit(&oid[info->theirs[i]]);
706 bitmap = ref_bitmap_at(ref_bitmap, c);
709 for (j = 0; j < bitmap_nr; j++)
711 update_refstatus(ref_status, info->ref->nr, *bitmap);
716 info->nr_theirs = dst;
718 memset(&ca, 0, sizeof(ca));
719 head_ref(add_ref, &ca);
720 for_each_ref(add_ref, &ca);
722 /* Remove unreachable shallow commits from "ours" */
723 for (i = dst = 0; i < info->nr_ours; i++) {
725 info->ours[dst] = info->ours[i];
726 c = lookup_commit(&oid[info->ours[i]]);
727 bitmap = ref_bitmap_at(ref_bitmap, c);
730 for (j = 0; j < bitmap_nr; j++)
732 /* Step 7, reachability test at commit level */
733 !in_merge_bases_many(c, ca.nr, ca.commits)) {
734 update_refstatus(ref_status, info->ref->nr, *bitmap);
744 /* (Delayed) step 7, reachability test at commit level */
745 int delayed_reachability_test(struct shallow_info *si, int c)
747 if (si->need_reachability_test[c]) {
748 struct commit *commit = lookup_commit(&si->shallow->oid[c]);
751 struct commit_array ca;
753 memset(&ca, 0, sizeof(ca));
754 head_ref(add_ref, &ca);
755 for_each_ref(add_ref, &ca);
756 si->commits = ca.commits;
757 si->nr_commits = ca.nr;
760 si->reachable[c] = in_merge_bases_many(commit,
763 si->need_reachability_test[c] = 0;
765 return si->reachable[c];