3 #include "refs-internal.h"
5 #include "../iterator.h"
6 #include "../dir-iterator.h"
7 #include "../lockfile.h"
14 struct object_id old_oid;
18 * Return true if refname, which has the specified oid and flags, can
19 * be resolved to an object in the database. If the referred-to object
20 * does not exist, emit a warning and return false.
22 static int ref_resolves_to_object(const char *refname,
23 const struct object_id *oid,
26 if (flags & REF_ISBROKEN)
28 if (!has_sha1_file(oid->hash)) {
29 error("%s does not point to a valid object!", refname);
35 struct packed_ref_cache {
36 struct ref_cache *cache;
39 * Count of references to the data structure in this instance,
40 * including the pointer from files_ref_store::packed if any.
41 * The data will not be freed as long as the reference count
44 unsigned int referrers;
46 /* The metadata from when this packed-refs cache was read */
47 struct stat_validity validity;
51 * A container for `packed-refs`-related data. It is not (yet) a
54 struct packed_ref_store {
55 unsigned int store_flags;
57 /* The path of the "packed-refs" file: */
61 * A cache of the values read from the `packed-refs` file, if
62 * it might still be current; otherwise, NULL.
64 struct packed_ref_cache *cache;
67 * Lock used for the "packed-refs" file. Note that this (and
68 * thus the enclosing `packed_ref_store`) must not be freed.
70 struct lock_file lock;
73 static struct packed_ref_store *packed_ref_store_create(
74 const char *path, unsigned int store_flags)
76 struct packed_ref_store *refs = xcalloc(1, sizeof(*refs));
78 refs->store_flags = store_flags;
79 refs->path = xstrdup(path);
84 * Die if refs is not the main ref store. caller is used in any
85 * necessary error messages.
87 static void packed_assert_main_repository(struct packed_ref_store *refs,
90 if (refs->store_flags & REF_STORE_MAIN)
93 die("BUG: operation %s only allowed for main ref store", caller);
97 * Future: need to be in "struct repository"
98 * when doing a full libification.
100 struct files_ref_store {
101 struct ref_store base;
102 unsigned int store_flags;
107 struct ref_cache *loose;
109 struct packed_ref_store *packed_ref_store;
113 * Increment the reference count of *packed_refs.
115 static void acquire_packed_ref_cache(struct packed_ref_cache *packed_refs)
117 packed_refs->referrers++;
121 * Decrease the reference count of *packed_refs. If it goes to zero,
122 * free *packed_refs and return true; otherwise return false.
124 static int release_packed_ref_cache(struct packed_ref_cache *packed_refs)
126 if (!--packed_refs->referrers) {
127 free_ref_cache(packed_refs->cache);
128 stat_validity_clear(&packed_refs->validity);
136 static void clear_packed_ref_cache(struct packed_ref_store *refs)
139 struct packed_ref_cache *cache = refs->cache;
141 if (is_lock_file_locked(&refs->lock))
142 die("BUG: packed-ref cache cleared while locked");
144 release_packed_ref_cache(cache);
148 static void clear_loose_ref_cache(struct files_ref_store *refs)
151 free_ref_cache(refs->loose);
157 * Create a new submodule ref cache and add it to the internal
160 static struct ref_store *files_ref_store_create(const char *gitdir,
163 struct files_ref_store *refs = xcalloc(1, sizeof(*refs));
164 struct ref_store *ref_store = (struct ref_store *)refs;
165 struct strbuf sb = STRBUF_INIT;
167 base_ref_store_init(ref_store, &refs_be_files);
168 refs->store_flags = flags;
170 refs->gitdir = xstrdup(gitdir);
171 get_common_dir_noenv(&sb, gitdir);
172 refs->gitcommondir = strbuf_detach(&sb, NULL);
173 strbuf_addf(&sb, "%s/packed-refs", refs->gitcommondir);
174 refs->packed_ref_store = packed_ref_store_create(sb.buf, flags);
181 * Die if refs is not the main ref store. caller is used in any
182 * necessary error messages.
184 static void files_assert_main_repository(struct files_ref_store *refs,
187 if (refs->store_flags & REF_STORE_MAIN)
190 die("BUG: operation %s only allowed for main ref store", caller);
194 * Downcast ref_store to files_ref_store. Die if ref_store is not a
195 * files_ref_store. required_flags is compared with ref_store's
196 * store_flags to ensure the ref_store has all required capabilities.
197 * "caller" is used in any necessary error messages.
199 static struct files_ref_store *files_downcast(struct ref_store *ref_store,
200 unsigned int required_flags,
203 struct files_ref_store *refs;
205 if (ref_store->be != &refs_be_files)
206 die("BUG: ref_store is type \"%s\" not \"files\" in %s",
207 ref_store->be->name, caller);
209 refs = (struct files_ref_store *)ref_store;
211 if ((refs->store_flags & required_flags) != required_flags)
212 die("BUG: operation %s requires abilities 0x%x, but only have 0x%x",
213 caller, required_flags, refs->store_flags);
218 /* The length of a peeled reference line in packed-refs, including EOL: */
219 #define PEELED_LINE_LENGTH 42
222 * The packed-refs header line that we write out. Perhaps other
223 * traits will be added later. The trailing space is required.
225 static const char PACKED_REFS_HEADER[] =
226 "# pack-refs with: peeled fully-peeled \n";
229 * Parse one line from a packed-refs file. Write the SHA1 to sha1.
230 * Return a pointer to the refname within the line (null-terminated),
231 * or NULL if there was a problem.
233 static const char *parse_ref_line(struct strbuf *line, struct object_id *oid)
237 if (parse_oid_hex(line->buf, oid, &ref) < 0)
239 if (!isspace(*ref++))
245 if (line->buf[line->len - 1] != '\n')
247 line->buf[--line->len] = 0;
253 * Read from `packed_refs_file` into a newly-allocated
254 * `packed_ref_cache` and return it. The return value will already
255 * have its reference count incremented.
257 * A comment line of the form "# pack-refs with: " may contain zero or
258 * more traits. We interpret the traits as follows:
262 * Probably no references are peeled. But if the file contains a
263 * peeled value for a reference, we will use it.
267 * References under "refs/tags/", if they *can* be peeled, *are*
268 * peeled in this file. References outside of "refs/tags/" are
269 * probably not peeled even if they could have been, but if we find
270 * a peeled value for such a reference we will use it.
274 * All references in the file that can be peeled are peeled.
275 * Inversely (and this is more important), any references in the
276 * file for which no peeled value is recorded is not peelable. This
277 * trait should typically be written alongside "peeled" for
278 * compatibility with older clients, but we do not require it
279 * (i.e., "peeled" is a no-op if "fully-peeled" is set).
281 static struct packed_ref_cache *read_packed_refs(const char *packed_refs_file)
284 struct packed_ref_cache *packed_refs = xcalloc(1, sizeof(*packed_refs));
285 struct ref_entry *last = NULL;
286 struct strbuf line = STRBUF_INIT;
287 enum { PEELED_NONE, PEELED_TAGS, PEELED_FULLY } peeled = PEELED_NONE;
290 acquire_packed_ref_cache(packed_refs);
291 packed_refs->cache = create_ref_cache(NULL, NULL);
292 packed_refs->cache->root->flag &= ~REF_INCOMPLETE;
294 f = fopen(packed_refs_file, "r");
296 if (errno == ENOENT) {
298 * This is OK; it just means that no
299 * "packed-refs" file has been written yet,
300 * which is equivalent to it being empty.
304 die_errno("couldn't read %s", packed_refs_file);
308 stat_validity_update(&packed_refs->validity, fileno(f));
310 dir = get_ref_dir(packed_refs->cache->root);
311 while (strbuf_getwholeline(&line, f, '\n') != EOF) {
312 struct object_id oid;
316 if (skip_prefix(line.buf, "# pack-refs with:", &traits)) {
317 if (strstr(traits, " fully-peeled "))
318 peeled = PEELED_FULLY;
319 else if (strstr(traits, " peeled "))
320 peeled = PEELED_TAGS;
321 /* perhaps other traits later as well */
325 refname = parse_ref_line(&line, &oid);
327 int flag = REF_ISPACKED;
329 if (check_refname_format(refname, REFNAME_ALLOW_ONELEVEL)) {
330 if (!refname_is_safe(refname))
331 die("packed refname is dangerous: %s", refname);
333 flag |= REF_BAD_NAME | REF_ISBROKEN;
335 last = create_ref_entry(refname, &oid, flag);
336 if (peeled == PEELED_FULLY ||
337 (peeled == PEELED_TAGS && starts_with(refname, "refs/tags/")))
338 last->flag |= REF_KNOWS_PEELED;
339 add_ref_entry(dir, last);
343 line.buf[0] == '^' &&
344 line.len == PEELED_LINE_LENGTH &&
345 line.buf[PEELED_LINE_LENGTH - 1] == '\n' &&
346 !get_oid_hex(line.buf + 1, &oid)) {
347 oidcpy(&last->u.value.peeled, &oid);
349 * Regardless of what the file header said,
350 * we definitely know the value of *this*
353 last->flag |= REF_KNOWS_PEELED;
358 strbuf_release(&line);
363 static void files_reflog_path(struct files_ref_store *refs,
369 * FIXME: of course this is wrong in multi worktree
370 * setting. To be fixed real soon.
372 strbuf_addf(sb, "%s/logs", refs->gitcommondir);
376 switch (ref_type(refname)) {
377 case REF_TYPE_PER_WORKTREE:
378 case REF_TYPE_PSEUDOREF:
379 strbuf_addf(sb, "%s/logs/%s", refs->gitdir, refname);
381 case REF_TYPE_NORMAL:
382 strbuf_addf(sb, "%s/logs/%s", refs->gitcommondir, refname);
385 die("BUG: unknown ref type %d of ref %s",
386 ref_type(refname), refname);
390 static void files_ref_path(struct files_ref_store *refs,
394 switch (ref_type(refname)) {
395 case REF_TYPE_PER_WORKTREE:
396 case REF_TYPE_PSEUDOREF:
397 strbuf_addf(sb, "%s/%s", refs->gitdir, refname);
399 case REF_TYPE_NORMAL:
400 strbuf_addf(sb, "%s/%s", refs->gitcommondir, refname);
403 die("BUG: unknown ref type %d of ref %s",
404 ref_type(refname), refname);
409 * Check that the packed refs cache (if any) still reflects the
410 * contents of the file. If not, clear the cache.
412 static void validate_packed_ref_cache(struct packed_ref_store *refs)
415 !stat_validity_check(&refs->cache->validity, refs->path))
416 clear_packed_ref_cache(refs);
420 * Get the packed_ref_cache for the specified packed_ref_store,
421 * creating and populating it if it hasn't been read before or if the
422 * file has been changed (according to its `validity` field) since it
423 * was last read. On the other hand, if we hold the lock, then assume
424 * that the file hasn't been changed out from under us, so skip the
425 * extra `stat()` call in `stat_validity_check()`.
427 static struct packed_ref_cache *get_packed_ref_cache(struct packed_ref_store *refs)
429 if (!is_lock_file_locked(&refs->lock))
430 validate_packed_ref_cache(refs);
433 refs->cache = read_packed_refs(refs->path);
438 static struct ref_dir *get_packed_ref_dir(struct packed_ref_cache *packed_ref_cache)
440 return get_ref_dir(packed_ref_cache->cache->root);
443 static struct ref_dir *get_packed_refs(struct packed_ref_store *refs)
445 return get_packed_ref_dir(get_packed_ref_cache(refs));
449 * Add or overwrite a reference in the in-memory packed reference
450 * cache. This may only be called while the packed-refs file is locked
451 * (see lock_packed_refs()). To actually write the packed-refs file,
452 * call commit_packed_refs().
454 static void add_packed_ref(struct packed_ref_store *refs,
455 const char *refname, const struct object_id *oid)
457 struct ref_dir *packed_refs;
458 struct ref_entry *packed_entry;
460 if (!is_lock_file_locked(&refs->lock))
461 die("BUG: packed refs not locked");
463 if (check_refname_format(refname, REFNAME_ALLOW_ONELEVEL))
464 die("Reference has invalid format: '%s'", refname);
466 packed_refs = get_packed_refs(refs);
467 packed_entry = find_ref_entry(packed_refs, refname);
469 /* Overwrite the existing entry: */
470 oidcpy(&packed_entry->u.value.oid, oid);
471 packed_entry->flag = REF_ISPACKED;
472 oidclr(&packed_entry->u.value.peeled);
474 packed_entry = create_ref_entry(refname, oid, REF_ISPACKED);
475 add_ref_entry(packed_refs, packed_entry);
480 * Read the loose references from the namespace dirname into dir
481 * (without recursing). dirname must end with '/'. dir must be the
482 * directory entry corresponding to dirname.
484 static void loose_fill_ref_dir(struct ref_store *ref_store,
485 struct ref_dir *dir, const char *dirname)
487 struct files_ref_store *refs =
488 files_downcast(ref_store, REF_STORE_READ, "fill_ref_dir");
491 int dirnamelen = strlen(dirname);
492 struct strbuf refname;
493 struct strbuf path = STRBUF_INIT;
496 files_ref_path(refs, &path, dirname);
497 path_baselen = path.len;
499 d = opendir(path.buf);
501 strbuf_release(&path);
505 strbuf_init(&refname, dirnamelen + 257);
506 strbuf_add(&refname, dirname, dirnamelen);
508 while ((de = readdir(d)) != NULL) {
509 struct object_id oid;
513 if (de->d_name[0] == '.')
515 if (ends_with(de->d_name, ".lock"))
517 strbuf_addstr(&refname, de->d_name);
518 strbuf_addstr(&path, de->d_name);
519 if (stat(path.buf, &st) < 0) {
520 ; /* silently ignore */
521 } else if (S_ISDIR(st.st_mode)) {
522 strbuf_addch(&refname, '/');
523 add_entry_to_dir(dir,
524 create_dir_entry(dir->cache, refname.buf,
527 if (!refs_resolve_ref_unsafe(&refs->base,
532 flag |= REF_ISBROKEN;
533 } else if (is_null_oid(&oid)) {
535 * It is so astronomically unlikely
536 * that NULL_SHA1 is the SHA-1 of an
537 * actual object that we consider its
538 * appearance in a loose reference
539 * file to be repo corruption
540 * (probably due to a software bug).
542 flag |= REF_ISBROKEN;
545 if (check_refname_format(refname.buf,
546 REFNAME_ALLOW_ONELEVEL)) {
547 if (!refname_is_safe(refname.buf))
548 die("loose refname is dangerous: %s", refname.buf);
550 flag |= REF_BAD_NAME | REF_ISBROKEN;
552 add_entry_to_dir(dir,
553 create_ref_entry(refname.buf, &oid, flag));
555 strbuf_setlen(&refname, dirnamelen);
556 strbuf_setlen(&path, path_baselen);
558 strbuf_release(&refname);
559 strbuf_release(&path);
563 * Manually add refs/bisect, which, being per-worktree, might
564 * not appear in the directory listing for refs/ in the main
567 if (!strcmp(dirname, "refs/")) {
568 int pos = search_ref_dir(dir, "refs/bisect/", 12);
571 struct ref_entry *child_entry = create_dir_entry(
572 dir->cache, "refs/bisect/", 12, 1);
573 add_entry_to_dir(dir, child_entry);
578 static struct ref_cache *get_loose_ref_cache(struct files_ref_store *refs)
582 * Mark the top-level directory complete because we
583 * are about to read the only subdirectory that can
586 refs->loose = create_ref_cache(&refs->base, loose_fill_ref_dir);
588 /* We're going to fill the top level ourselves: */
589 refs->loose->root->flag &= ~REF_INCOMPLETE;
592 * Add an incomplete entry for "refs/" (to be filled
595 add_entry_to_dir(get_ref_dir(refs->loose->root),
596 create_dir_entry(refs->loose, "refs/", 5, 1));
602 * Return the ref_entry for the given refname from the packed
603 * references. If it does not exist, return NULL.
605 static struct ref_entry *get_packed_ref(struct packed_ref_store *refs,
608 return find_ref_entry(get_packed_refs(refs), refname);
611 static int packed_read_raw_ref(struct packed_ref_store *refs,
612 const char *refname, unsigned char *sha1,
613 struct strbuf *referent, unsigned int *type)
615 struct ref_entry *entry;
619 entry = get_packed_ref(refs, refname);
625 hashcpy(sha1, entry->u.value.oid.hash);
626 *type = REF_ISPACKED;
630 static int files_read_raw_ref(struct ref_store *ref_store,
631 const char *refname, unsigned char *sha1,
632 struct strbuf *referent, unsigned int *type)
634 struct files_ref_store *refs =
635 files_downcast(ref_store, REF_STORE_READ, "read_raw_ref");
636 struct strbuf sb_contents = STRBUF_INIT;
637 struct strbuf sb_path = STRBUF_INIT;
644 int remaining_retries = 3;
647 strbuf_reset(&sb_path);
649 files_ref_path(refs, &sb_path, refname);
655 * We might have to loop back here to avoid a race
656 * condition: first we lstat() the file, then we try
657 * to read it as a link or as a file. But if somebody
658 * changes the type of the file (file <-> directory
659 * <-> symlink) between the lstat() and reading, then
660 * we don't want to report that as an error but rather
661 * try again starting with the lstat().
663 * We'll keep a count of the retries, though, just to avoid
664 * any confusing situation sending us into an infinite loop.
667 if (remaining_retries-- <= 0)
670 if (lstat(path, &st) < 0) {
673 if (packed_read_raw_ref(refs->packed_ref_store, refname,
674 sha1, referent, type)) {
682 /* Follow "normalized" - ie "refs/.." symlinks by hand */
683 if (S_ISLNK(st.st_mode)) {
684 strbuf_reset(&sb_contents);
685 if (strbuf_readlink(&sb_contents, path, 0) < 0) {
686 if (errno == ENOENT || errno == EINVAL)
687 /* inconsistent with lstat; retry */
692 if (starts_with(sb_contents.buf, "refs/") &&
693 !check_refname_format(sb_contents.buf, 0)) {
694 strbuf_swap(&sb_contents, referent);
695 *type |= REF_ISSYMREF;
700 * It doesn't look like a refname; fall through to just
701 * treating it like a non-symlink, and reading whatever it
706 /* Is it a directory? */
707 if (S_ISDIR(st.st_mode)) {
709 * Even though there is a directory where the loose
710 * ref is supposed to be, there could still be a
713 if (packed_read_raw_ref(refs->packed_ref_store, refname,
714 sha1, referent, type)) {
723 * Anything else, just open it and try to use it as
726 fd = open(path, O_RDONLY);
728 if (errno == ENOENT && !S_ISLNK(st.st_mode))
729 /* inconsistent with lstat; retry */
734 strbuf_reset(&sb_contents);
735 if (strbuf_read(&sb_contents, fd, 256) < 0) {
736 int save_errno = errno;
742 strbuf_rtrim(&sb_contents);
743 buf = sb_contents.buf;
744 if (starts_with(buf, "ref:")) {
746 while (isspace(*buf))
749 strbuf_reset(referent);
750 strbuf_addstr(referent, buf);
751 *type |= REF_ISSYMREF;
757 * Please note that FETCH_HEAD has additional
758 * data after the sha.
760 if (get_sha1_hex(buf, sha1) ||
761 (buf[40] != '\0' && !isspace(buf[40]))) {
762 *type |= REF_ISBROKEN;
771 strbuf_release(&sb_path);
772 strbuf_release(&sb_contents);
777 static void unlock_ref(struct ref_lock *lock)
779 /* Do not free lock->lk -- atexit() still looks at them */
781 rollback_lock_file(lock->lk);
782 free(lock->ref_name);
787 * Lock refname, without following symrefs, and set *lock_p to point
788 * at a newly-allocated lock object. Fill in lock->old_oid, referent,
789 * and type similarly to read_raw_ref().
791 * The caller must verify that refname is a "safe" reference name (in
792 * the sense of refname_is_safe()) before calling this function.
794 * If the reference doesn't already exist, verify that refname doesn't
795 * have a D/F conflict with any existing references. extras and skip
796 * are passed to refs_verify_refname_available() for this check.
798 * If mustexist is not set and the reference is not found or is
799 * broken, lock the reference anyway but clear sha1.
801 * Return 0 on success. On failure, write an error message to err and
802 * return TRANSACTION_NAME_CONFLICT or TRANSACTION_GENERIC_ERROR.
804 * Implementation note: This function is basically
809 * but it includes a lot more code to
810 * - Deal with possible races with other processes
811 * - Avoid calling refs_verify_refname_available() when it can be
812 * avoided, namely if we were successfully able to read the ref
813 * - Generate informative error messages in the case of failure
815 static int lock_raw_ref(struct files_ref_store *refs,
816 const char *refname, int mustexist,
817 const struct string_list *extras,
818 const struct string_list *skip,
819 struct ref_lock **lock_p,
820 struct strbuf *referent,
824 struct ref_lock *lock;
825 struct strbuf ref_file = STRBUF_INIT;
826 int attempts_remaining = 3;
827 int ret = TRANSACTION_GENERIC_ERROR;
830 files_assert_main_repository(refs, "lock_raw_ref");
834 /* First lock the file so it can't change out from under us. */
836 *lock_p = lock = xcalloc(1, sizeof(*lock));
838 lock->ref_name = xstrdup(refname);
839 files_ref_path(refs, &ref_file, refname);
842 switch (safe_create_leading_directories(ref_file.buf)) {
847 * Suppose refname is "refs/foo/bar". We just failed
848 * to create the containing directory, "refs/foo",
849 * because there was a non-directory in the way. This
850 * indicates a D/F conflict, probably because of
851 * another reference such as "refs/foo". There is no
852 * reason to expect this error to be transitory.
854 if (refs_verify_refname_available(&refs->base, refname,
855 extras, skip, err)) {
858 * To the user the relevant error is
859 * that the "mustexist" reference is
863 strbuf_addf(err, "unable to resolve reference '%s'",
867 * The error message set by
868 * refs_verify_refname_available() is
871 ret = TRANSACTION_NAME_CONFLICT;
875 * The file that is in the way isn't a loose
876 * reference. Report it as a low-level
879 strbuf_addf(err, "unable to create lock file %s.lock; "
880 "non-directory in the way",
885 /* Maybe another process was tidying up. Try again. */
886 if (--attempts_remaining > 0)
890 strbuf_addf(err, "unable to create directory for %s",
896 lock->lk = xcalloc(1, sizeof(struct lock_file));
898 if (hold_lock_file_for_update(lock->lk, ref_file.buf, LOCK_NO_DEREF) < 0) {
899 if (errno == ENOENT && --attempts_remaining > 0) {
901 * Maybe somebody just deleted one of the
902 * directories leading to ref_file. Try
907 unable_to_lock_message(ref_file.buf, errno, err);
913 * Now we hold the lock and can read the reference without
914 * fear that its value will change.
917 if (files_read_raw_ref(&refs->base, refname,
918 lock->old_oid.hash, referent, type)) {
919 if (errno == ENOENT) {
921 /* Garden variety missing reference. */
922 strbuf_addf(err, "unable to resolve reference '%s'",
927 * Reference is missing, but that's OK. We
928 * know that there is not a conflict with
929 * another loose reference because
930 * (supposing that we are trying to lock
931 * reference "refs/foo/bar"):
933 * - We were successfully able to create
934 * the lockfile refs/foo/bar.lock, so we
935 * know there cannot be a loose reference
938 * - We got ENOENT and not EISDIR, so we
939 * know that there cannot be a loose
940 * reference named "refs/foo/bar/baz".
943 } else if (errno == EISDIR) {
945 * There is a directory in the way. It might have
946 * contained references that have been deleted. If
947 * we don't require that the reference already
948 * exists, try to remove the directory so that it
949 * doesn't cause trouble when we want to rename the
950 * lockfile into place later.
953 /* Garden variety missing reference. */
954 strbuf_addf(err, "unable to resolve reference '%s'",
957 } else if (remove_dir_recursively(&ref_file,
958 REMOVE_DIR_EMPTY_ONLY)) {
959 if (refs_verify_refname_available(
960 &refs->base, refname,
961 extras, skip, err)) {
963 * The error message set by
964 * verify_refname_available() is OK.
966 ret = TRANSACTION_NAME_CONFLICT;
970 * We can't delete the directory,
971 * but we also don't know of any
972 * references that it should
975 strbuf_addf(err, "there is a non-empty directory '%s' "
976 "blocking reference '%s'",
977 ref_file.buf, refname);
981 } else if (errno == EINVAL && (*type & REF_ISBROKEN)) {
982 strbuf_addf(err, "unable to resolve reference '%s': "
983 "reference broken", refname);
986 strbuf_addf(err, "unable to resolve reference '%s': %s",
987 refname, strerror(errno));
992 * If the ref did not exist and we are creating it,
993 * make sure there is no existing ref that conflicts
996 if (refs_verify_refname_available(
997 &refs->base, refname,
1010 strbuf_release(&ref_file);
1014 static int packed_peel_ref(struct packed_ref_store *refs,
1015 const char *refname, unsigned char *sha1)
1017 struct ref_entry *r = get_packed_ref(refs, refname);
1019 if (!r || peel_entry(r, 0))
1022 hashcpy(sha1, r->u.value.peeled.hash);
1026 static int files_peel_ref(struct ref_store *ref_store,
1027 const char *refname, unsigned char *sha1)
1029 struct files_ref_store *refs =
1030 files_downcast(ref_store, REF_STORE_READ | REF_STORE_ODB,
1033 unsigned char base[20];
1035 if (current_ref_iter && current_ref_iter->refname == refname) {
1036 struct object_id peeled;
1038 if (ref_iterator_peel(current_ref_iter, &peeled))
1040 hashcpy(sha1, peeled.hash);
1044 if (refs_read_ref_full(ref_store, refname,
1045 RESOLVE_REF_READING, base, &flag))
1049 * If the reference is packed, read its ref_entry from the
1050 * cache in the hope that we already know its peeled value.
1051 * We only try this optimization on packed references because
1052 * (a) forcing the filling of the loose reference cache could
1053 * be expensive and (b) loose references anyway usually do not
1054 * have REF_KNOWS_PEELED.
1056 if (flag & REF_ISPACKED &&
1057 !packed_peel_ref(refs->packed_ref_store, refname, sha1))
1060 return peel_object(base, sha1);
1063 struct packed_ref_iterator {
1064 struct ref_iterator base;
1066 struct packed_ref_cache *cache;
1067 struct ref_iterator *iter0;
1071 static int packed_ref_iterator_advance(struct ref_iterator *ref_iterator)
1073 struct packed_ref_iterator *iter =
1074 (struct packed_ref_iterator *)ref_iterator;
1077 while ((ok = ref_iterator_advance(iter->iter0)) == ITER_OK) {
1078 if (iter->flags & DO_FOR_EACH_PER_WORKTREE_ONLY &&
1079 ref_type(iter->iter0->refname) != REF_TYPE_PER_WORKTREE)
1082 if (!(iter->flags & DO_FOR_EACH_INCLUDE_BROKEN) &&
1083 !ref_resolves_to_object(iter->iter0->refname,
1085 iter->iter0->flags))
1088 iter->base.refname = iter->iter0->refname;
1089 iter->base.oid = iter->iter0->oid;
1090 iter->base.flags = iter->iter0->flags;
1095 if (ref_iterator_abort(ref_iterator) != ITER_DONE)
1101 static int packed_ref_iterator_peel(struct ref_iterator *ref_iterator,
1102 struct object_id *peeled)
1104 struct packed_ref_iterator *iter =
1105 (struct packed_ref_iterator *)ref_iterator;
1107 return ref_iterator_peel(iter->iter0, peeled);
1110 static int packed_ref_iterator_abort(struct ref_iterator *ref_iterator)
1112 struct packed_ref_iterator *iter =
1113 (struct packed_ref_iterator *)ref_iterator;
1117 ok = ref_iterator_abort(iter->iter0);
1119 release_packed_ref_cache(iter->cache);
1120 base_ref_iterator_free(ref_iterator);
1124 static struct ref_iterator_vtable packed_ref_iterator_vtable = {
1125 packed_ref_iterator_advance,
1126 packed_ref_iterator_peel,
1127 packed_ref_iterator_abort
1130 static struct ref_iterator *packed_ref_iterator_begin(
1131 struct packed_ref_store *refs,
1132 const char *prefix, unsigned int flags)
1134 struct packed_ref_iterator *iter;
1135 struct ref_iterator *ref_iterator;
1137 iter = xcalloc(1, sizeof(*iter));
1138 ref_iterator = &iter->base;
1139 base_ref_iterator_init(ref_iterator, &packed_ref_iterator_vtable);
1142 * Note that get_packed_ref_cache() internally checks whether
1143 * the packed-ref cache is up to date with what is on disk,
1144 * and re-reads it if not.
1147 iter->cache = get_packed_ref_cache(refs);
1148 acquire_packed_ref_cache(iter->cache);
1149 iter->iter0 = cache_ref_iterator_begin(iter->cache->cache, prefix, 0);
1151 iter->flags = flags;
1153 return ref_iterator;
1156 struct files_ref_iterator {
1157 struct ref_iterator base;
1159 struct ref_iterator *iter0;
1163 static int files_ref_iterator_advance(struct ref_iterator *ref_iterator)
1165 struct files_ref_iterator *iter =
1166 (struct files_ref_iterator *)ref_iterator;
1169 while ((ok = ref_iterator_advance(iter->iter0)) == ITER_OK) {
1170 if (iter->flags & DO_FOR_EACH_PER_WORKTREE_ONLY &&
1171 ref_type(iter->iter0->refname) != REF_TYPE_PER_WORKTREE)
1174 if (!(iter->flags & DO_FOR_EACH_INCLUDE_BROKEN) &&
1175 !ref_resolves_to_object(iter->iter0->refname,
1177 iter->iter0->flags))
1180 iter->base.refname = iter->iter0->refname;
1181 iter->base.oid = iter->iter0->oid;
1182 iter->base.flags = iter->iter0->flags;
1187 if (ref_iterator_abort(ref_iterator) != ITER_DONE)
1193 static int files_ref_iterator_peel(struct ref_iterator *ref_iterator,
1194 struct object_id *peeled)
1196 struct files_ref_iterator *iter =
1197 (struct files_ref_iterator *)ref_iterator;
1199 return ref_iterator_peel(iter->iter0, peeled);
1202 static int files_ref_iterator_abort(struct ref_iterator *ref_iterator)
1204 struct files_ref_iterator *iter =
1205 (struct files_ref_iterator *)ref_iterator;
1209 ok = ref_iterator_abort(iter->iter0);
1211 base_ref_iterator_free(ref_iterator);
1215 static struct ref_iterator_vtable files_ref_iterator_vtable = {
1216 files_ref_iterator_advance,
1217 files_ref_iterator_peel,
1218 files_ref_iterator_abort
1221 static struct ref_iterator *files_ref_iterator_begin(
1222 struct ref_store *ref_store,
1223 const char *prefix, unsigned int flags)
1225 struct files_ref_store *refs;
1226 struct ref_iterator *loose_iter, *packed_iter;
1227 struct files_ref_iterator *iter;
1228 struct ref_iterator *ref_iterator;
1229 unsigned int required_flags = REF_STORE_READ;
1231 if (!(flags & DO_FOR_EACH_INCLUDE_BROKEN))
1232 required_flags |= REF_STORE_ODB;
1234 refs = files_downcast(ref_store, required_flags, "ref_iterator_begin");
1236 iter = xcalloc(1, sizeof(*iter));
1237 ref_iterator = &iter->base;
1238 base_ref_iterator_init(ref_iterator, &files_ref_iterator_vtable);
1241 * We must make sure that all loose refs are read before
1242 * accessing the packed-refs file; this avoids a race
1243 * condition if loose refs are migrated to the packed-refs
1244 * file by a simultaneous process, but our in-memory view is
1245 * from before the migration. We ensure this as follows:
1246 * First, we call start the loose refs iteration with its
1247 * `prime_ref` argument set to true. This causes the loose
1248 * references in the subtree to be pre-read into the cache.
1249 * (If they've already been read, that's OK; we only need to
1250 * guarantee that they're read before the packed refs, not
1251 * *how much* before.) After that, we call
1252 * packed_ref_iterator_begin(), which internally checks
1253 * whether the packed-ref cache is up to date with what is on
1254 * disk, and re-reads it if not.
1257 loose_iter = cache_ref_iterator_begin(get_loose_ref_cache(refs),
1261 * The packed-refs file might contain broken references, for
1262 * example an old version of a reference that points at an
1263 * object that has since been garbage-collected. This is OK as
1264 * long as there is a corresponding loose reference that
1265 * overrides it, and we don't want to emit an error message in
1266 * this case. So ask the packed_ref_store for all of its
1267 * references, and (if needed) do our own check for broken
1268 * ones in files_ref_iterator_advance(), after we have merged
1269 * the packed and loose references.
1271 packed_iter = packed_ref_iterator_begin(
1272 refs->packed_ref_store, prefix,
1273 DO_FOR_EACH_INCLUDE_BROKEN);
1275 iter->iter0 = overlay_ref_iterator_begin(loose_iter, packed_iter);
1276 iter->flags = flags;
1278 return ref_iterator;
1282 * Verify that the reference locked by lock has the value old_sha1.
1283 * Fail if the reference doesn't exist and mustexist is set. Return 0
1284 * on success. On error, write an error message to err, set errno, and
1285 * return a negative value.
1287 static int verify_lock(struct ref_store *ref_store, struct ref_lock *lock,
1288 const unsigned char *old_sha1, int mustexist,
1293 if (refs_read_ref_full(ref_store, lock->ref_name,
1294 mustexist ? RESOLVE_REF_READING : 0,
1295 lock->old_oid.hash, NULL)) {
1297 int save_errno = errno;
1298 strbuf_addf(err, "can't verify ref '%s'", lock->ref_name);
1302 oidclr(&lock->old_oid);
1306 if (old_sha1 && hashcmp(lock->old_oid.hash, old_sha1)) {
1307 strbuf_addf(err, "ref '%s' is at %s but expected %s",
1309 oid_to_hex(&lock->old_oid),
1310 sha1_to_hex(old_sha1));
1317 static int remove_empty_directories(struct strbuf *path)
1320 * we want to create a file but there is a directory there;
1321 * if that is an empty directory (or a directory that contains
1322 * only empty directories), remove them.
1324 return remove_dir_recursively(path, REMOVE_DIR_EMPTY_ONLY);
1327 static int create_reflock(const char *path, void *cb)
1329 struct lock_file *lk = cb;
1331 return hold_lock_file_for_update(lk, path, LOCK_NO_DEREF) < 0 ? -1 : 0;
1335 * Locks a ref returning the lock on success and NULL on failure.
1336 * On failure errno is set to something meaningful.
1338 static struct ref_lock *lock_ref_sha1_basic(struct files_ref_store *refs,
1339 const char *refname,
1340 const unsigned char *old_sha1,
1341 const struct string_list *extras,
1342 const struct string_list *skip,
1343 unsigned int flags, int *type,
1346 struct strbuf ref_file = STRBUF_INIT;
1347 struct ref_lock *lock;
1349 int mustexist = (old_sha1 && !is_null_sha1(old_sha1));
1350 int resolve_flags = RESOLVE_REF_NO_RECURSE;
1353 files_assert_main_repository(refs, "lock_ref_sha1_basic");
1356 lock = xcalloc(1, sizeof(struct ref_lock));
1359 resolve_flags |= RESOLVE_REF_READING;
1360 if (flags & REF_DELETING)
1361 resolve_flags |= RESOLVE_REF_ALLOW_BAD_NAME;
1363 files_ref_path(refs, &ref_file, refname);
1364 resolved = !!refs_resolve_ref_unsafe(&refs->base,
1365 refname, resolve_flags,
1366 lock->old_oid.hash, type);
1367 if (!resolved && errno == EISDIR) {
1369 * we are trying to lock foo but we used to
1370 * have foo/bar which now does not exist;
1371 * it is normal for the empty directory 'foo'
1374 if (remove_empty_directories(&ref_file)) {
1376 if (!refs_verify_refname_available(
1378 refname, extras, skip, err))
1379 strbuf_addf(err, "there are still refs under '%s'",
1383 resolved = !!refs_resolve_ref_unsafe(&refs->base,
1384 refname, resolve_flags,
1385 lock->old_oid.hash, type);
1389 if (last_errno != ENOTDIR ||
1390 !refs_verify_refname_available(&refs->base, refname,
1392 strbuf_addf(err, "unable to resolve reference '%s': %s",
1393 refname, strerror(last_errno));
1399 * If the ref did not exist and we are creating it, make sure
1400 * there is no existing packed ref whose name begins with our
1401 * refname, nor a packed ref whose name is a proper prefix of
1404 if (is_null_oid(&lock->old_oid) &&
1405 refs_verify_refname_available(&refs->base, refname,
1406 extras, skip, err)) {
1407 last_errno = ENOTDIR;
1411 lock->lk = xcalloc(1, sizeof(struct lock_file));
1413 lock->ref_name = xstrdup(refname);
1415 if (raceproof_create_file(ref_file.buf, create_reflock, lock->lk)) {
1417 unable_to_lock_message(ref_file.buf, errno, err);
1421 if (verify_lock(&refs->base, lock, old_sha1, mustexist, err)) {
1432 strbuf_release(&ref_file);
1438 * Write an entry to the packed-refs file for the specified refname.
1439 * If peeled is non-NULL, write it as the entry's peeled value.
1441 static void write_packed_entry(FILE *fh, const char *refname,
1442 const unsigned char *sha1,
1443 const unsigned char *peeled)
1445 fprintf_or_die(fh, "%s %s\n", sha1_to_hex(sha1), refname);
1447 fprintf_or_die(fh, "^%s\n", sha1_to_hex(peeled));
1451 * Lock the packed-refs file for writing. Flags is passed to
1452 * hold_lock_file_for_update(). Return 0 on success. On errors, set
1453 * errno appropriately and return a nonzero value.
1455 static int lock_packed_refs(struct packed_ref_store *refs, int flags)
1457 static int timeout_configured = 0;
1458 static int timeout_value = 1000;
1459 struct packed_ref_cache *packed_ref_cache;
1461 packed_assert_main_repository(refs, "lock_packed_refs");
1463 if (!timeout_configured) {
1464 git_config_get_int("core.packedrefstimeout", &timeout_value);
1465 timeout_configured = 1;
1468 if (hold_lock_file_for_update_timeout(
1471 flags, timeout_value) < 0)
1475 * Now that we hold the `packed-refs` lock, make sure that our
1476 * cache matches the current version of the file. Normally
1477 * `get_packed_ref_cache()` does that for us, but that
1478 * function assumes that when the file is locked, any existing
1479 * cache is still valid. We've just locked the file, but it
1480 * might have changed the moment *before* we locked it.
1482 validate_packed_ref_cache(refs);
1484 packed_ref_cache = get_packed_ref_cache(refs);
1485 /* Increment the reference count to prevent it from being freed: */
1486 acquire_packed_ref_cache(packed_ref_cache);
1491 * Write the current version of the packed refs cache from memory to
1492 * disk. The packed-refs file must already be locked for writing (see
1493 * lock_packed_refs()). Return zero on success. On errors, set errno
1494 * and return a nonzero value
1496 static int commit_packed_refs(struct packed_ref_store *refs)
1498 struct packed_ref_cache *packed_ref_cache =
1499 get_packed_ref_cache(refs);
1503 struct ref_iterator *iter;
1505 packed_assert_main_repository(refs, "commit_packed_refs");
1507 if (!is_lock_file_locked(&refs->lock))
1508 die("BUG: packed-refs not locked");
1510 out = fdopen_lock_file(&refs->lock, "w");
1512 die_errno("unable to fdopen packed-refs descriptor");
1514 fprintf_or_die(out, "%s", PACKED_REFS_HEADER);
1516 iter = cache_ref_iterator_begin(packed_ref_cache->cache, NULL, 0);
1517 while ((ok = ref_iterator_advance(iter)) == ITER_OK) {
1518 struct object_id peeled;
1519 int peel_error = ref_iterator_peel(iter, &peeled);
1521 write_packed_entry(out, iter->refname, iter->oid->hash,
1522 peel_error ? NULL : peeled.hash);
1525 if (ok != ITER_DONE)
1526 die("error while iterating over references");
1528 if (commit_lock_file(&refs->lock)) {
1532 release_packed_ref_cache(packed_ref_cache);
1538 * Rollback the lockfile for the packed-refs file, and discard the
1539 * in-memory packed reference cache. (The packed-refs file will be
1540 * read anew if it is needed again after this function is called.)
1542 static void rollback_packed_refs(struct packed_ref_store *refs)
1544 struct packed_ref_cache *packed_ref_cache = get_packed_ref_cache(refs);
1546 packed_assert_main_repository(refs, "rollback_packed_refs");
1548 if (!is_lock_file_locked(&refs->lock))
1549 die("BUG: packed-refs not locked");
1550 rollback_lock_file(&refs->lock);
1551 release_packed_ref_cache(packed_ref_cache);
1552 clear_packed_ref_cache(refs);
1555 struct ref_to_prune {
1556 struct ref_to_prune *next;
1557 unsigned char sha1[20];
1558 char name[FLEX_ARRAY];
1562 REMOVE_EMPTY_PARENTS_REF = 0x01,
1563 REMOVE_EMPTY_PARENTS_REFLOG = 0x02
1567 * Remove empty parent directories associated with the specified
1568 * reference and/or its reflog, but spare [logs/]refs/ and immediate
1569 * subdirs. flags is a combination of REMOVE_EMPTY_PARENTS_REF and/or
1570 * REMOVE_EMPTY_PARENTS_REFLOG.
1572 static void try_remove_empty_parents(struct files_ref_store *refs,
1573 const char *refname,
1576 struct strbuf buf = STRBUF_INIT;
1577 struct strbuf sb = STRBUF_INIT;
1581 strbuf_addstr(&buf, refname);
1583 for (i = 0; i < 2; i++) { /* refs/{heads,tags,...}/ */
1584 while (*p && *p != '/')
1586 /* tolerate duplicate slashes; see check_refname_format() */
1590 q = buf.buf + buf.len;
1591 while (flags & (REMOVE_EMPTY_PARENTS_REF | REMOVE_EMPTY_PARENTS_REFLOG)) {
1592 while (q > p && *q != '/')
1594 while (q > p && *(q-1) == '/')
1598 strbuf_setlen(&buf, q - buf.buf);
1601 files_ref_path(refs, &sb, buf.buf);
1602 if ((flags & REMOVE_EMPTY_PARENTS_REF) && rmdir(sb.buf))
1603 flags &= ~REMOVE_EMPTY_PARENTS_REF;
1606 files_reflog_path(refs, &sb, buf.buf);
1607 if ((flags & REMOVE_EMPTY_PARENTS_REFLOG) && rmdir(sb.buf))
1608 flags &= ~REMOVE_EMPTY_PARENTS_REFLOG;
1610 strbuf_release(&buf);
1611 strbuf_release(&sb);
1614 /* make sure nobody touched the ref, and unlink */
1615 static void prune_ref(struct files_ref_store *refs, struct ref_to_prune *r)
1617 struct ref_transaction *transaction;
1618 struct strbuf err = STRBUF_INIT;
1620 if (check_refname_format(r->name, 0))
1623 transaction = ref_store_transaction_begin(&refs->base, &err);
1625 ref_transaction_delete(transaction, r->name, r->sha1,
1626 REF_ISPRUNING | REF_NODEREF, NULL, &err) ||
1627 ref_transaction_commit(transaction, &err)) {
1628 ref_transaction_free(transaction);
1629 error("%s", err.buf);
1630 strbuf_release(&err);
1633 ref_transaction_free(transaction);
1634 strbuf_release(&err);
1637 static void prune_refs(struct files_ref_store *refs, struct ref_to_prune *r)
1646 * Return true if the specified reference should be packed.
1648 static int should_pack_ref(const char *refname,
1649 const struct object_id *oid, unsigned int ref_flags,
1650 unsigned int pack_flags)
1652 /* Do not pack per-worktree refs: */
1653 if (ref_type(refname) != REF_TYPE_NORMAL)
1656 /* Do not pack non-tags unless PACK_REFS_ALL is set: */
1657 if (!(pack_flags & PACK_REFS_ALL) && !starts_with(refname, "refs/tags/"))
1660 /* Do not pack symbolic refs: */
1661 if (ref_flags & REF_ISSYMREF)
1664 /* Do not pack broken refs: */
1665 if (!ref_resolves_to_object(refname, oid, ref_flags))
1671 static int files_pack_refs(struct ref_store *ref_store, unsigned int flags)
1673 struct files_ref_store *refs =
1674 files_downcast(ref_store, REF_STORE_WRITE | REF_STORE_ODB,
1676 struct ref_iterator *iter;
1678 struct ref_to_prune *refs_to_prune = NULL;
1680 lock_packed_refs(refs->packed_ref_store, LOCK_DIE_ON_ERROR);
1682 iter = cache_ref_iterator_begin(get_loose_ref_cache(refs), NULL, 0);
1683 while ((ok = ref_iterator_advance(iter)) == ITER_OK) {
1685 * If the loose reference can be packed, add an entry
1686 * in the packed ref cache. If the reference should be
1687 * pruned, also add it to refs_to_prune.
1689 if (!should_pack_ref(iter->refname, iter->oid, iter->flags,
1694 * Create an entry in the packed-refs cache equivalent
1695 * to the one from the loose ref cache, except that
1696 * we don't copy the peeled status, because we want it
1699 add_packed_ref(refs->packed_ref_store, iter->refname, iter->oid);
1701 /* Schedule the loose reference for pruning if requested. */
1702 if ((flags & PACK_REFS_PRUNE)) {
1703 struct ref_to_prune *n;
1704 FLEX_ALLOC_STR(n, name, iter->refname);
1705 hashcpy(n->sha1, iter->oid->hash);
1706 n->next = refs_to_prune;
1710 if (ok != ITER_DONE)
1711 die("error while iterating over references");
1713 if (commit_packed_refs(refs->packed_ref_store))
1714 die_errno("unable to overwrite old ref-pack file");
1716 prune_refs(refs, refs_to_prune);
1721 * Rewrite the packed-refs file, omitting any refs listed in
1722 * 'refnames'. On error, leave packed-refs unchanged, write an error
1723 * message to 'err', and return a nonzero value.
1725 * The refs in 'refnames' needn't be sorted. `err` must not be NULL.
1727 static int repack_without_refs(struct packed_ref_store *refs,
1728 struct string_list *refnames, struct strbuf *err)
1730 struct ref_dir *packed;
1731 struct string_list_item *refname;
1732 int ret, needs_repacking = 0, removed = 0;
1734 packed_assert_main_repository(refs, "repack_without_refs");
1737 /* Look for a packed ref */
1738 for_each_string_list_item(refname, refnames) {
1739 if (get_packed_ref(refs, refname->string)) {
1740 needs_repacking = 1;
1745 /* Avoid locking if we have nothing to do */
1746 if (!needs_repacking)
1747 return 0; /* no refname exists in packed refs */
1749 if (lock_packed_refs(refs, 0)) {
1750 unable_to_lock_message(refs->path, errno, err);
1753 packed = get_packed_refs(refs);
1755 /* Remove refnames from the cache */
1756 for_each_string_list_item(refname, refnames)
1757 if (remove_entry_from_dir(packed, refname->string) != -1)
1761 * All packed entries disappeared while we were
1762 * acquiring the lock.
1764 rollback_packed_refs(refs);
1768 /* Write what remains */
1769 ret = commit_packed_refs(refs);
1771 strbuf_addf(err, "unable to overwrite old ref-pack file: %s",
1776 static int files_delete_refs(struct ref_store *ref_store, const char *msg,
1777 struct string_list *refnames, unsigned int flags)
1779 struct files_ref_store *refs =
1780 files_downcast(ref_store, REF_STORE_WRITE, "delete_refs");
1781 struct strbuf err = STRBUF_INIT;
1787 result = repack_without_refs(refs->packed_ref_store, refnames, &err);
1790 * If we failed to rewrite the packed-refs file, then
1791 * it is unsafe to try to remove loose refs, because
1792 * doing so might expose an obsolete packed value for
1793 * a reference that might even point at an object that
1794 * has been garbage collected.
1796 if (refnames->nr == 1)
1797 error(_("could not delete reference %s: %s"),
1798 refnames->items[0].string, err.buf);
1800 error(_("could not delete references: %s"), err.buf);
1805 for (i = 0; i < refnames->nr; i++) {
1806 const char *refname = refnames->items[i].string;
1808 if (refs_delete_ref(&refs->base, msg, refname, NULL, flags))
1809 result |= error(_("could not remove reference %s"), refname);
1813 strbuf_release(&err);
1818 * People using contrib's git-new-workdir have .git/logs/refs ->
1819 * /some/other/path/.git/logs/refs, and that may live on another device.
1821 * IOW, to avoid cross device rename errors, the temporary renamed log must
1822 * live into logs/refs.
1824 #define TMP_RENAMED_LOG "refs/.tmp-renamed-log"
1827 const char *tmp_renamed_log;
1831 static int rename_tmp_log_callback(const char *path, void *cb_data)
1833 struct rename_cb *cb = cb_data;
1835 if (rename(cb->tmp_renamed_log, path)) {
1837 * rename(a, b) when b is an existing directory ought
1838 * to result in ISDIR, but Solaris 5.8 gives ENOTDIR.
1839 * Sheesh. Record the true errno for error reporting,
1840 * but report EISDIR to raceproof_create_file() so
1841 * that it knows to retry.
1843 cb->true_errno = errno;
1844 if (errno == ENOTDIR)
1852 static int rename_tmp_log(struct files_ref_store *refs, const char *newrefname)
1854 struct strbuf path = STRBUF_INIT;
1855 struct strbuf tmp = STRBUF_INIT;
1856 struct rename_cb cb;
1859 files_reflog_path(refs, &path, newrefname);
1860 files_reflog_path(refs, &tmp, TMP_RENAMED_LOG);
1861 cb.tmp_renamed_log = tmp.buf;
1862 ret = raceproof_create_file(path.buf, rename_tmp_log_callback, &cb);
1864 if (errno == EISDIR)
1865 error("directory not empty: %s", path.buf);
1867 error("unable to move logfile %s to %s: %s",
1869 strerror(cb.true_errno));
1872 strbuf_release(&path);
1873 strbuf_release(&tmp);
1877 static int write_ref_to_lockfile(struct ref_lock *lock,
1878 const struct object_id *oid, struct strbuf *err);
1879 static int commit_ref_update(struct files_ref_store *refs,
1880 struct ref_lock *lock,
1881 const struct object_id *oid, const char *logmsg,
1882 struct strbuf *err);
1884 static int files_rename_ref(struct ref_store *ref_store,
1885 const char *oldrefname, const char *newrefname,
1888 struct files_ref_store *refs =
1889 files_downcast(ref_store, REF_STORE_WRITE, "rename_ref");
1890 struct object_id oid, orig_oid;
1891 int flag = 0, logmoved = 0;
1892 struct ref_lock *lock;
1893 struct stat loginfo;
1894 struct strbuf sb_oldref = STRBUF_INIT;
1895 struct strbuf sb_newref = STRBUF_INIT;
1896 struct strbuf tmp_renamed_log = STRBUF_INIT;
1898 struct strbuf err = STRBUF_INIT;
1900 files_reflog_path(refs, &sb_oldref, oldrefname);
1901 files_reflog_path(refs, &sb_newref, newrefname);
1902 files_reflog_path(refs, &tmp_renamed_log, TMP_RENAMED_LOG);
1904 log = !lstat(sb_oldref.buf, &loginfo);
1905 if (log && S_ISLNK(loginfo.st_mode)) {
1906 ret = error("reflog for %s is a symlink", oldrefname);
1910 if (!refs_resolve_ref_unsafe(&refs->base, oldrefname,
1911 RESOLVE_REF_READING | RESOLVE_REF_NO_RECURSE,
1912 orig_oid.hash, &flag)) {
1913 ret = error("refname %s not found", oldrefname);
1917 if (flag & REF_ISSYMREF) {
1918 ret = error("refname %s is a symbolic ref, renaming it is not supported",
1922 if (!refs_rename_ref_available(&refs->base, oldrefname, newrefname)) {
1927 if (log && rename(sb_oldref.buf, tmp_renamed_log.buf)) {
1928 ret = error("unable to move logfile logs/%s to logs/"TMP_RENAMED_LOG": %s",
1929 oldrefname, strerror(errno));
1933 if (refs_delete_ref(&refs->base, logmsg, oldrefname,
1934 orig_oid.hash, REF_NODEREF)) {
1935 error("unable to delete old %s", oldrefname);
1940 * Since we are doing a shallow lookup, oid is not the
1941 * correct value to pass to delete_ref as old_oid. But that
1942 * doesn't matter, because an old_oid check wouldn't add to
1943 * the safety anyway; we want to delete the reference whatever
1944 * its current value.
1946 if (!refs_read_ref_full(&refs->base, newrefname,
1947 RESOLVE_REF_READING | RESOLVE_REF_NO_RECURSE,
1949 refs_delete_ref(&refs->base, NULL, newrefname,
1950 NULL, REF_NODEREF)) {
1951 if (errno == EISDIR) {
1952 struct strbuf path = STRBUF_INIT;
1955 files_ref_path(refs, &path, newrefname);
1956 result = remove_empty_directories(&path);
1957 strbuf_release(&path);
1960 error("Directory not empty: %s", newrefname);
1964 error("unable to delete existing %s", newrefname);
1969 if (log && rename_tmp_log(refs, newrefname))
1974 lock = lock_ref_sha1_basic(refs, newrefname, NULL, NULL, NULL,
1975 REF_NODEREF, NULL, &err);
1977 error("unable to rename '%s' to '%s': %s", oldrefname, newrefname, err.buf);
1978 strbuf_release(&err);
1981 oidcpy(&lock->old_oid, &orig_oid);
1983 if (write_ref_to_lockfile(lock, &orig_oid, &err) ||
1984 commit_ref_update(refs, lock, &orig_oid, logmsg, &err)) {
1985 error("unable to write current sha1 into %s: %s", newrefname, err.buf);
1986 strbuf_release(&err);
1994 lock = lock_ref_sha1_basic(refs, oldrefname, NULL, NULL, NULL,
1995 REF_NODEREF, NULL, &err);
1997 error("unable to lock %s for rollback: %s", oldrefname, err.buf);
1998 strbuf_release(&err);
2002 flag = log_all_ref_updates;
2003 log_all_ref_updates = LOG_REFS_NONE;
2004 if (write_ref_to_lockfile(lock, &orig_oid, &err) ||
2005 commit_ref_update(refs, lock, &orig_oid, NULL, &err)) {
2006 error("unable to write current sha1 into %s: %s", oldrefname, err.buf);
2007 strbuf_release(&err);
2009 log_all_ref_updates = flag;
2012 if (logmoved && rename(sb_newref.buf, sb_oldref.buf))
2013 error("unable to restore logfile %s from %s: %s",
2014 oldrefname, newrefname, strerror(errno));
2015 if (!logmoved && log &&
2016 rename(tmp_renamed_log.buf, sb_oldref.buf))
2017 error("unable to restore logfile %s from logs/"TMP_RENAMED_LOG": %s",
2018 oldrefname, strerror(errno));
2021 strbuf_release(&sb_newref);
2022 strbuf_release(&sb_oldref);
2023 strbuf_release(&tmp_renamed_log);
2028 static int close_ref(struct ref_lock *lock)
2030 if (close_lock_file(lock->lk))
2035 static int commit_ref(struct ref_lock *lock)
2037 char *path = get_locked_file_path(lock->lk);
2040 if (!lstat(path, &st) && S_ISDIR(st.st_mode)) {
2042 * There is a directory at the path we want to rename
2043 * the lockfile to. Hopefully it is empty; try to
2046 size_t len = strlen(path);
2047 struct strbuf sb_path = STRBUF_INIT;
2049 strbuf_attach(&sb_path, path, len, len);
2052 * If this fails, commit_lock_file() will also fail
2053 * and will report the problem.
2055 remove_empty_directories(&sb_path);
2056 strbuf_release(&sb_path);
2061 if (commit_lock_file(lock->lk))
2066 static int open_or_create_logfile(const char *path, void *cb)
2070 *fd = open(path, O_APPEND | O_WRONLY | O_CREAT, 0666);
2071 return (*fd < 0) ? -1 : 0;
2075 * Create a reflog for a ref. If force_create = 0, only create the
2076 * reflog for certain refs (those for which should_autocreate_reflog
2077 * returns non-zero). Otherwise, create it regardless of the reference
2078 * name. If the logfile already existed or was created, return 0 and
2079 * set *logfd to the file descriptor opened for appending to the file.
2080 * If no logfile exists and we decided not to create one, return 0 and
2081 * set *logfd to -1. On failure, fill in *err, set *logfd to -1, and
2084 static int log_ref_setup(struct files_ref_store *refs,
2085 const char *refname, int force_create,
2086 int *logfd, struct strbuf *err)
2088 struct strbuf logfile_sb = STRBUF_INIT;
2091 files_reflog_path(refs, &logfile_sb, refname);
2092 logfile = strbuf_detach(&logfile_sb, NULL);
2094 if (force_create || should_autocreate_reflog(refname)) {
2095 if (raceproof_create_file(logfile, open_or_create_logfile, logfd)) {
2096 if (errno == ENOENT)
2097 strbuf_addf(err, "unable to create directory for '%s': "
2098 "%s", logfile, strerror(errno));
2099 else if (errno == EISDIR)
2100 strbuf_addf(err, "there are still logs under '%s'",
2103 strbuf_addf(err, "unable to append to '%s': %s",
2104 logfile, strerror(errno));
2109 *logfd = open(logfile, O_APPEND | O_WRONLY, 0666);
2111 if (errno == ENOENT || errno == EISDIR) {
2113 * The logfile doesn't already exist,
2114 * but that is not an error; it only
2115 * means that we won't write log
2120 strbuf_addf(err, "unable to append to '%s': %s",
2121 logfile, strerror(errno));
2128 adjust_shared_perm(logfile);
2138 static int files_create_reflog(struct ref_store *ref_store,
2139 const char *refname, int force_create,
2142 struct files_ref_store *refs =
2143 files_downcast(ref_store, REF_STORE_WRITE, "create_reflog");
2146 if (log_ref_setup(refs, refname, force_create, &fd, err))
2155 static int log_ref_write_fd(int fd, const struct object_id *old_oid,
2156 const struct object_id *new_oid,
2157 const char *committer, const char *msg)
2159 int msglen, written;
2160 unsigned maxlen, len;
2163 msglen = msg ? strlen(msg) : 0;
2164 maxlen = strlen(committer) + msglen + 100;
2165 logrec = xmalloc(maxlen);
2166 len = xsnprintf(logrec, maxlen, "%s %s %s\n",
2167 oid_to_hex(old_oid),
2168 oid_to_hex(new_oid),
2171 len += copy_reflog_msg(logrec + len - 1, msg) - 1;
2173 written = len <= maxlen ? write_in_full(fd, logrec, len) : -1;
2181 static int files_log_ref_write(struct files_ref_store *refs,
2182 const char *refname, const struct object_id *old_oid,
2183 const struct object_id *new_oid, const char *msg,
2184 int flags, struct strbuf *err)
2188 if (log_all_ref_updates == LOG_REFS_UNSET)
2189 log_all_ref_updates = is_bare_repository() ? LOG_REFS_NONE : LOG_REFS_NORMAL;
2191 result = log_ref_setup(refs, refname,
2192 flags & REF_FORCE_CREATE_REFLOG,
2200 result = log_ref_write_fd(logfd, old_oid, new_oid,
2201 git_committer_info(0), msg);
2203 struct strbuf sb = STRBUF_INIT;
2204 int save_errno = errno;
2206 files_reflog_path(refs, &sb, refname);
2207 strbuf_addf(err, "unable to append to '%s': %s",
2208 sb.buf, strerror(save_errno));
2209 strbuf_release(&sb);
2214 struct strbuf sb = STRBUF_INIT;
2215 int save_errno = errno;
2217 files_reflog_path(refs, &sb, refname);
2218 strbuf_addf(err, "unable to append to '%s': %s",
2219 sb.buf, strerror(save_errno));
2220 strbuf_release(&sb);
2227 * Write sha1 into the open lockfile, then close the lockfile. On
2228 * errors, rollback the lockfile, fill in *err and
2231 static int write_ref_to_lockfile(struct ref_lock *lock,
2232 const struct object_id *oid, struct strbuf *err)
2234 static char term = '\n';
2238 o = parse_object(oid);
2241 "trying to write ref '%s' with nonexistent object %s",
2242 lock->ref_name, oid_to_hex(oid));
2246 if (o->type != OBJ_COMMIT && is_branch(lock->ref_name)) {
2248 "trying to write non-commit object %s to branch '%s'",
2249 oid_to_hex(oid), lock->ref_name);
2253 fd = get_lock_file_fd(lock->lk);
2254 if (write_in_full(fd, oid_to_hex(oid), GIT_SHA1_HEXSZ) != GIT_SHA1_HEXSZ ||
2255 write_in_full(fd, &term, 1) != 1 ||
2256 close_ref(lock) < 0) {
2258 "couldn't write '%s'", get_lock_file_path(lock->lk));
2266 * Commit a change to a loose reference that has already been written
2267 * to the loose reference lockfile. Also update the reflogs if
2268 * necessary, using the specified lockmsg (which can be NULL).
2270 static int commit_ref_update(struct files_ref_store *refs,
2271 struct ref_lock *lock,
2272 const struct object_id *oid, const char *logmsg,
2275 files_assert_main_repository(refs, "commit_ref_update");
2277 clear_loose_ref_cache(refs);
2278 if (files_log_ref_write(refs, lock->ref_name,
2279 &lock->old_oid, oid,
2281 char *old_msg = strbuf_detach(err, NULL);
2282 strbuf_addf(err, "cannot update the ref '%s': %s",
2283 lock->ref_name, old_msg);
2289 if (strcmp(lock->ref_name, "HEAD") != 0) {
2291 * Special hack: If a branch is updated directly and HEAD
2292 * points to it (may happen on the remote side of a push
2293 * for example) then logically the HEAD reflog should be
2295 * A generic solution implies reverse symref information,
2296 * but finding all symrefs pointing to the given branch
2297 * would be rather costly for this rare event (the direct
2298 * update of a branch) to be worth it. So let's cheat and
2299 * check with HEAD only which should cover 99% of all usage
2300 * scenarios (even 100% of the default ones).
2302 struct object_id head_oid;
2304 const char *head_ref;
2306 head_ref = refs_resolve_ref_unsafe(&refs->base, "HEAD",
2307 RESOLVE_REF_READING,
2308 head_oid.hash, &head_flag);
2309 if (head_ref && (head_flag & REF_ISSYMREF) &&
2310 !strcmp(head_ref, lock->ref_name)) {
2311 struct strbuf log_err = STRBUF_INIT;
2312 if (files_log_ref_write(refs, "HEAD",
2313 &lock->old_oid, oid,
2314 logmsg, 0, &log_err)) {
2315 error("%s", log_err.buf);
2316 strbuf_release(&log_err);
2321 if (commit_ref(lock)) {
2322 strbuf_addf(err, "couldn't set '%s'", lock->ref_name);
2331 static int create_ref_symlink(struct ref_lock *lock, const char *target)
2334 #ifndef NO_SYMLINK_HEAD
2335 char *ref_path = get_locked_file_path(lock->lk);
2337 ret = symlink(target, ref_path);
2341 fprintf(stderr, "no symlink - falling back to symbolic ref\n");
2346 static void update_symref_reflog(struct files_ref_store *refs,
2347 struct ref_lock *lock, const char *refname,
2348 const char *target, const char *logmsg)
2350 struct strbuf err = STRBUF_INIT;
2351 struct object_id new_oid;
2353 !refs_read_ref_full(&refs->base, target,
2354 RESOLVE_REF_READING, new_oid.hash, NULL) &&
2355 files_log_ref_write(refs, refname, &lock->old_oid,
2356 &new_oid, logmsg, 0, &err)) {
2357 error("%s", err.buf);
2358 strbuf_release(&err);
2362 static int create_symref_locked(struct files_ref_store *refs,
2363 struct ref_lock *lock, const char *refname,
2364 const char *target, const char *logmsg)
2366 if (prefer_symlink_refs && !create_ref_symlink(lock, target)) {
2367 update_symref_reflog(refs, lock, refname, target, logmsg);
2371 if (!fdopen_lock_file(lock->lk, "w"))
2372 return error("unable to fdopen %s: %s",
2373 lock->lk->tempfile.filename.buf, strerror(errno));
2375 update_symref_reflog(refs, lock, refname, target, logmsg);
2377 /* no error check; commit_ref will check ferror */
2378 fprintf(lock->lk->tempfile.fp, "ref: %s\n", target);
2379 if (commit_ref(lock) < 0)
2380 return error("unable to write symref for %s: %s", refname,
2385 static int files_create_symref(struct ref_store *ref_store,
2386 const char *refname, const char *target,
2389 struct files_ref_store *refs =
2390 files_downcast(ref_store, REF_STORE_WRITE, "create_symref");
2391 struct strbuf err = STRBUF_INIT;
2392 struct ref_lock *lock;
2395 lock = lock_ref_sha1_basic(refs, refname, NULL,
2396 NULL, NULL, REF_NODEREF, NULL,
2399 error("%s", err.buf);
2400 strbuf_release(&err);
2404 ret = create_symref_locked(refs, lock, refname, target, logmsg);
2409 static int files_reflog_exists(struct ref_store *ref_store,
2410 const char *refname)
2412 struct files_ref_store *refs =
2413 files_downcast(ref_store, REF_STORE_READ, "reflog_exists");
2414 struct strbuf sb = STRBUF_INIT;
2418 files_reflog_path(refs, &sb, refname);
2419 ret = !lstat(sb.buf, &st) && S_ISREG(st.st_mode);
2420 strbuf_release(&sb);
2424 static int files_delete_reflog(struct ref_store *ref_store,
2425 const char *refname)
2427 struct files_ref_store *refs =
2428 files_downcast(ref_store, REF_STORE_WRITE, "delete_reflog");
2429 struct strbuf sb = STRBUF_INIT;
2432 files_reflog_path(refs, &sb, refname);
2433 ret = remove_path(sb.buf);
2434 strbuf_release(&sb);
2438 static int show_one_reflog_ent(struct strbuf *sb, each_reflog_ent_fn fn, void *cb_data)
2440 struct object_id ooid, noid;
2441 char *email_end, *message;
2442 timestamp_t timestamp;
2444 const char *p = sb->buf;
2446 /* old SP new SP name <email> SP time TAB msg LF */
2447 if (!sb->len || sb->buf[sb->len - 1] != '\n' ||
2448 parse_oid_hex(p, &ooid, &p) || *p++ != ' ' ||
2449 parse_oid_hex(p, &noid, &p) || *p++ != ' ' ||
2450 !(email_end = strchr(p, '>')) ||
2451 email_end[1] != ' ' ||
2452 !(timestamp = parse_timestamp(email_end + 2, &message, 10)) ||
2453 !message || message[0] != ' ' ||
2454 (message[1] != '+' && message[1] != '-') ||
2455 !isdigit(message[2]) || !isdigit(message[3]) ||
2456 !isdigit(message[4]) || !isdigit(message[5]))
2457 return 0; /* corrupt? */
2458 email_end[1] = '\0';
2459 tz = strtol(message + 1, NULL, 10);
2460 if (message[6] != '\t')
2464 return fn(&ooid, &noid, p, timestamp, tz, message, cb_data);
2467 static char *find_beginning_of_line(char *bob, char *scan)
2469 while (bob < scan && *(--scan) != '\n')
2470 ; /* keep scanning backwards */
2472 * Return either beginning of the buffer, or LF at the end of
2473 * the previous line.
2478 static int files_for_each_reflog_ent_reverse(struct ref_store *ref_store,
2479 const char *refname,
2480 each_reflog_ent_fn fn,
2483 struct files_ref_store *refs =
2484 files_downcast(ref_store, REF_STORE_READ,
2485 "for_each_reflog_ent_reverse");
2486 struct strbuf sb = STRBUF_INIT;
2489 int ret = 0, at_tail = 1;
2491 files_reflog_path(refs, &sb, refname);
2492 logfp = fopen(sb.buf, "r");
2493 strbuf_release(&sb);
2497 /* Jump to the end */
2498 if (fseek(logfp, 0, SEEK_END) < 0)
2499 ret = error("cannot seek back reflog for %s: %s",
2500 refname, strerror(errno));
2502 while (!ret && 0 < pos) {
2508 /* Fill next block from the end */
2509 cnt = (sizeof(buf) < pos) ? sizeof(buf) : pos;
2510 if (fseek(logfp, pos - cnt, SEEK_SET)) {
2511 ret = error("cannot seek back reflog for %s: %s",
2512 refname, strerror(errno));
2515 nread = fread(buf, cnt, 1, logfp);
2517 ret = error("cannot read %d bytes from reflog for %s: %s",
2518 cnt, refname, strerror(errno));
2523 scanp = endp = buf + cnt;
2524 if (at_tail && scanp[-1] == '\n')
2525 /* Looking at the final LF at the end of the file */
2529 while (buf < scanp) {
2531 * terminating LF of the previous line, or the beginning
2536 bp = find_beginning_of_line(buf, scanp);
2540 * The newline is the end of the previous line,
2541 * so we know we have complete line starting
2542 * at (bp + 1). Prefix it onto any prior data
2543 * we collected for the line and process it.
2545 strbuf_splice(&sb, 0, 0, bp + 1, endp - (bp + 1));
2548 ret = show_one_reflog_ent(&sb, fn, cb_data);
2554 * We are at the start of the buffer, and the
2555 * start of the file; there is no previous
2556 * line, and we have everything for this one.
2557 * Process it, and we can end the loop.
2559 strbuf_splice(&sb, 0, 0, buf, endp - buf);
2560 ret = show_one_reflog_ent(&sb, fn, cb_data);
2567 * We are at the start of the buffer, and there
2568 * is more file to read backwards. Which means
2569 * we are in the middle of a line. Note that we
2570 * may get here even if *bp was a newline; that
2571 * just means we are at the exact end of the
2572 * previous line, rather than some spot in the
2575 * Save away what we have to be combined with
2576 * the data from the next read.
2578 strbuf_splice(&sb, 0, 0, buf, endp - buf);
2585 die("BUG: reverse reflog parser had leftover data");
2588 strbuf_release(&sb);
2592 static int files_for_each_reflog_ent(struct ref_store *ref_store,
2593 const char *refname,
2594 each_reflog_ent_fn fn, void *cb_data)
2596 struct files_ref_store *refs =
2597 files_downcast(ref_store, REF_STORE_READ,
2598 "for_each_reflog_ent");
2600 struct strbuf sb = STRBUF_INIT;
2603 files_reflog_path(refs, &sb, refname);
2604 logfp = fopen(sb.buf, "r");
2605 strbuf_release(&sb);
2609 while (!ret && !strbuf_getwholeline(&sb, logfp, '\n'))
2610 ret = show_one_reflog_ent(&sb, fn, cb_data);
2612 strbuf_release(&sb);
2616 struct files_reflog_iterator {
2617 struct ref_iterator base;
2619 struct ref_store *ref_store;
2620 struct dir_iterator *dir_iterator;
2621 struct object_id oid;
2624 static int files_reflog_iterator_advance(struct ref_iterator *ref_iterator)
2626 struct files_reflog_iterator *iter =
2627 (struct files_reflog_iterator *)ref_iterator;
2628 struct dir_iterator *diter = iter->dir_iterator;
2631 while ((ok = dir_iterator_advance(diter)) == ITER_OK) {
2634 if (!S_ISREG(diter->st.st_mode))
2636 if (diter->basename[0] == '.')
2638 if (ends_with(diter->basename, ".lock"))
2641 if (refs_read_ref_full(iter->ref_store,
2642 diter->relative_path, 0,
2643 iter->oid.hash, &flags)) {
2644 error("bad ref for %s", diter->path.buf);
2648 iter->base.refname = diter->relative_path;
2649 iter->base.oid = &iter->oid;
2650 iter->base.flags = flags;
2654 iter->dir_iterator = NULL;
2655 if (ref_iterator_abort(ref_iterator) == ITER_ERROR)
2660 static int files_reflog_iterator_peel(struct ref_iterator *ref_iterator,
2661 struct object_id *peeled)
2663 die("BUG: ref_iterator_peel() called for reflog_iterator");
2666 static int files_reflog_iterator_abort(struct ref_iterator *ref_iterator)
2668 struct files_reflog_iterator *iter =
2669 (struct files_reflog_iterator *)ref_iterator;
2672 if (iter->dir_iterator)
2673 ok = dir_iterator_abort(iter->dir_iterator);
2675 base_ref_iterator_free(ref_iterator);
2679 static struct ref_iterator_vtable files_reflog_iterator_vtable = {
2680 files_reflog_iterator_advance,
2681 files_reflog_iterator_peel,
2682 files_reflog_iterator_abort
2685 static struct ref_iterator *files_reflog_iterator_begin(struct ref_store *ref_store)
2687 struct files_ref_store *refs =
2688 files_downcast(ref_store, REF_STORE_READ,
2689 "reflog_iterator_begin");
2690 struct files_reflog_iterator *iter = xcalloc(1, sizeof(*iter));
2691 struct ref_iterator *ref_iterator = &iter->base;
2692 struct strbuf sb = STRBUF_INIT;
2694 base_ref_iterator_init(ref_iterator, &files_reflog_iterator_vtable);
2695 files_reflog_path(refs, &sb, NULL);
2696 iter->dir_iterator = dir_iterator_begin(sb.buf);
2697 iter->ref_store = ref_store;
2698 strbuf_release(&sb);
2699 return ref_iterator;
2703 * If update is a direct update of head_ref (the reference pointed to
2704 * by HEAD), then add an extra REF_LOG_ONLY update for HEAD.
2706 static int split_head_update(struct ref_update *update,
2707 struct ref_transaction *transaction,
2708 const char *head_ref,
2709 struct string_list *affected_refnames,
2712 struct string_list_item *item;
2713 struct ref_update *new_update;
2715 if ((update->flags & REF_LOG_ONLY) ||
2716 (update->flags & REF_ISPRUNING) ||
2717 (update->flags & REF_UPDATE_VIA_HEAD))
2720 if (strcmp(update->refname, head_ref))
2724 * First make sure that HEAD is not already in the
2725 * transaction. This insertion is O(N) in the transaction
2726 * size, but it happens at most once per transaction.
2728 item = string_list_insert(affected_refnames, "HEAD");
2730 /* An entry already existed */
2732 "multiple updates for 'HEAD' (including one "
2733 "via its referent '%s') are not allowed",
2735 return TRANSACTION_NAME_CONFLICT;
2738 new_update = ref_transaction_add_update(
2739 transaction, "HEAD",
2740 update->flags | REF_LOG_ONLY | REF_NODEREF,
2741 update->new_oid.hash, update->old_oid.hash,
2744 item->util = new_update;
2750 * update is for a symref that points at referent and doesn't have
2751 * REF_NODEREF set. Split it into two updates:
2752 * - The original update, but with REF_LOG_ONLY and REF_NODEREF set
2753 * - A new, separate update for the referent reference
2754 * Note that the new update will itself be subject to splitting when
2755 * the iteration gets to it.
2757 static int split_symref_update(struct files_ref_store *refs,
2758 struct ref_update *update,
2759 const char *referent,
2760 struct ref_transaction *transaction,
2761 struct string_list *affected_refnames,
2764 struct string_list_item *item;
2765 struct ref_update *new_update;
2766 unsigned int new_flags;
2769 * First make sure that referent is not already in the
2770 * transaction. This insertion is O(N) in the transaction
2771 * size, but it happens at most once per symref in a
2774 item = string_list_insert(affected_refnames, referent);
2776 /* An entry already existed */
2778 "multiple updates for '%s' (including one "
2779 "via symref '%s') are not allowed",
2780 referent, update->refname);
2781 return TRANSACTION_NAME_CONFLICT;
2784 new_flags = update->flags;
2785 if (!strcmp(update->refname, "HEAD")) {
2787 * Record that the new update came via HEAD, so that
2788 * when we process it, split_head_update() doesn't try
2789 * to add another reflog update for HEAD. Note that
2790 * this bit will be propagated if the new_update
2791 * itself needs to be split.
2793 new_flags |= REF_UPDATE_VIA_HEAD;
2796 new_update = ref_transaction_add_update(
2797 transaction, referent, new_flags,
2798 update->new_oid.hash, update->old_oid.hash,
2801 new_update->parent_update = update;
2804 * Change the symbolic ref update to log only. Also, it
2805 * doesn't need to check its old SHA-1 value, as that will be
2806 * done when new_update is processed.
2808 update->flags |= REF_LOG_ONLY | REF_NODEREF;
2809 update->flags &= ~REF_HAVE_OLD;
2811 item->util = new_update;
2817 * Return the refname under which update was originally requested.
2819 static const char *original_update_refname(struct ref_update *update)
2821 while (update->parent_update)
2822 update = update->parent_update;
2824 return update->refname;
2828 * Check whether the REF_HAVE_OLD and old_oid values stored in update
2829 * are consistent with oid, which is the reference's current value. If
2830 * everything is OK, return 0; otherwise, write an error message to
2831 * err and return -1.
2833 static int check_old_oid(struct ref_update *update, struct object_id *oid,
2836 if (!(update->flags & REF_HAVE_OLD) ||
2837 !oidcmp(oid, &update->old_oid))
2840 if (is_null_oid(&update->old_oid))
2841 strbuf_addf(err, "cannot lock ref '%s': "
2842 "reference already exists",
2843 original_update_refname(update));
2844 else if (is_null_oid(oid))
2845 strbuf_addf(err, "cannot lock ref '%s': "
2846 "reference is missing but expected %s",
2847 original_update_refname(update),
2848 oid_to_hex(&update->old_oid));
2850 strbuf_addf(err, "cannot lock ref '%s': "
2851 "is at %s but expected %s",
2852 original_update_refname(update),
2854 oid_to_hex(&update->old_oid));
2860 * Prepare for carrying out update:
2861 * - Lock the reference referred to by update.
2862 * - Read the reference under lock.
2863 * - Check that its old SHA-1 value (if specified) is correct, and in
2864 * any case record it in update->lock->old_oid for later use when
2865 * writing the reflog.
2866 * - If it is a symref update without REF_NODEREF, split it up into a
2867 * REF_LOG_ONLY update of the symref and add a separate update for
2868 * the referent to transaction.
2869 * - If it is an update of head_ref, add a corresponding REF_LOG_ONLY
2872 static int lock_ref_for_update(struct files_ref_store *refs,
2873 struct ref_update *update,
2874 struct ref_transaction *transaction,
2875 const char *head_ref,
2876 struct string_list *affected_refnames,
2879 struct strbuf referent = STRBUF_INIT;
2880 int mustexist = (update->flags & REF_HAVE_OLD) &&
2881 !is_null_oid(&update->old_oid);
2883 struct ref_lock *lock;
2885 files_assert_main_repository(refs, "lock_ref_for_update");
2887 if ((update->flags & REF_HAVE_NEW) && is_null_oid(&update->new_oid))
2888 update->flags |= REF_DELETING;
2891 ret = split_head_update(update, transaction, head_ref,
2892 affected_refnames, err);
2897 ret = lock_raw_ref(refs, update->refname, mustexist,
2898 affected_refnames, NULL,
2900 &update->type, err);
2904 reason = strbuf_detach(err, NULL);
2905 strbuf_addf(err, "cannot lock ref '%s': %s",
2906 original_update_refname(update), reason);
2911 update->backend_data = lock;
2913 if (update->type & REF_ISSYMREF) {
2914 if (update->flags & REF_NODEREF) {
2916 * We won't be reading the referent as part of
2917 * the transaction, so we have to read it here
2918 * to record and possibly check old_sha1:
2920 if (refs_read_ref_full(&refs->base,
2922 lock->old_oid.hash, NULL)) {
2923 if (update->flags & REF_HAVE_OLD) {
2924 strbuf_addf(err, "cannot lock ref '%s': "
2925 "error reading reference",
2926 original_update_refname(update));
2929 } else if (check_old_oid(update, &lock->old_oid, err)) {
2930 return TRANSACTION_GENERIC_ERROR;
2934 * Create a new update for the reference this
2935 * symref is pointing at. Also, we will record
2936 * and verify old_sha1 for this update as part
2937 * of processing the split-off update, so we
2938 * don't have to do it here.
2940 ret = split_symref_update(refs, update,
2941 referent.buf, transaction,
2942 affected_refnames, err);
2947 struct ref_update *parent_update;
2949 if (check_old_oid(update, &lock->old_oid, err))
2950 return TRANSACTION_GENERIC_ERROR;
2953 * If this update is happening indirectly because of a
2954 * symref update, record the old SHA-1 in the parent
2957 for (parent_update = update->parent_update;
2959 parent_update = parent_update->parent_update) {
2960 struct ref_lock *parent_lock = parent_update->backend_data;
2961 oidcpy(&parent_lock->old_oid, &lock->old_oid);
2965 if ((update->flags & REF_HAVE_NEW) &&
2966 !(update->flags & REF_DELETING) &&
2967 !(update->flags & REF_LOG_ONLY)) {
2968 if (!(update->type & REF_ISSYMREF) &&
2969 !oidcmp(&lock->old_oid, &update->new_oid)) {
2971 * The reference already has the desired
2972 * value, so we don't need to write it.
2974 } else if (write_ref_to_lockfile(lock, &update->new_oid,
2976 char *write_err = strbuf_detach(err, NULL);
2979 * The lock was freed upon failure of
2980 * write_ref_to_lockfile():
2982 update->backend_data = NULL;
2984 "cannot update ref '%s': %s",
2985 update->refname, write_err);
2987 return TRANSACTION_GENERIC_ERROR;
2989 update->flags |= REF_NEEDS_COMMIT;
2992 if (!(update->flags & REF_NEEDS_COMMIT)) {
2994 * We didn't call write_ref_to_lockfile(), so
2995 * the lockfile is still open. Close it to
2996 * free up the file descriptor:
2998 if (close_ref(lock)) {
2999 strbuf_addf(err, "couldn't close '%s.lock'",
3001 return TRANSACTION_GENERIC_ERROR;
3008 * Unlock any references in `transaction` that are still locked, and
3009 * mark the transaction closed.
3011 static void files_transaction_cleanup(struct ref_transaction *transaction)
3015 for (i = 0; i < transaction->nr; i++) {
3016 struct ref_update *update = transaction->updates[i];
3017 struct ref_lock *lock = update->backend_data;
3021 update->backend_data = NULL;
3025 transaction->state = REF_TRANSACTION_CLOSED;
3028 static int files_transaction_prepare(struct ref_store *ref_store,
3029 struct ref_transaction *transaction,
3032 struct files_ref_store *refs =
3033 files_downcast(ref_store, REF_STORE_WRITE,
3034 "ref_transaction_prepare");
3037 struct string_list affected_refnames = STRING_LIST_INIT_NODUP;
3038 char *head_ref = NULL;
3040 struct object_id head_oid;
3044 if (!transaction->nr)
3048 * Fail if a refname appears more than once in the
3049 * transaction. (If we end up splitting up any updates using
3050 * split_symref_update() or split_head_update(), those
3051 * functions will check that the new updates don't have the
3052 * same refname as any existing ones.)
3054 for (i = 0; i < transaction->nr; i++) {
3055 struct ref_update *update = transaction->updates[i];
3056 struct string_list_item *item =
3057 string_list_append(&affected_refnames, update->refname);
3060 * We store a pointer to update in item->util, but at
3061 * the moment we never use the value of this field
3062 * except to check whether it is non-NULL.
3064 item->util = update;
3066 string_list_sort(&affected_refnames);
3067 if (ref_update_reject_duplicates(&affected_refnames, err)) {
3068 ret = TRANSACTION_GENERIC_ERROR;
3073 * Special hack: If a branch is updated directly and HEAD
3074 * points to it (may happen on the remote side of a push
3075 * for example) then logically the HEAD reflog should be
3078 * A generic solution would require reverse symref lookups,
3079 * but finding all symrefs pointing to a given branch would be
3080 * rather costly for this rare event (the direct update of a
3081 * branch) to be worth it. So let's cheat and check with HEAD
3082 * only, which should cover 99% of all usage scenarios (even
3083 * 100% of the default ones).
3085 * So if HEAD is a symbolic reference, then record the name of
3086 * the reference that it points to. If we see an update of
3087 * head_ref within the transaction, then split_head_update()
3088 * arranges for the reflog of HEAD to be updated, too.
3090 head_ref = refs_resolve_refdup(ref_store, "HEAD",
3091 RESOLVE_REF_NO_RECURSE,
3092 head_oid.hash, &head_type);
3094 if (head_ref && !(head_type & REF_ISSYMREF)) {
3100 * Acquire all locks, verify old values if provided, check
3101 * that new values are valid, and write new values to the
3102 * lockfiles, ready to be activated. Only keep one lockfile
3103 * open at a time to avoid running out of file descriptors.
3104 * Note that lock_ref_for_update() might append more updates
3105 * to the transaction.
3107 for (i = 0; i < transaction->nr; i++) {
3108 struct ref_update *update = transaction->updates[i];
3110 ret = lock_ref_for_update(refs, update, transaction,
3111 head_ref, &affected_refnames, err);
3118 string_list_clear(&affected_refnames, 0);
3121 files_transaction_cleanup(transaction);
3123 transaction->state = REF_TRANSACTION_PREPARED;
3128 static int files_transaction_finish(struct ref_store *ref_store,
3129 struct ref_transaction *transaction,
3132 struct files_ref_store *refs =
3133 files_downcast(ref_store, 0, "ref_transaction_finish");
3136 struct string_list refs_to_delete = STRING_LIST_INIT_NODUP;
3137 struct string_list_item *ref_to_delete;
3138 struct strbuf sb = STRBUF_INIT;
3142 if (!transaction->nr) {
3143 transaction->state = REF_TRANSACTION_CLOSED;
3147 /* Perform updates first so live commits remain referenced */
3148 for (i = 0; i < transaction->nr; i++) {
3149 struct ref_update *update = transaction->updates[i];
3150 struct ref_lock *lock = update->backend_data;
3152 if (update->flags & REF_NEEDS_COMMIT ||
3153 update->flags & REF_LOG_ONLY) {
3154 if (files_log_ref_write(refs,
3158 update->msg, update->flags,
3160 char *old_msg = strbuf_detach(err, NULL);
3162 strbuf_addf(err, "cannot update the ref '%s': %s",
3163 lock->ref_name, old_msg);
3166 update->backend_data = NULL;
3167 ret = TRANSACTION_GENERIC_ERROR;
3171 if (update->flags & REF_NEEDS_COMMIT) {
3172 clear_loose_ref_cache(refs);
3173 if (commit_ref(lock)) {
3174 strbuf_addf(err, "couldn't set '%s'", lock->ref_name);
3176 update->backend_data = NULL;
3177 ret = TRANSACTION_GENERIC_ERROR;
3182 /* Perform deletes now that updates are safely completed */
3183 for (i = 0; i < transaction->nr; i++) {
3184 struct ref_update *update = transaction->updates[i];
3185 struct ref_lock *lock = update->backend_data;
3187 if (update->flags & REF_DELETING &&
3188 !(update->flags & REF_LOG_ONLY)) {
3189 if (!(update->type & REF_ISPACKED) ||
3190 update->type & REF_ISSYMREF) {
3191 /* It is a loose reference. */
3193 files_ref_path(refs, &sb, lock->ref_name);
3194 if (unlink_or_msg(sb.buf, err)) {
3195 ret = TRANSACTION_GENERIC_ERROR;
3198 update->flags |= REF_DELETED_LOOSE;
3201 if (!(update->flags & REF_ISPRUNING))
3202 string_list_append(&refs_to_delete,
3207 if (repack_without_refs(refs->packed_ref_store, &refs_to_delete, err)) {
3208 ret = TRANSACTION_GENERIC_ERROR;
3212 /* Delete the reflogs of any references that were deleted: */
3213 for_each_string_list_item(ref_to_delete, &refs_to_delete) {
3215 files_reflog_path(refs, &sb, ref_to_delete->string);
3216 if (!unlink_or_warn(sb.buf))
3217 try_remove_empty_parents(refs, ref_to_delete->string,
3218 REMOVE_EMPTY_PARENTS_REFLOG);
3221 clear_loose_ref_cache(refs);
3224 files_transaction_cleanup(transaction);
3226 for (i = 0; i < transaction->nr; i++) {
3227 struct ref_update *update = transaction->updates[i];
3229 if (update->flags & REF_DELETED_LOOSE) {
3231 * The loose reference was deleted. Delete any
3232 * empty parent directories. (Note that this
3233 * can only work because we have already
3234 * removed the lockfile.)
3236 try_remove_empty_parents(refs, update->refname,
3237 REMOVE_EMPTY_PARENTS_REF);
3241 strbuf_release(&sb);
3242 string_list_clear(&refs_to_delete, 0);
3246 static int files_transaction_abort(struct ref_store *ref_store,
3247 struct ref_transaction *transaction,
3250 files_transaction_cleanup(transaction);
3254 static int ref_present(const char *refname,
3255 const struct object_id *oid, int flags, void *cb_data)
3257 struct string_list *affected_refnames = cb_data;
3259 return string_list_has_string(affected_refnames, refname);
3262 static int files_initial_transaction_commit(struct ref_store *ref_store,
3263 struct ref_transaction *transaction,
3266 struct files_ref_store *refs =
3267 files_downcast(ref_store, REF_STORE_WRITE,
3268 "initial_ref_transaction_commit");
3271 struct string_list affected_refnames = STRING_LIST_INIT_NODUP;
3275 if (transaction->state != REF_TRANSACTION_OPEN)
3276 die("BUG: commit called for transaction that is not open");
3278 /* Fail if a refname appears more than once in the transaction: */
3279 for (i = 0; i < transaction->nr; i++)
3280 string_list_append(&affected_refnames,
3281 transaction->updates[i]->refname);
3282 string_list_sort(&affected_refnames);
3283 if (ref_update_reject_duplicates(&affected_refnames, err)) {
3284 ret = TRANSACTION_GENERIC_ERROR;
3289 * It's really undefined to call this function in an active
3290 * repository or when there are existing references: we are
3291 * only locking and changing packed-refs, so (1) any
3292 * simultaneous processes might try to change a reference at
3293 * the same time we do, and (2) any existing loose versions of
3294 * the references that we are setting would have precedence
3295 * over our values. But some remote helpers create the remote
3296 * "HEAD" and "master" branches before calling this function,
3297 * so here we really only check that none of the references
3298 * that we are creating already exists.
3300 if (refs_for_each_rawref(&refs->base, ref_present,
3301 &affected_refnames))
3302 die("BUG: initial ref transaction called with existing refs");
3304 for (i = 0; i < transaction->nr; i++) {
3305 struct ref_update *update = transaction->updates[i];
3307 if ((update->flags & REF_HAVE_OLD) &&
3308 !is_null_oid(&update->old_oid))
3309 die("BUG: initial ref transaction with old_sha1 set");
3310 if (refs_verify_refname_available(&refs->base, update->refname,
3311 &affected_refnames, NULL,
3313 ret = TRANSACTION_NAME_CONFLICT;
3318 if (lock_packed_refs(refs->packed_ref_store, 0)) {
3319 strbuf_addf(err, "unable to lock packed-refs file: %s",
3321 ret = TRANSACTION_GENERIC_ERROR;
3325 for (i = 0; i < transaction->nr; i++) {
3326 struct ref_update *update = transaction->updates[i];
3328 if ((update->flags & REF_HAVE_NEW) &&
3329 !is_null_oid(&update->new_oid))
3330 add_packed_ref(refs->packed_ref_store, update->refname,
3334 if (commit_packed_refs(refs->packed_ref_store)) {
3335 strbuf_addf(err, "unable to commit packed-refs file: %s",
3337 ret = TRANSACTION_GENERIC_ERROR;
3342 transaction->state = REF_TRANSACTION_CLOSED;
3343 string_list_clear(&affected_refnames, 0);
3347 struct expire_reflog_cb {
3349 reflog_expiry_should_prune_fn *should_prune_fn;
3352 struct object_id last_kept_oid;
3355 static int expire_reflog_ent(struct object_id *ooid, struct object_id *noid,
3356 const char *email, timestamp_t timestamp, int tz,
3357 const char *message, void *cb_data)
3359 struct expire_reflog_cb *cb = cb_data;
3360 struct expire_reflog_policy_cb *policy_cb = cb->policy_cb;
3362 if (cb->flags & EXPIRE_REFLOGS_REWRITE)
3363 ooid = &cb->last_kept_oid;
3365 if ((*cb->should_prune_fn)(ooid, noid, email, timestamp, tz,
3366 message, policy_cb)) {
3368 printf("would prune %s", message);
3369 else if (cb->flags & EXPIRE_REFLOGS_VERBOSE)
3370 printf("prune %s", message);
3373 fprintf(cb->newlog, "%s %s %s %"PRItime" %+05d\t%s",
3374 oid_to_hex(ooid), oid_to_hex(noid),
3375 email, timestamp, tz, message);
3376 oidcpy(&cb->last_kept_oid, noid);
3378 if (cb->flags & EXPIRE_REFLOGS_VERBOSE)
3379 printf("keep %s", message);
3384 static int files_reflog_expire(struct ref_store *ref_store,
3385 const char *refname, const unsigned char *sha1,
3387 reflog_expiry_prepare_fn prepare_fn,
3388 reflog_expiry_should_prune_fn should_prune_fn,
3389 reflog_expiry_cleanup_fn cleanup_fn,
3390 void *policy_cb_data)
3392 struct files_ref_store *refs =
3393 files_downcast(ref_store, REF_STORE_WRITE, "reflog_expire");
3394 static struct lock_file reflog_lock;
3395 struct expire_reflog_cb cb;
3396 struct ref_lock *lock;
3397 struct strbuf log_file_sb = STRBUF_INIT;
3401 struct strbuf err = STRBUF_INIT;
3402 struct object_id oid;
3404 memset(&cb, 0, sizeof(cb));
3406 cb.policy_cb = policy_cb_data;
3407 cb.should_prune_fn = should_prune_fn;
3410 * The reflog file is locked by holding the lock on the
3411 * reference itself, plus we might need to update the
3412 * reference if --updateref was specified:
3414 lock = lock_ref_sha1_basic(refs, refname, sha1,
3415 NULL, NULL, REF_NODEREF,
3418 error("cannot lock ref '%s': %s", refname, err.buf);
3419 strbuf_release(&err);
3422 if (!refs_reflog_exists(ref_store, refname)) {
3427 files_reflog_path(refs, &log_file_sb, refname);
3428 log_file = strbuf_detach(&log_file_sb, NULL);
3429 if (!(flags & EXPIRE_REFLOGS_DRY_RUN)) {
3431 * Even though holding $GIT_DIR/logs/$reflog.lock has
3432 * no locking implications, we use the lock_file
3433 * machinery here anyway because it does a lot of the
3434 * work we need, including cleaning up if the program
3435 * exits unexpectedly.
3437 if (hold_lock_file_for_update(&reflog_lock, log_file, 0) < 0) {
3438 struct strbuf err = STRBUF_INIT;
3439 unable_to_lock_message(log_file, errno, &err);
3440 error("%s", err.buf);
3441 strbuf_release(&err);
3444 cb.newlog = fdopen_lock_file(&reflog_lock, "w");
3446 error("cannot fdopen %s (%s)",
3447 get_lock_file_path(&reflog_lock), strerror(errno));
3452 hashcpy(oid.hash, sha1);
3454 (*prepare_fn)(refname, &oid, cb.policy_cb);
3455 refs_for_each_reflog_ent(ref_store, refname, expire_reflog_ent, &cb);
3456 (*cleanup_fn)(cb.policy_cb);
3458 if (!(flags & EXPIRE_REFLOGS_DRY_RUN)) {
3460 * It doesn't make sense to adjust a reference pointed
3461 * to by a symbolic ref based on expiring entries in
3462 * the symbolic reference's reflog. Nor can we update
3463 * a reference if there are no remaining reflog
3466 int update = (flags & EXPIRE_REFLOGS_UPDATE_REF) &&
3467 !(type & REF_ISSYMREF) &&
3468 !is_null_oid(&cb.last_kept_oid);
3470 if (close_lock_file(&reflog_lock)) {
3471 status |= error("couldn't write %s: %s", log_file,
3473 } else if (update &&
3474 (write_in_full(get_lock_file_fd(lock->lk),
3475 oid_to_hex(&cb.last_kept_oid), GIT_SHA1_HEXSZ) != GIT_SHA1_HEXSZ ||
3476 write_str_in_full(get_lock_file_fd(lock->lk), "\n") != 1 ||
3477 close_ref(lock) < 0)) {
3478 status |= error("couldn't write %s",
3479 get_lock_file_path(lock->lk));
3480 rollback_lock_file(&reflog_lock);
3481 } else if (commit_lock_file(&reflog_lock)) {
3482 status |= error("unable to write reflog '%s' (%s)",
3483 log_file, strerror(errno));
3484 } else if (update && commit_ref(lock)) {
3485 status |= error("couldn't set %s", lock->ref_name);
3493 rollback_lock_file(&reflog_lock);
3499 static int files_init_db(struct ref_store *ref_store, struct strbuf *err)
3501 struct files_ref_store *refs =
3502 files_downcast(ref_store, REF_STORE_WRITE, "init_db");
3503 struct strbuf sb = STRBUF_INIT;
3506 * Create .git/refs/{heads,tags}
3508 files_ref_path(refs, &sb, "refs/heads");
3509 safe_create_dir(sb.buf, 1);
3512 files_ref_path(refs, &sb, "refs/tags");
3513 safe_create_dir(sb.buf, 1);
3515 strbuf_release(&sb);
3519 struct ref_storage_be refs_be_files = {
3522 files_ref_store_create,
3524 files_transaction_prepare,
3525 files_transaction_finish,
3526 files_transaction_abort,
3527 files_initial_transaction_commit,
3531 files_create_symref,
3535 files_ref_iterator_begin,
3538 files_reflog_iterator_begin,
3539 files_for_each_reflog_ent,
3540 files_for_each_reflog_ent_reverse,
3541 files_reflog_exists,
3542 files_create_reflog,
3543 files_delete_reflog,