mirror of
https://github.com/git/git.git
synced 2024-12-01 05:54:16 +08:00
86829f3f3e
Unlike "git add" and other end-user facing commands, where it is diagnosed as an error to give a pathspec with an element that does not match any path, the diff machinery does not care if some elements of the pathspec do not match. Given that the diff machinery is heavily used in pathspec-limited "git log" machinery, and it is common for a path to come and go while traversing the project history, this is usually a good thing. However, in some cases we would want to know if all the pathspec elements matched. For example, "git add -u <pathspec>" internally uses the machinery used by "git diff-files" to decide contents from what paths to add to the index, and as an end-user facing command, "git add -u" would want to report an unmatched pathspec element. Add a new .ps_matched member next to the .prune_data member in "struct rev_info" so that we can optionally keep track of the use of .prune_data pathspec elements that can be inspected by the caller. Signed-off-by: Junio C Hamano <gitster@pobox.com>
490 lines
17 KiB
C
490 lines
17 KiB
C
#ifndef READ_CACHE_LL_H
|
|
#define READ_CACHE_LL_H
|
|
|
|
#include "hash-ll.h"
|
|
#include "hashmap.h"
|
|
#include "statinfo.h"
|
|
|
|
/*
|
|
* Basic data structures for the directory cache
|
|
*/
|
|
|
|
#define CACHE_SIGNATURE 0x44495243 /* "DIRC" */
|
|
struct cache_header {
|
|
uint32_t hdr_signature;
|
|
uint32_t hdr_version;
|
|
uint32_t hdr_entries;
|
|
};
|
|
|
|
#define INDEX_FORMAT_LB 2
|
|
#define INDEX_FORMAT_UB 4
|
|
|
|
struct cache_entry {
|
|
struct hashmap_entry ent;
|
|
struct stat_data ce_stat_data;
|
|
unsigned int ce_mode;
|
|
unsigned int ce_flags;
|
|
unsigned int mem_pool_allocated;
|
|
unsigned int ce_namelen;
|
|
unsigned int index; /* for link extension */
|
|
struct object_id oid;
|
|
char name[FLEX_ARRAY]; /* more */
|
|
};
|
|
|
|
#define CE_STAGEMASK (0x3000)
|
|
#define CE_EXTENDED (0x4000)
|
|
#define CE_VALID (0x8000)
|
|
#define CE_STAGESHIFT 12
|
|
|
|
/*
|
|
* Range 0xFFFF0FFF in ce_flags is divided into
|
|
* two parts: in-memory flags and on-disk ones.
|
|
* Flags in CE_EXTENDED_FLAGS will get saved on-disk
|
|
* if you want to save a new flag, add it in
|
|
* CE_EXTENDED_FLAGS
|
|
*
|
|
* In-memory only flags
|
|
*/
|
|
#define CE_UPDATE (1 << 16)
|
|
#define CE_REMOVE (1 << 17)
|
|
#define CE_UPTODATE (1 << 18)
|
|
#define CE_ADDED (1 << 19)
|
|
|
|
#define CE_HASHED (1 << 20)
|
|
#define CE_FSMONITOR_VALID (1 << 21)
|
|
#define CE_WT_REMOVE (1 << 22) /* remove in work directory */
|
|
#define CE_CONFLICTED (1 << 23)
|
|
|
|
#define CE_UNPACKED (1 << 24)
|
|
#define CE_NEW_SKIP_WORKTREE (1 << 25)
|
|
|
|
/* used to temporarily mark paths matched by pathspecs */
|
|
#define CE_MATCHED (1 << 26)
|
|
|
|
#define CE_UPDATE_IN_BASE (1 << 27)
|
|
#define CE_STRIP_NAME (1 << 28)
|
|
|
|
/*
|
|
* Extended on-disk flags
|
|
*/
|
|
#define CE_INTENT_TO_ADD (1 << 29)
|
|
#define CE_SKIP_WORKTREE (1 << 30)
|
|
/* CE_EXTENDED2 is for future extension */
|
|
#define CE_EXTENDED2 (1U << 31)
|
|
|
|
#define CE_EXTENDED_FLAGS (CE_INTENT_TO_ADD | CE_SKIP_WORKTREE)
|
|
|
|
/*
|
|
* Safeguard to avoid saving wrong flags:
|
|
* - CE_EXTENDED2 won't get saved until its semantic is known
|
|
* - Bits in 0x0000FFFF have been saved in ce_flags already
|
|
* - Bits in 0x003F0000 are currently in-memory flags
|
|
*/
|
|
#if CE_EXTENDED_FLAGS & 0x803FFFFF
|
|
#error "CE_EXTENDED_FLAGS out of range"
|
|
#endif
|
|
|
|
/* Forward structure decls */
|
|
struct pathspec;
|
|
struct tree;
|
|
|
|
/*
|
|
* Copy the sha1 and stat state of a cache entry from one to
|
|
* another. But we never change the name, or the hash state!
|
|
*/
|
|
static inline void copy_cache_entry(struct cache_entry *dst,
|
|
const struct cache_entry *src)
|
|
{
|
|
unsigned int state = dst->ce_flags & CE_HASHED;
|
|
int mem_pool_allocated = dst->mem_pool_allocated;
|
|
|
|
/* Don't copy hash chain and name */
|
|
memcpy(&dst->ce_stat_data, &src->ce_stat_data,
|
|
offsetof(struct cache_entry, name) -
|
|
offsetof(struct cache_entry, ce_stat_data));
|
|
|
|
/* Restore the hash state */
|
|
dst->ce_flags = (dst->ce_flags & ~CE_HASHED) | state;
|
|
|
|
/* Restore the mem_pool_allocated flag */
|
|
dst->mem_pool_allocated = mem_pool_allocated;
|
|
}
|
|
|
|
static inline unsigned create_ce_flags(unsigned stage)
|
|
{
|
|
return (stage << CE_STAGESHIFT);
|
|
}
|
|
|
|
#define ce_namelen(ce) ((ce)->ce_namelen)
|
|
#define ce_size(ce) cache_entry_size(ce_namelen(ce))
|
|
#define ce_stage(ce) ((CE_STAGEMASK & (ce)->ce_flags) >> CE_STAGESHIFT)
|
|
#define ce_uptodate(ce) ((ce)->ce_flags & CE_UPTODATE)
|
|
#define ce_skip_worktree(ce) ((ce)->ce_flags & CE_SKIP_WORKTREE)
|
|
#define ce_mark_uptodate(ce) ((ce)->ce_flags |= CE_UPTODATE)
|
|
#define ce_intent_to_add(ce) ((ce)->ce_flags & CE_INTENT_TO_ADD)
|
|
|
|
#define cache_entry_size(len) (offsetof(struct cache_entry,name) + (len) + 1)
|
|
|
|
#define SOMETHING_CHANGED (1 << 0) /* unclassified changes go here */
|
|
#define CE_ENTRY_CHANGED (1 << 1)
|
|
#define CE_ENTRY_REMOVED (1 << 2)
|
|
#define CE_ENTRY_ADDED (1 << 3)
|
|
#define RESOLVE_UNDO_CHANGED (1 << 4)
|
|
#define CACHE_TREE_CHANGED (1 << 5)
|
|
#define SPLIT_INDEX_ORDERED (1 << 6)
|
|
#define UNTRACKED_CHANGED (1 << 7)
|
|
#define FSMONITOR_CHANGED (1 << 8)
|
|
|
|
struct split_index;
|
|
struct untracked_cache;
|
|
struct progress;
|
|
struct pattern_list;
|
|
|
|
enum sparse_index_mode {
|
|
/*
|
|
* There are no sparse directories in the index at all.
|
|
*
|
|
* Repositories that don't use cone-mode sparse-checkout will
|
|
* always have their indexes in this mode.
|
|
*/
|
|
INDEX_EXPANDED = 0,
|
|
|
|
/*
|
|
* The index has already been collapsed to sparse directories
|
|
* whereever possible.
|
|
*/
|
|
INDEX_COLLAPSED,
|
|
|
|
/*
|
|
* The sparse directories that exist are outside the
|
|
* sparse-checkout boundary, but it is possible that some file
|
|
* entries could collapse to sparse directory entries.
|
|
*/
|
|
INDEX_PARTIALLY_SPARSE,
|
|
};
|
|
|
|
struct index_state {
|
|
struct cache_entry **cache;
|
|
unsigned int version;
|
|
unsigned int cache_nr, cache_alloc, cache_changed;
|
|
struct string_list *resolve_undo;
|
|
struct cache_tree *cache_tree;
|
|
struct split_index *split_index;
|
|
struct cache_time timestamp;
|
|
unsigned name_hash_initialized : 1,
|
|
initialized : 1,
|
|
drop_cache_tree : 1,
|
|
updated_workdir : 1,
|
|
updated_skipworktree : 1,
|
|
fsmonitor_has_run_once : 1;
|
|
enum sparse_index_mode sparse_index;
|
|
struct hashmap name_hash;
|
|
struct hashmap dir_hash;
|
|
struct object_id oid;
|
|
struct untracked_cache *untracked;
|
|
char *fsmonitor_last_update;
|
|
struct ewah_bitmap *fsmonitor_dirty;
|
|
struct mem_pool *ce_mem_pool;
|
|
struct progress *progress;
|
|
struct repository *repo;
|
|
struct pattern_list *sparse_checkout_patterns;
|
|
};
|
|
|
|
/**
|
|
* A "struct index_state istate" must be initialized with
|
|
* INDEX_STATE_INIT or the corresponding index_state_init().
|
|
*
|
|
* If the variable won't be used again, use release_index() to free()
|
|
* its resources. If it needs to be used again use discard_index(),
|
|
* which does the same thing, but will use use index_state_init() at
|
|
* the end. The discard_index() will use its own "istate->repo" as the
|
|
* "r" argument to index_state_init() in that case.
|
|
*/
|
|
#define INDEX_STATE_INIT(r) { \
|
|
.repo = (r), \
|
|
}
|
|
void index_state_init(struct index_state *istate, struct repository *r);
|
|
void release_index(struct index_state *istate);
|
|
|
|
/* Cache entry creation and cleanup */
|
|
|
|
/*
|
|
* Create cache_entry intended for use in the specified index. Caller
|
|
* is responsible for discarding the cache_entry with
|
|
* `discard_cache_entry`.
|
|
*/
|
|
struct cache_entry *make_cache_entry(struct index_state *istate,
|
|
unsigned int mode,
|
|
const struct object_id *oid,
|
|
const char *path,
|
|
int stage,
|
|
unsigned int refresh_options);
|
|
|
|
struct cache_entry *make_empty_cache_entry(struct index_state *istate,
|
|
size_t name_len);
|
|
|
|
/*
|
|
* Create a cache_entry that is not intended to be added to an index. If
|
|
* `ce_mem_pool` is not NULL, the entry is allocated within the given memory
|
|
* pool. Caller is responsible for discarding "loose" entries with
|
|
* `discard_cache_entry()` and the memory pool with
|
|
* `mem_pool_discard(ce_mem_pool, should_validate_cache_entries())`.
|
|
*/
|
|
struct cache_entry *make_transient_cache_entry(unsigned int mode,
|
|
const struct object_id *oid,
|
|
const char *path,
|
|
int stage,
|
|
struct mem_pool *ce_mem_pool);
|
|
|
|
struct cache_entry *make_empty_transient_cache_entry(size_t len,
|
|
struct mem_pool *ce_mem_pool);
|
|
|
|
/*
|
|
* Discard cache entry.
|
|
*/
|
|
void discard_cache_entry(struct cache_entry *ce);
|
|
|
|
/*
|
|
* Check configuration if we should perform extra validation on cache
|
|
* entries.
|
|
*/
|
|
int should_validate_cache_entries(void);
|
|
|
|
/*
|
|
* Duplicate a cache_entry. Allocate memory for the new entry from a
|
|
* memory_pool. Takes into account cache_entry fields that are meant
|
|
* for managing the underlying memory allocation of the cache_entry.
|
|
*/
|
|
struct cache_entry *dup_cache_entry(const struct cache_entry *ce, struct index_state *istate);
|
|
|
|
/*
|
|
* Validate the cache entries in the index. This is an internal
|
|
* consistency check that the cache_entry structs are allocated from
|
|
* the expected memory pool.
|
|
*/
|
|
void validate_cache_entries(const struct index_state *istate);
|
|
|
|
/*
|
|
* Bulk prefetch all missing cache entries that are not GITLINKs and that match
|
|
* the given predicate. This function should only be called if
|
|
* repo_has_promisor_remote() returns true.
|
|
*/
|
|
typedef int (*must_prefetch_predicate)(const struct cache_entry *);
|
|
void prefetch_cache_entries(const struct index_state *istate,
|
|
must_prefetch_predicate must_prefetch);
|
|
|
|
/* Initialize and use the cache information */
|
|
struct lock_file;
|
|
int do_read_index(struct index_state *istate, const char *path,
|
|
int must_exist); /* for testting only! */
|
|
int read_index_from(struct index_state *, const char *path,
|
|
const char *gitdir);
|
|
int is_index_unborn(struct index_state *);
|
|
|
|
/* For use with `write_locked_index()`. */
|
|
#define COMMIT_LOCK (1 << 0)
|
|
#define SKIP_IF_UNCHANGED (1 << 1)
|
|
|
|
/*
|
|
* Write the index while holding an already-taken lock. Close the lock,
|
|
* and if `COMMIT_LOCK` is given, commit it.
|
|
*
|
|
* Unless a split index is in use, write the index into the lockfile.
|
|
*
|
|
* With a split index, write the shared index to a temporary file,
|
|
* adjust its permissions and rename it into place, then write the
|
|
* split index to the lockfile. If the temporary file for the shared
|
|
* index cannot be created, fall back to the behavior described in
|
|
* the previous paragraph.
|
|
*
|
|
* With `COMMIT_LOCK`, the lock is always committed or rolled back.
|
|
* Without it, the lock is closed, but neither committed nor rolled
|
|
* back.
|
|
*
|
|
* If `SKIP_IF_UNCHANGED` is given and the index is unchanged, nothing
|
|
* is written (and the lock is rolled back if `COMMIT_LOCK` is given).
|
|
*/
|
|
int write_locked_index(struct index_state *, struct lock_file *lock, unsigned flags);
|
|
|
|
void discard_index(struct index_state *);
|
|
void move_index_extensions(struct index_state *dst, struct index_state *src);
|
|
int unmerged_index(const struct index_state *);
|
|
|
|
/**
|
|
* Returns 1 if istate differs from tree, 0 otherwise. If tree is NULL,
|
|
* compares istate to HEAD. If tree is NULL and on an unborn branch,
|
|
* returns 1 if there are entries in istate, 0 otherwise. If an strbuf is
|
|
* provided, the space-separated list of files that differ will be appended
|
|
* to it.
|
|
*/
|
|
int repo_index_has_changes(struct repository *repo,
|
|
struct tree *tree,
|
|
struct strbuf *sb);
|
|
|
|
int verify_path(const char *path, unsigned mode);
|
|
int strcmp_offset(const char *s1, const char *s2, size_t *first_change);
|
|
|
|
/*
|
|
* Searches for an entry defined by name and namelen in the given index.
|
|
* If the return value is positive (including 0) it is the position of an
|
|
* exact match. If the return value is negative, the negated value minus 1
|
|
* is the position where the entry would be inserted.
|
|
* Example: The current index consists of these files and its stages:
|
|
*
|
|
* b#0, d#0, f#1, f#3
|
|
*
|
|
* index_name_pos(&index, "a", 1) -> -1
|
|
* index_name_pos(&index, "b", 1) -> 0
|
|
* index_name_pos(&index, "c", 1) -> -2
|
|
* index_name_pos(&index, "d", 1) -> 1
|
|
* index_name_pos(&index, "e", 1) -> -3
|
|
* index_name_pos(&index, "f", 1) -> -3
|
|
* index_name_pos(&index, "g", 1) -> -5
|
|
*/
|
|
int index_name_pos(struct index_state *, const char *name, int namelen);
|
|
|
|
/*
|
|
* Like index_name_pos, returns the position of an entry of the given name in
|
|
* the index if one exists, otherwise returns a negative value where the negated
|
|
* value minus 1 is the position where the index entry would be inserted. Unlike
|
|
* index_name_pos, however, a sparse index is not expanded to find an entry
|
|
* inside a sparse directory.
|
|
*/
|
|
int index_name_pos_sparse(struct index_state *, const char *name, int namelen);
|
|
|
|
/*
|
|
* Determines whether an entry with the given name exists within the
|
|
* given index. The return value is 1 if an exact match is found, otherwise
|
|
* it is 0. Note that, unlike index_name_pos, this function does not expand
|
|
* the index if it is sparse. If an item exists within the full index but it
|
|
* is contained within a sparse directory (and not in the sparse index), 0 is
|
|
* returned.
|
|
*/
|
|
int index_entry_exists(struct index_state *, const char *name, int namelen);
|
|
|
|
/*
|
|
* Some functions return the negative complement of an insert position when a
|
|
* precise match was not found but a position was found where the entry would
|
|
* need to be inserted. This helper protects that logic from any integer
|
|
* underflow.
|
|
*/
|
|
static inline int index_pos_to_insert_pos(uintmax_t pos)
|
|
{
|
|
if (pos > INT_MAX)
|
|
die("overflow: -1 - %"PRIuMAX, pos);
|
|
return -1 - (int)pos;
|
|
}
|
|
|
|
#define ADD_CACHE_OK_TO_ADD 1 /* Ok to add */
|
|
#define ADD_CACHE_OK_TO_REPLACE 2 /* Ok to replace file/directory */
|
|
#define ADD_CACHE_SKIP_DFCHECK 4 /* Ok to skip DF conflict checks */
|
|
#define ADD_CACHE_JUST_APPEND 8 /* Append only */
|
|
#define ADD_CACHE_NEW_ONLY 16 /* Do not replace existing ones */
|
|
#define ADD_CACHE_KEEP_CACHE_TREE 32 /* Do not invalidate cache-tree */
|
|
#define ADD_CACHE_RENORMALIZE 64 /* Pass along HASH_RENORMALIZE */
|
|
int add_index_entry(struct index_state *, struct cache_entry *ce, int option);
|
|
void rename_index_entry_at(struct index_state *, int pos, const char *new_name);
|
|
|
|
/* Remove entry, return true if there are more entries to go. */
|
|
int remove_index_entry_at(struct index_state *, int pos);
|
|
|
|
void remove_marked_cache_entries(struct index_state *istate, int invalidate);
|
|
int remove_file_from_index(struct index_state *, const char *path);
|
|
#define ADD_CACHE_VERBOSE 1
|
|
#define ADD_CACHE_PRETEND 2
|
|
#define ADD_CACHE_IGNORE_ERRORS 4
|
|
#define ADD_CACHE_IGNORE_REMOVAL 8
|
|
#define ADD_CACHE_INTENT 16
|
|
/*
|
|
* These two are used to add the contents of the file at path
|
|
* to the index, marking the working tree up-to-date by storing
|
|
* the cached stat info in the resulting cache entry. A caller
|
|
* that has already run lstat(2) on the path can call
|
|
* add_to_index(), and all others can call add_file_to_index();
|
|
* the latter will do necessary lstat(2) internally before
|
|
* calling the former.
|
|
*/
|
|
int add_to_index(struct index_state *, const char *path, struct stat *, int flags);
|
|
int add_file_to_index(struct index_state *, const char *path, int flags);
|
|
|
|
int chmod_index_entry(struct index_state *, struct cache_entry *ce, char flip);
|
|
int ce_same_name(const struct cache_entry *a, const struct cache_entry *b);
|
|
void set_object_name_for_intent_to_add_entry(struct cache_entry *ce);
|
|
int index_name_is_other(struct index_state *, const char *, int);
|
|
void *read_blob_data_from_index(struct index_state *, const char *, unsigned long *);
|
|
|
|
/* do stat comparison even if CE_VALID is true */
|
|
#define CE_MATCH_IGNORE_VALID 01
|
|
/* do not check the contents but report dirty on racily-clean entries */
|
|
#define CE_MATCH_RACY_IS_DIRTY 02
|
|
/* do stat comparison even if CE_SKIP_WORKTREE is true */
|
|
#define CE_MATCH_IGNORE_SKIP_WORKTREE 04
|
|
/* ignore non-existent files during stat update */
|
|
#define CE_MATCH_IGNORE_MISSING 0x08
|
|
/* enable stat refresh */
|
|
#define CE_MATCH_REFRESH 0x10
|
|
/* don't refresh_fsmonitor state or do stat comparison even if CE_FSMONITOR_VALID is true */
|
|
#define CE_MATCH_IGNORE_FSMONITOR 0X20
|
|
int is_racy_timestamp(const struct index_state *istate,
|
|
const struct cache_entry *ce);
|
|
int has_racy_timestamp(struct index_state *istate);
|
|
int ie_match_stat(struct index_state *, const struct cache_entry *, struct stat *, unsigned int);
|
|
int ie_modified(struct index_state *, const struct cache_entry *, struct stat *, unsigned int);
|
|
|
|
int match_stat_data_racy(const struct index_state *istate,
|
|
const struct stat_data *sd, struct stat *st);
|
|
|
|
void fill_stat_cache_info(struct index_state *istate, struct cache_entry *ce, struct stat *st);
|
|
|
|
/*
|
|
* Fill members of st by members of sd enough to convince match_stat()
|
|
* to consider that they match. It should be usable as a replacement
|
|
* for lstat() for a tracked path that is known to be up-to-date via
|
|
* some out-of-line means (like fsmonitor).
|
|
*/
|
|
int fake_lstat(const struct cache_entry *ce, struct stat *st);
|
|
|
|
#define REFRESH_REALLY (1 << 0) /* ignore_valid */
|
|
#define REFRESH_UNMERGED (1 << 1) /* allow unmerged */
|
|
#define REFRESH_QUIET (1 << 2) /* be quiet about it */
|
|
#define REFRESH_IGNORE_MISSING (1 << 3) /* ignore non-existent */
|
|
#define REFRESH_IGNORE_SUBMODULES (1 << 4) /* ignore submodules */
|
|
#define REFRESH_IN_PORCELAIN (1 << 5) /* user friendly output, not "needs update" */
|
|
#define REFRESH_PROGRESS (1 << 6) /* show progress bar if stderr is tty */
|
|
#define REFRESH_IGNORE_SKIP_WORKTREE (1 << 7) /* ignore skip_worktree entries */
|
|
int refresh_index(struct index_state *, unsigned int flags, const struct pathspec *pathspec, char *seen, const char *header_msg);
|
|
/*
|
|
* Refresh the index and write it to disk.
|
|
*
|
|
* 'refresh_flags' is passed directly to 'refresh_index()', while
|
|
* 'COMMIT_LOCK | write_flags' is passed to 'write_locked_index()', so
|
|
* the lockfile is always either committed or rolled back.
|
|
*
|
|
* If 'gentle' is passed, errors locking the index are ignored.
|
|
*
|
|
* Return 1 if refreshing the index returns an error, -1 if writing
|
|
* the index to disk fails, 0 on success.
|
|
*
|
|
* Note that if refreshing the index returns an error, we still write
|
|
* out the index (unless locking fails).
|
|
*/
|
|
int repo_refresh_and_write_index(struct repository*, unsigned int refresh_flags, unsigned int write_flags, int gentle, const struct pathspec *, char *seen, const char *header_msg);
|
|
|
|
struct cache_entry *refresh_cache_entry(struct index_state *, struct cache_entry *, unsigned int);
|
|
|
|
void set_alternate_index_output(const char *);
|
|
|
|
extern int verify_index_checksum;
|
|
extern int verify_ce_order;
|
|
|
|
int cmp_cache_name_compare(const void *a_, const void *b_);
|
|
|
|
int add_files_to_cache(struct repository *repo, const char *prefix,
|
|
const struct pathspec *pathspec, char *ps_matched,
|
|
int include_sparse, int flags);
|
|
|
|
void overlay_tree_on_index(struct index_state *istate,
|
|
const char *tree_name, const char *prefix);
|
|
|
|
#endif /* READ_CACHE_LL_H */
|