mirror of
https://github.com/git/git.git
synced 2024-11-27 12:03:55 +08:00
50c37ee839
The function add_files_to_cache() is used by all three of builtin/{add, checkout, commit}.c. That suggests this is common library code, and should be moved somewhere else, like read-cache.c. However, the function and its helpers made use of two global variables that made straight code movement difficult: * the_index * include_sparse The latter was perhaps more problematic since it was only accessible in builtin/add.c but was still affecting builtin/checkout.c and builtin/commit.c without this fact being very clear from the code. I'm not sure if the other two callers would want to add a `--sparse` flag similar to add.c to get non-default behavior, but exposing this dependence will help if we ever decide we do want to add such a flag. Modify add_files_to_cache() and its helpers to accept the necessary arguments instead of relying on globals. Signed-off-by: Elijah Newren <newren@gmail.com> Signed-off-by: Junio C Hamano <gitster@pobox.com>
610 lines
21 KiB
C
610 lines
21 KiB
C
#ifndef CACHE_H
|
|
#define CACHE_H
|
|
|
|
#include "git-compat-util.h"
|
|
#include "strbuf.h"
|
|
#include "hashmap.h"
|
|
#include "pathspec.h"
|
|
#include "object.h"
|
|
#include "statinfo.h"
|
|
|
|
/*
|
|
* Basic data structures for the directory cache
|
|
*/
|
|
|
|
#define CACHE_SIGNATURE 0x44495243 /* "DIRC" */
|
|
struct cache_header {
|
|
uint32_t hdr_signature;
|
|
uint32_t hdr_version;
|
|
uint32_t hdr_entries;
|
|
};
|
|
|
|
#define INDEX_FORMAT_LB 2
|
|
#define INDEX_FORMAT_UB 4
|
|
|
|
struct cache_entry {
|
|
struct hashmap_entry ent;
|
|
struct stat_data ce_stat_data;
|
|
unsigned int ce_mode;
|
|
unsigned int ce_flags;
|
|
unsigned int mem_pool_allocated;
|
|
unsigned int ce_namelen;
|
|
unsigned int index; /* for link extension */
|
|
struct object_id oid;
|
|
char name[FLEX_ARRAY]; /* more */
|
|
};
|
|
|
|
#define CE_STAGEMASK (0x3000)
|
|
#define CE_EXTENDED (0x4000)
|
|
#define CE_VALID (0x8000)
|
|
#define CE_STAGESHIFT 12
|
|
|
|
/*
|
|
* Range 0xFFFF0FFF in ce_flags is divided into
|
|
* two parts: in-memory flags and on-disk ones.
|
|
* Flags in CE_EXTENDED_FLAGS will get saved on-disk
|
|
* if you want to save a new flag, add it in
|
|
* CE_EXTENDED_FLAGS
|
|
*
|
|
* In-memory only flags
|
|
*/
|
|
#define CE_UPDATE (1 << 16)
|
|
#define CE_REMOVE (1 << 17)
|
|
#define CE_UPTODATE (1 << 18)
|
|
#define CE_ADDED (1 << 19)
|
|
|
|
#define CE_HASHED (1 << 20)
|
|
#define CE_FSMONITOR_VALID (1 << 21)
|
|
#define CE_WT_REMOVE (1 << 22) /* remove in work directory */
|
|
#define CE_CONFLICTED (1 << 23)
|
|
|
|
#define CE_UNPACKED (1 << 24)
|
|
#define CE_NEW_SKIP_WORKTREE (1 << 25)
|
|
|
|
/* used to temporarily mark paths matched by pathspecs */
|
|
#define CE_MATCHED (1 << 26)
|
|
|
|
#define CE_UPDATE_IN_BASE (1 << 27)
|
|
#define CE_STRIP_NAME (1 << 28)
|
|
|
|
/*
|
|
* Extended on-disk flags
|
|
*/
|
|
#define CE_INTENT_TO_ADD (1 << 29)
|
|
#define CE_SKIP_WORKTREE (1 << 30)
|
|
/* CE_EXTENDED2 is for future extension */
|
|
#define CE_EXTENDED2 (1U << 31)
|
|
|
|
#define CE_EXTENDED_FLAGS (CE_INTENT_TO_ADD | CE_SKIP_WORKTREE)
|
|
|
|
/*
|
|
* Safeguard to avoid saving wrong flags:
|
|
* - CE_EXTENDED2 won't get saved until its semantic is known
|
|
* - Bits in 0x0000FFFF have been saved in ce_flags already
|
|
* - Bits in 0x003F0000 are currently in-memory flags
|
|
*/
|
|
#if CE_EXTENDED_FLAGS & 0x803FFFFF
|
|
#error "CE_EXTENDED_FLAGS out of range"
|
|
#endif
|
|
|
|
/* Forward structure decls */
|
|
struct pathspec;
|
|
struct tree;
|
|
|
|
/*
|
|
* Copy the sha1 and stat state of a cache entry from one to
|
|
* another. But we never change the name, or the hash state!
|
|
*/
|
|
static inline void copy_cache_entry(struct cache_entry *dst,
|
|
const struct cache_entry *src)
|
|
{
|
|
unsigned int state = dst->ce_flags & CE_HASHED;
|
|
int mem_pool_allocated = dst->mem_pool_allocated;
|
|
|
|
/* Don't copy hash chain and name */
|
|
memcpy(&dst->ce_stat_data, &src->ce_stat_data,
|
|
offsetof(struct cache_entry, name) -
|
|
offsetof(struct cache_entry, ce_stat_data));
|
|
|
|
/* Restore the hash state */
|
|
dst->ce_flags = (dst->ce_flags & ~CE_HASHED) | state;
|
|
|
|
/* Restore the mem_pool_allocated flag */
|
|
dst->mem_pool_allocated = mem_pool_allocated;
|
|
}
|
|
|
|
static inline unsigned create_ce_flags(unsigned stage)
|
|
{
|
|
return (stage << CE_STAGESHIFT);
|
|
}
|
|
|
|
#define ce_namelen(ce) ((ce)->ce_namelen)
|
|
#define ce_size(ce) cache_entry_size(ce_namelen(ce))
|
|
#define ce_stage(ce) ((CE_STAGEMASK & (ce)->ce_flags) >> CE_STAGESHIFT)
|
|
#define ce_uptodate(ce) ((ce)->ce_flags & CE_UPTODATE)
|
|
#define ce_skip_worktree(ce) ((ce)->ce_flags & CE_SKIP_WORKTREE)
|
|
#define ce_mark_uptodate(ce) ((ce)->ce_flags |= CE_UPTODATE)
|
|
#define ce_intent_to_add(ce) ((ce)->ce_flags & CE_INTENT_TO_ADD)
|
|
|
|
static inline unsigned int ce_mode_from_stat(const struct cache_entry *ce,
|
|
unsigned int mode)
|
|
{
|
|
extern int trust_executable_bit, has_symlinks;
|
|
if (!has_symlinks && S_ISREG(mode) &&
|
|
ce && S_ISLNK(ce->ce_mode))
|
|
return ce->ce_mode;
|
|
if (!trust_executable_bit && S_ISREG(mode)) {
|
|
if (ce && S_ISREG(ce->ce_mode))
|
|
return ce->ce_mode;
|
|
return create_ce_mode(0666);
|
|
}
|
|
return create_ce_mode(mode);
|
|
}
|
|
static inline int ce_to_dtype(const struct cache_entry *ce)
|
|
{
|
|
unsigned ce_mode = ntohl(ce->ce_mode);
|
|
if (S_ISREG(ce_mode))
|
|
return DT_REG;
|
|
else if (S_ISDIR(ce_mode) || S_ISGITLINK(ce_mode))
|
|
return DT_DIR;
|
|
else if (S_ISLNK(ce_mode))
|
|
return DT_LNK;
|
|
else
|
|
return DT_UNKNOWN;
|
|
}
|
|
|
|
static inline int ce_path_match(struct index_state *istate,
|
|
const struct cache_entry *ce,
|
|
const struct pathspec *pathspec,
|
|
char *seen)
|
|
{
|
|
return match_pathspec(istate, pathspec, ce->name, ce_namelen(ce), 0, seen,
|
|
S_ISDIR(ce->ce_mode) || S_ISGITLINK(ce->ce_mode));
|
|
}
|
|
|
|
#define cache_entry_size(len) (offsetof(struct cache_entry,name) + (len) + 1)
|
|
|
|
#define SOMETHING_CHANGED (1 << 0) /* unclassified changes go here */
|
|
#define CE_ENTRY_CHANGED (1 << 1)
|
|
#define CE_ENTRY_REMOVED (1 << 2)
|
|
#define CE_ENTRY_ADDED (1 << 3)
|
|
#define RESOLVE_UNDO_CHANGED (1 << 4)
|
|
#define CACHE_TREE_CHANGED (1 << 5)
|
|
#define SPLIT_INDEX_ORDERED (1 << 6)
|
|
#define UNTRACKED_CHANGED (1 << 7)
|
|
#define FSMONITOR_CHANGED (1 << 8)
|
|
|
|
struct split_index;
|
|
struct untracked_cache;
|
|
struct progress;
|
|
struct pattern_list;
|
|
|
|
enum sparse_index_mode {
|
|
/*
|
|
* There are no sparse directories in the index at all.
|
|
*
|
|
* Repositories that don't use cone-mode sparse-checkout will
|
|
* always have their indexes in this mode.
|
|
*/
|
|
INDEX_EXPANDED = 0,
|
|
|
|
/*
|
|
* The index has already been collapsed to sparse directories
|
|
* whereever possible.
|
|
*/
|
|
INDEX_COLLAPSED,
|
|
|
|
/*
|
|
* The sparse directories that exist are outside the
|
|
* sparse-checkout boundary, but it is possible that some file
|
|
* entries could collapse to sparse directory entries.
|
|
*/
|
|
INDEX_PARTIALLY_SPARSE,
|
|
};
|
|
|
|
struct index_state {
|
|
struct cache_entry **cache;
|
|
unsigned int version;
|
|
unsigned int cache_nr, cache_alloc, cache_changed;
|
|
struct string_list *resolve_undo;
|
|
struct cache_tree *cache_tree;
|
|
struct split_index *split_index;
|
|
struct cache_time timestamp;
|
|
unsigned name_hash_initialized : 1,
|
|
initialized : 1,
|
|
drop_cache_tree : 1,
|
|
updated_workdir : 1,
|
|
updated_skipworktree : 1,
|
|
fsmonitor_has_run_once : 1;
|
|
enum sparse_index_mode sparse_index;
|
|
struct hashmap name_hash;
|
|
struct hashmap dir_hash;
|
|
struct object_id oid;
|
|
struct untracked_cache *untracked;
|
|
char *fsmonitor_last_update;
|
|
struct ewah_bitmap *fsmonitor_dirty;
|
|
struct mem_pool *ce_mem_pool;
|
|
struct progress *progress;
|
|
struct repository *repo;
|
|
struct pattern_list *sparse_checkout_patterns;
|
|
};
|
|
|
|
/**
|
|
* A "struct index_state istate" must be initialized with
|
|
* INDEX_STATE_INIT or the corresponding index_state_init().
|
|
*
|
|
* If the variable won't be used again, use release_index() to free()
|
|
* its resources. If it needs to be used again use discard_index(),
|
|
* which does the same thing, but will use use index_state_init() at
|
|
* the end. The discard_index() will use its own "istate->repo" as the
|
|
* "r" argument to index_state_init() in that case.
|
|
*/
|
|
#define INDEX_STATE_INIT(r) { \
|
|
.repo = (r), \
|
|
}
|
|
void index_state_init(struct index_state *istate, struct repository *r);
|
|
void release_index(struct index_state *istate);
|
|
|
|
/* Name hashing */
|
|
int test_lazy_init_name_hash(struct index_state *istate, int try_threaded);
|
|
void add_name_hash(struct index_state *istate, struct cache_entry *ce);
|
|
void remove_name_hash(struct index_state *istate, struct cache_entry *ce);
|
|
void free_name_hash(struct index_state *istate);
|
|
|
|
/* Cache entry creation and cleanup */
|
|
|
|
/*
|
|
* Create cache_entry intended for use in the specified index. Caller
|
|
* is responsible for discarding the cache_entry with
|
|
* `discard_cache_entry`.
|
|
*/
|
|
struct cache_entry *make_cache_entry(struct index_state *istate,
|
|
unsigned int mode,
|
|
const struct object_id *oid,
|
|
const char *path,
|
|
int stage,
|
|
unsigned int refresh_options);
|
|
|
|
struct cache_entry *make_empty_cache_entry(struct index_state *istate,
|
|
size_t name_len);
|
|
|
|
/*
|
|
* Create a cache_entry that is not intended to be added to an index. If
|
|
* `ce_mem_pool` is not NULL, the entry is allocated within the given memory
|
|
* pool. Caller is responsible for discarding "loose" entries with
|
|
* `discard_cache_entry()` and the memory pool with
|
|
* `mem_pool_discard(ce_mem_pool, should_validate_cache_entries())`.
|
|
*/
|
|
struct cache_entry *make_transient_cache_entry(unsigned int mode,
|
|
const struct object_id *oid,
|
|
const char *path,
|
|
int stage,
|
|
struct mem_pool *ce_mem_pool);
|
|
|
|
struct cache_entry *make_empty_transient_cache_entry(size_t len,
|
|
struct mem_pool *ce_mem_pool);
|
|
|
|
/*
|
|
* Discard cache entry.
|
|
*/
|
|
void discard_cache_entry(struct cache_entry *ce);
|
|
|
|
/*
|
|
* Check configuration if we should perform extra validation on cache
|
|
* entries.
|
|
*/
|
|
int should_validate_cache_entries(void);
|
|
|
|
/*
|
|
* Duplicate a cache_entry. Allocate memory for the new entry from a
|
|
* memory_pool. Takes into account cache_entry fields that are meant
|
|
* for managing the underlying memory allocation of the cache_entry.
|
|
*/
|
|
struct cache_entry *dup_cache_entry(const struct cache_entry *ce, struct index_state *istate);
|
|
|
|
/*
|
|
* Validate the cache entries in the index. This is an internal
|
|
* consistency check that the cache_entry structs are allocated from
|
|
* the expected memory pool.
|
|
*/
|
|
void validate_cache_entries(const struct index_state *istate);
|
|
|
|
/*
|
|
* Bulk prefetch all missing cache entries that are not GITLINKs and that match
|
|
* the given predicate. This function should only be called if
|
|
* repo_has_promisor_remote() returns true.
|
|
*/
|
|
typedef int (*must_prefetch_predicate)(const struct cache_entry *);
|
|
void prefetch_cache_entries(const struct index_state *istate,
|
|
must_prefetch_predicate must_prefetch);
|
|
|
|
#ifdef USE_THE_INDEX_VARIABLE
|
|
extern struct index_state the_index;
|
|
#endif
|
|
|
|
/* Initialize and use the cache information */
|
|
struct lock_file;
|
|
void preload_index(struct index_state *index,
|
|
const struct pathspec *pathspec,
|
|
unsigned int refresh_flags);
|
|
int do_read_index(struct index_state *istate, const char *path,
|
|
int must_exist); /* for testting only! */
|
|
int read_index_from(struct index_state *, const char *path,
|
|
const char *gitdir);
|
|
int is_index_unborn(struct index_state *);
|
|
|
|
void ensure_full_index(struct index_state *istate);
|
|
|
|
/* For use with `write_locked_index()`. */
|
|
#define COMMIT_LOCK (1 << 0)
|
|
#define SKIP_IF_UNCHANGED (1 << 1)
|
|
|
|
/*
|
|
* Write the index while holding an already-taken lock. Close the lock,
|
|
* and if `COMMIT_LOCK` is given, commit it.
|
|
*
|
|
* Unless a split index is in use, write the index into the lockfile.
|
|
*
|
|
* With a split index, write the shared index to a temporary file,
|
|
* adjust its permissions and rename it into place, then write the
|
|
* split index to the lockfile. If the temporary file for the shared
|
|
* index cannot be created, fall back to the behavior described in
|
|
* the previous paragraph.
|
|
*
|
|
* With `COMMIT_LOCK`, the lock is always committed or rolled back.
|
|
* Without it, the lock is closed, but neither committed nor rolled
|
|
* back.
|
|
*
|
|
* If `SKIP_IF_UNCHANGED` is given and the index is unchanged, nothing
|
|
* is written (and the lock is rolled back if `COMMIT_LOCK` is given).
|
|
*/
|
|
int write_locked_index(struct index_state *, struct lock_file *lock, unsigned flags);
|
|
|
|
void discard_index(struct index_state *);
|
|
void move_index_extensions(struct index_state *dst, struct index_state *src);
|
|
int unmerged_index(const struct index_state *);
|
|
|
|
/**
|
|
* Returns 1 if istate differs from tree, 0 otherwise. If tree is NULL,
|
|
* compares istate to HEAD. If tree is NULL and on an unborn branch,
|
|
* returns 1 if there are entries in istate, 0 otherwise. If an strbuf is
|
|
* provided, the space-separated list of files that differ will be appended
|
|
* to it.
|
|
*/
|
|
int repo_index_has_changes(struct repository *repo,
|
|
struct tree *tree,
|
|
struct strbuf *sb);
|
|
|
|
int verify_path(const char *path, unsigned mode);
|
|
int strcmp_offset(const char *s1, const char *s2, size_t *first_change);
|
|
int index_dir_exists(struct index_state *istate, const char *name, int namelen);
|
|
void adjust_dirname_case(struct index_state *istate, char *name);
|
|
struct cache_entry *index_file_exists(struct index_state *istate, const char *name, int namelen, int igncase);
|
|
|
|
/*
|
|
* Searches for an entry defined by name and namelen in the given index.
|
|
* If the return value is positive (including 0) it is the position of an
|
|
* exact match. If the return value is negative, the negated value minus 1
|
|
* is the position where the entry would be inserted.
|
|
* Example: The current index consists of these files and its stages:
|
|
*
|
|
* b#0, d#0, f#1, f#3
|
|
*
|
|
* index_name_pos(&index, "a", 1) -> -1
|
|
* index_name_pos(&index, "b", 1) -> 0
|
|
* index_name_pos(&index, "c", 1) -> -2
|
|
* index_name_pos(&index, "d", 1) -> 1
|
|
* index_name_pos(&index, "e", 1) -> -3
|
|
* index_name_pos(&index, "f", 1) -> -3
|
|
* index_name_pos(&index, "g", 1) -> -5
|
|
*/
|
|
int index_name_pos(struct index_state *, const char *name, int namelen);
|
|
|
|
/*
|
|
* Like index_name_pos, returns the position of an entry of the given name in
|
|
* the index if one exists, otherwise returns a negative value where the negated
|
|
* value minus 1 is the position where the index entry would be inserted. Unlike
|
|
* index_name_pos, however, a sparse index is not expanded to find an entry
|
|
* inside a sparse directory.
|
|
*/
|
|
int index_name_pos_sparse(struct index_state *, const char *name, int namelen);
|
|
|
|
/*
|
|
* Determines whether an entry with the given name exists within the
|
|
* given index. The return value is 1 if an exact match is found, otherwise
|
|
* it is 0. Note that, unlike index_name_pos, this function does not expand
|
|
* the index if it is sparse. If an item exists within the full index but it
|
|
* is contained within a sparse directory (and not in the sparse index), 0 is
|
|
* returned.
|
|
*/
|
|
int index_entry_exists(struct index_state *, const char *name, int namelen);
|
|
|
|
/*
|
|
* Some functions return the negative complement of an insert position when a
|
|
* precise match was not found but a position was found where the entry would
|
|
* need to be inserted. This helper protects that logic from any integer
|
|
* underflow.
|
|
*/
|
|
static inline int index_pos_to_insert_pos(uintmax_t pos)
|
|
{
|
|
if (pos > INT_MAX)
|
|
die("overflow: -1 - %"PRIuMAX, pos);
|
|
return -1 - (int)pos;
|
|
}
|
|
|
|
#define ADD_CACHE_OK_TO_ADD 1 /* Ok to add */
|
|
#define ADD_CACHE_OK_TO_REPLACE 2 /* Ok to replace file/directory */
|
|
#define ADD_CACHE_SKIP_DFCHECK 4 /* Ok to skip DF conflict checks */
|
|
#define ADD_CACHE_JUST_APPEND 8 /* Append only */
|
|
#define ADD_CACHE_NEW_ONLY 16 /* Do not replace existing ones */
|
|
#define ADD_CACHE_KEEP_CACHE_TREE 32 /* Do not invalidate cache-tree */
|
|
#define ADD_CACHE_RENORMALIZE 64 /* Pass along HASH_RENORMALIZE */
|
|
int add_index_entry(struct index_state *, struct cache_entry *ce, int option);
|
|
void rename_index_entry_at(struct index_state *, int pos, const char *new_name);
|
|
|
|
/* Remove entry, return true if there are more entries to go. */
|
|
int remove_index_entry_at(struct index_state *, int pos);
|
|
|
|
void remove_marked_cache_entries(struct index_state *istate, int invalidate);
|
|
int remove_file_from_index(struct index_state *, const char *path);
|
|
#define ADD_CACHE_VERBOSE 1
|
|
#define ADD_CACHE_PRETEND 2
|
|
#define ADD_CACHE_IGNORE_ERRORS 4
|
|
#define ADD_CACHE_IGNORE_REMOVAL 8
|
|
#define ADD_CACHE_INTENT 16
|
|
/*
|
|
* These two are used to add the contents of the file at path
|
|
* to the index, marking the working tree up-to-date by storing
|
|
* the cached stat info in the resulting cache entry. A caller
|
|
* that has already run lstat(2) on the path can call
|
|
* add_to_index(), and all others can call add_file_to_index();
|
|
* the latter will do necessary lstat(2) internally before
|
|
* calling the former.
|
|
*/
|
|
int add_to_index(struct index_state *, const char *path, struct stat *, int flags);
|
|
int add_file_to_index(struct index_state *, const char *path, int flags);
|
|
|
|
int chmod_index_entry(struct index_state *, struct cache_entry *ce, char flip);
|
|
int ce_same_name(const struct cache_entry *a, const struct cache_entry *b);
|
|
void set_object_name_for_intent_to_add_entry(struct cache_entry *ce);
|
|
int index_name_is_other(struct index_state *, const char *, int);
|
|
void *read_blob_data_from_index(struct index_state *, const char *, unsigned long *);
|
|
|
|
/* do stat comparison even if CE_VALID is true */
|
|
#define CE_MATCH_IGNORE_VALID 01
|
|
/* do not check the contents but report dirty on racily-clean entries */
|
|
#define CE_MATCH_RACY_IS_DIRTY 02
|
|
/* do stat comparison even if CE_SKIP_WORKTREE is true */
|
|
#define CE_MATCH_IGNORE_SKIP_WORKTREE 04
|
|
/* ignore non-existent files during stat update */
|
|
#define CE_MATCH_IGNORE_MISSING 0x08
|
|
/* enable stat refresh */
|
|
#define CE_MATCH_REFRESH 0x10
|
|
/* don't refresh_fsmonitor state or do stat comparison even if CE_FSMONITOR_VALID is true */
|
|
#define CE_MATCH_IGNORE_FSMONITOR 0X20
|
|
int is_racy_timestamp(const struct index_state *istate,
|
|
const struct cache_entry *ce);
|
|
int has_racy_timestamp(struct index_state *istate);
|
|
int ie_match_stat(struct index_state *, const struct cache_entry *, struct stat *, unsigned int);
|
|
int ie_modified(struct index_state *, const struct cache_entry *, struct stat *, unsigned int);
|
|
|
|
/*
|
|
* Record to sd the data from st that we use to check whether a file
|
|
* might have changed.
|
|
*/
|
|
void fill_stat_data(struct stat_data *sd, struct stat *st);
|
|
|
|
/*
|
|
* Return 0 if st is consistent with a file not having been changed
|
|
* since sd was filled. If there are differences, return a
|
|
* combination of MTIME_CHANGED, CTIME_CHANGED, OWNER_CHANGED,
|
|
* INODE_CHANGED, and DATA_CHANGED.
|
|
*/
|
|
int match_stat_data(const struct stat_data *sd, struct stat *st);
|
|
int match_stat_data_racy(const struct index_state *istate,
|
|
const struct stat_data *sd, struct stat *st);
|
|
|
|
void fill_stat_cache_info(struct index_state *istate, struct cache_entry *ce, struct stat *st);
|
|
|
|
#define REFRESH_REALLY (1 << 0) /* ignore_valid */
|
|
#define REFRESH_UNMERGED (1 << 1) /* allow unmerged */
|
|
#define REFRESH_QUIET (1 << 2) /* be quiet about it */
|
|
#define REFRESH_IGNORE_MISSING (1 << 3) /* ignore non-existent */
|
|
#define REFRESH_IGNORE_SUBMODULES (1 << 4) /* ignore submodules */
|
|
#define REFRESH_IN_PORCELAIN (1 << 5) /* user friendly output, not "needs update" */
|
|
#define REFRESH_PROGRESS (1 << 6) /* show progress bar if stderr is tty */
|
|
#define REFRESH_IGNORE_SKIP_WORKTREE (1 << 7) /* ignore skip_worktree entries */
|
|
int refresh_index(struct index_state *, unsigned int flags, const struct pathspec *pathspec, char *seen, const char *header_msg);
|
|
/*
|
|
* Refresh the index and write it to disk.
|
|
*
|
|
* 'refresh_flags' is passed directly to 'refresh_index()', while
|
|
* 'COMMIT_LOCK | write_flags' is passed to 'write_locked_index()', so
|
|
* the lockfile is always either committed or rolled back.
|
|
*
|
|
* If 'gentle' is passed, errors locking the index are ignored.
|
|
*
|
|
* Return 1 if refreshing the index returns an error, -1 if writing
|
|
* the index to disk fails, 0 on success.
|
|
*
|
|
* Note that if refreshing the index returns an error, we still write
|
|
* out the index (unless locking fails).
|
|
*/
|
|
int repo_refresh_and_write_index(struct repository*, unsigned int refresh_flags, unsigned int write_flags, int gentle, const struct pathspec *, char *seen, const char *header_msg);
|
|
|
|
struct cache_entry *refresh_cache_entry(struct index_state *, struct cache_entry *, unsigned int);
|
|
|
|
void set_alternate_index_output(const char *);
|
|
|
|
extern int verify_index_checksum;
|
|
extern int verify_ce_order;
|
|
|
|
#define MTIME_CHANGED 0x0001
|
|
#define CTIME_CHANGED 0x0002
|
|
#define OWNER_CHANGED 0x0004
|
|
#define MODE_CHANGED 0x0008
|
|
#define INODE_CHANGED 0x0010
|
|
#define DATA_CHANGED 0x0020
|
|
#define TYPE_CHANGED 0x0040
|
|
|
|
int cmp_cache_name_compare(const void *a_, const void *b_);
|
|
|
|
/* add */
|
|
/*
|
|
* return 0 if success, 1 - if addition of a file failed and
|
|
* ADD_FILES_IGNORE_ERRORS was specified in flags
|
|
*/
|
|
int add_files_to_cache(struct repository *repo, const char *prefix,
|
|
const struct pathspec *pathspec, int include_sparse,
|
|
int flags);
|
|
|
|
/* diff.c */
|
|
extern int diff_auto_refresh_index;
|
|
|
|
/* ls-files */
|
|
void overlay_tree_on_index(struct index_state *istate,
|
|
const char *tree_name, const char *prefix);
|
|
|
|
/* merge.c */
|
|
struct commit_list;
|
|
int try_merge_command(struct repository *r,
|
|
const char *strategy, size_t xopts_nr,
|
|
const char **xopts, struct commit_list *common,
|
|
const char *head_arg, struct commit_list *remotes);
|
|
int checkout_fast_forward(struct repository *r,
|
|
const struct object_id *from,
|
|
const struct object_id *to,
|
|
int overwrite_ignore);
|
|
|
|
|
|
int sane_execvp(const char *file, char *const argv[]);
|
|
|
|
/*
|
|
* A struct to encapsulate the concept of whether a file has changed
|
|
* since we last checked it. This uses criteria similar to those used
|
|
* for the index.
|
|
*/
|
|
struct stat_validity {
|
|
struct stat_data *sd;
|
|
};
|
|
|
|
void stat_validity_clear(struct stat_validity *sv);
|
|
|
|
/*
|
|
* Returns 1 if the path is a regular file (or a symlink to a regular
|
|
* file) and matches the saved stat_validity, 0 otherwise. A missing
|
|
* or inaccessible file is considered a match if the struct was just
|
|
* initialized, or if the previous update found an inaccessible file.
|
|
*/
|
|
int stat_validity_check(struct stat_validity *sv, const char *path);
|
|
|
|
/*
|
|
* Update the stat_validity from a file opened at descriptor fd. If
|
|
* the file is missing, inaccessible, or not a regular file, then
|
|
* future calls to stat_validity_check will match iff one of those
|
|
* conditions continues to be true.
|
|
*/
|
|
void stat_validity_update(struct stat_validity *sv, int fd);
|
|
|
|
#endif /* CACHE_H */
|