mirror of
https://github.com/git/git.git
synced 2024-11-30 21:44:02 +08:00
82823118b9
If a full hexadecimal hash is given as a --negotiation-tip to "git fetch", and that hash does not correspond to an object, "git fetch" will segfault if --negotiate-only is given and will silently ignore that hash otherwise. Make these cases fatal errors, just like the case when an invalid ref name or abbreviated hash is given. While at it, mark the error messages as translatable. Signed-off-by: Jonathan Tan <jonathantanmy@google.com> Signed-off-by: Junio C Hamano <gitster@pobox.com>
2142 lines
59 KiB
C
2142 lines
59 KiB
C
/*
|
|
* "git fetch"
|
|
*/
|
|
#include "cache.h"
|
|
#include "config.h"
|
|
#include "repository.h"
|
|
#include "refs.h"
|
|
#include "refspec.h"
|
|
#include "object-store.h"
|
|
#include "oidset.h"
|
|
#include "commit.h"
|
|
#include "builtin.h"
|
|
#include "string-list.h"
|
|
#include "remote.h"
|
|
#include "transport.h"
|
|
#include "run-command.h"
|
|
#include "parse-options.h"
|
|
#include "sigchain.h"
|
|
#include "submodule-config.h"
|
|
#include "submodule.h"
|
|
#include "connected.h"
|
|
#include "strvec.h"
|
|
#include "utf8.h"
|
|
#include "packfile.h"
|
|
#include "list-objects-filter-options.h"
|
|
#include "commit-reach.h"
|
|
#include "branch.h"
|
|
#include "promisor-remote.h"
|
|
#include "commit-graph.h"
|
|
#include "shallow.h"
|
|
|
|
#define FORCED_UPDATES_DELAY_WARNING_IN_MS (10 * 1000)
|
|
|
|
static const char * const builtin_fetch_usage[] = {
|
|
N_("git fetch [<options>] [<repository> [<refspec>...]]"),
|
|
N_("git fetch [<options>] <group>"),
|
|
N_("git fetch --multiple [<options>] [(<repository> | <group>)...]"),
|
|
N_("git fetch --all [<options>]"),
|
|
NULL
|
|
};
|
|
|
|
enum {
|
|
TAGS_UNSET = 0,
|
|
TAGS_DEFAULT = 1,
|
|
TAGS_SET = 2
|
|
};
|
|
|
|
static int fetch_prune_config = -1; /* unspecified */
|
|
static int fetch_show_forced_updates = 1;
|
|
static uint64_t forced_updates_ms = 0;
|
|
static int prefetch = 0;
|
|
static int prune = -1; /* unspecified */
|
|
#define PRUNE_BY_DEFAULT 0 /* do we prune by default? */
|
|
|
|
static int fetch_prune_tags_config = -1; /* unspecified */
|
|
static int prune_tags = -1; /* unspecified */
|
|
#define PRUNE_TAGS_BY_DEFAULT 0 /* do we prune tags by default? */
|
|
|
|
static int all, append, dry_run, force, keep, multiple, update_head_ok;
|
|
static int write_fetch_head = 1;
|
|
static int verbosity, deepen_relative, set_upstream;
|
|
static int progress = -1;
|
|
static int enable_auto_gc = 1;
|
|
static int tags = TAGS_DEFAULT, unshallow, update_shallow, deepen;
|
|
static int max_jobs = -1, submodule_fetch_jobs_config = -1;
|
|
static int fetch_parallel_config = 1;
|
|
static int atomic_fetch;
|
|
static enum transport_family family;
|
|
static const char *depth;
|
|
static const char *deepen_since;
|
|
static const char *upload_pack;
|
|
static struct string_list deepen_not = STRING_LIST_INIT_NODUP;
|
|
static struct strbuf default_rla = STRBUF_INIT;
|
|
static struct transport *gtransport;
|
|
static struct transport *gsecondary;
|
|
static const char *submodule_prefix = "";
|
|
static int recurse_submodules = RECURSE_SUBMODULES_DEFAULT;
|
|
static int recurse_submodules_default = RECURSE_SUBMODULES_ON_DEMAND;
|
|
static int shown_url = 0;
|
|
static struct refspec refmap = REFSPEC_INIT_FETCH;
|
|
static struct list_objects_filter_options filter_options;
|
|
static struct string_list server_options = STRING_LIST_INIT_DUP;
|
|
static struct string_list negotiation_tip = STRING_LIST_INIT_NODUP;
|
|
static int fetch_write_commit_graph = -1;
|
|
static int stdin_refspecs = 0;
|
|
static int negotiate_only;
|
|
|
|
static int git_fetch_config(const char *k, const char *v, void *cb)
|
|
{
|
|
if (!strcmp(k, "fetch.prune")) {
|
|
fetch_prune_config = git_config_bool(k, v);
|
|
return 0;
|
|
}
|
|
|
|
if (!strcmp(k, "fetch.prunetags")) {
|
|
fetch_prune_tags_config = git_config_bool(k, v);
|
|
return 0;
|
|
}
|
|
|
|
if (!strcmp(k, "fetch.showforcedupdates")) {
|
|
fetch_show_forced_updates = git_config_bool(k, v);
|
|
return 0;
|
|
}
|
|
|
|
if (!strcmp(k, "submodule.recurse")) {
|
|
int r = git_config_bool(k, v) ?
|
|
RECURSE_SUBMODULES_ON : RECURSE_SUBMODULES_OFF;
|
|
recurse_submodules = r;
|
|
}
|
|
|
|
if (!strcmp(k, "submodule.fetchjobs")) {
|
|
submodule_fetch_jobs_config = parse_submodule_fetchjobs(k, v);
|
|
return 0;
|
|
} else if (!strcmp(k, "fetch.recursesubmodules")) {
|
|
recurse_submodules = parse_fetch_recurse_submodules_arg(k, v);
|
|
return 0;
|
|
}
|
|
|
|
if (!strcmp(k, "fetch.parallel")) {
|
|
fetch_parallel_config = git_config_int(k, v);
|
|
if (fetch_parallel_config < 0)
|
|
die(_("fetch.parallel cannot be negative"));
|
|
return 0;
|
|
}
|
|
|
|
return git_default_config(k, v, cb);
|
|
}
|
|
|
|
static int parse_refmap_arg(const struct option *opt, const char *arg, int unset)
|
|
{
|
|
BUG_ON_OPT_NEG(unset);
|
|
|
|
/*
|
|
* "git fetch --refmap='' origin foo"
|
|
* can be used to tell the command not to store anywhere
|
|
*/
|
|
refspec_append(&refmap, arg);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct option builtin_fetch_options[] = {
|
|
OPT__VERBOSITY(&verbosity),
|
|
OPT_BOOL(0, "all", &all,
|
|
N_("fetch from all remotes")),
|
|
OPT_BOOL(0, "set-upstream", &set_upstream,
|
|
N_("set upstream for git pull/fetch")),
|
|
OPT_BOOL('a', "append", &append,
|
|
N_("append to .git/FETCH_HEAD instead of overwriting")),
|
|
OPT_BOOL(0, "atomic", &atomic_fetch,
|
|
N_("use atomic transaction to update references")),
|
|
OPT_STRING(0, "upload-pack", &upload_pack, N_("path"),
|
|
N_("path to upload pack on remote end")),
|
|
OPT__FORCE(&force, N_("force overwrite of local reference"), 0),
|
|
OPT_BOOL('m', "multiple", &multiple,
|
|
N_("fetch from multiple remotes")),
|
|
OPT_SET_INT('t', "tags", &tags,
|
|
N_("fetch all tags and associated objects"), TAGS_SET),
|
|
OPT_SET_INT('n', NULL, &tags,
|
|
N_("do not fetch all tags (--no-tags)"), TAGS_UNSET),
|
|
OPT_INTEGER('j', "jobs", &max_jobs,
|
|
N_("number of submodules fetched in parallel")),
|
|
OPT_BOOL(0, "prefetch", &prefetch,
|
|
N_("modify the refspec to place all refs within refs/prefetch/")),
|
|
OPT_BOOL('p', "prune", &prune,
|
|
N_("prune remote-tracking branches no longer on remote")),
|
|
OPT_BOOL('P', "prune-tags", &prune_tags,
|
|
N_("prune local tags no longer on remote and clobber changed tags")),
|
|
OPT_CALLBACK_F(0, "recurse-submodules", &recurse_submodules, N_("on-demand"),
|
|
N_("control recursive fetching of submodules"),
|
|
PARSE_OPT_OPTARG, option_fetch_parse_recurse_submodules),
|
|
OPT_BOOL(0, "dry-run", &dry_run,
|
|
N_("dry run")),
|
|
OPT_BOOL(0, "write-fetch-head", &write_fetch_head,
|
|
N_("write fetched references to the FETCH_HEAD file")),
|
|
OPT_BOOL('k', "keep", &keep, N_("keep downloaded pack")),
|
|
OPT_BOOL('u', "update-head-ok", &update_head_ok,
|
|
N_("allow updating of HEAD ref")),
|
|
OPT_BOOL(0, "progress", &progress, N_("force progress reporting")),
|
|
OPT_STRING(0, "depth", &depth, N_("depth"),
|
|
N_("deepen history of shallow clone")),
|
|
OPT_STRING(0, "shallow-since", &deepen_since, N_("time"),
|
|
N_("deepen history of shallow repository based on time")),
|
|
OPT_STRING_LIST(0, "shallow-exclude", &deepen_not, N_("revision"),
|
|
N_("deepen history of shallow clone, excluding rev")),
|
|
OPT_INTEGER(0, "deepen", &deepen_relative,
|
|
N_("deepen history of shallow clone")),
|
|
OPT_SET_INT_F(0, "unshallow", &unshallow,
|
|
N_("convert to a complete repository"),
|
|
1, PARSE_OPT_NONEG),
|
|
{ OPTION_STRING, 0, "submodule-prefix", &submodule_prefix, N_("dir"),
|
|
N_("prepend this to submodule path output"), PARSE_OPT_HIDDEN },
|
|
OPT_CALLBACK_F(0, "recurse-submodules-default",
|
|
&recurse_submodules_default, N_("on-demand"),
|
|
N_("default for recursive fetching of submodules "
|
|
"(lower priority than config files)"),
|
|
PARSE_OPT_HIDDEN, option_fetch_parse_recurse_submodules),
|
|
OPT_BOOL(0, "update-shallow", &update_shallow,
|
|
N_("accept refs that update .git/shallow")),
|
|
OPT_CALLBACK_F(0, "refmap", NULL, N_("refmap"),
|
|
N_("specify fetch refmap"), PARSE_OPT_NONEG, parse_refmap_arg),
|
|
OPT_STRING_LIST('o', "server-option", &server_options, N_("server-specific"), N_("option to transmit")),
|
|
OPT_SET_INT('4', "ipv4", &family, N_("use IPv4 addresses only"),
|
|
TRANSPORT_FAMILY_IPV4),
|
|
OPT_SET_INT('6', "ipv6", &family, N_("use IPv6 addresses only"),
|
|
TRANSPORT_FAMILY_IPV6),
|
|
OPT_STRING_LIST(0, "negotiation-tip", &negotiation_tip, N_("revision"),
|
|
N_("report that we have only objects reachable from this object")),
|
|
OPT_BOOL(0, "negotiate-only", &negotiate_only,
|
|
N_("do not fetch a packfile; instead, print ancestors of negotiation tips")),
|
|
OPT_PARSE_LIST_OBJECTS_FILTER(&filter_options),
|
|
OPT_BOOL(0, "auto-maintenance", &enable_auto_gc,
|
|
N_("run 'maintenance --auto' after fetching")),
|
|
OPT_BOOL(0, "auto-gc", &enable_auto_gc,
|
|
N_("run 'maintenance --auto' after fetching")),
|
|
OPT_BOOL(0, "show-forced-updates", &fetch_show_forced_updates,
|
|
N_("check for forced-updates on all updated branches")),
|
|
OPT_BOOL(0, "write-commit-graph", &fetch_write_commit_graph,
|
|
N_("write the commit-graph after fetching")),
|
|
OPT_BOOL(0, "stdin", &stdin_refspecs,
|
|
N_("accept refspecs from stdin")),
|
|
OPT_END()
|
|
};
|
|
|
|
static void unlock_pack(void)
|
|
{
|
|
if (gtransport)
|
|
transport_unlock_pack(gtransport);
|
|
if (gsecondary)
|
|
transport_unlock_pack(gsecondary);
|
|
}
|
|
|
|
static void unlock_pack_on_signal(int signo)
|
|
{
|
|
unlock_pack();
|
|
sigchain_pop(signo);
|
|
raise(signo);
|
|
}
|
|
|
|
static void add_merge_config(struct ref **head,
|
|
const struct ref *remote_refs,
|
|
struct branch *branch,
|
|
struct ref ***tail)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < branch->merge_nr; i++) {
|
|
struct ref *rm, **old_tail = *tail;
|
|
struct refspec_item refspec;
|
|
|
|
for (rm = *head; rm; rm = rm->next) {
|
|
if (branch_merge_matches(branch, i, rm->name)) {
|
|
rm->fetch_head_status = FETCH_HEAD_MERGE;
|
|
break;
|
|
}
|
|
}
|
|
if (rm)
|
|
continue;
|
|
|
|
/*
|
|
* Not fetched to a remote-tracking branch? We need to fetch
|
|
* it anyway to allow this branch's "branch.$name.merge"
|
|
* to be honored by 'git pull', but we do not have to
|
|
* fail if branch.$name.merge is misconfigured to point
|
|
* at a nonexisting branch. If we were indeed called by
|
|
* 'git pull', it will notice the misconfiguration because
|
|
* there is no entry in the resulting FETCH_HEAD marked
|
|
* for merging.
|
|
*/
|
|
memset(&refspec, 0, sizeof(refspec));
|
|
refspec.src = branch->merge[i]->src;
|
|
get_fetch_map(remote_refs, &refspec, tail, 1);
|
|
for (rm = *old_tail; rm; rm = rm->next)
|
|
rm->fetch_head_status = FETCH_HEAD_MERGE;
|
|
}
|
|
}
|
|
|
|
static void create_fetch_oidset(struct ref **head, struct oidset *out)
|
|
{
|
|
struct ref *rm = *head;
|
|
while (rm) {
|
|
oidset_insert(out, &rm->old_oid);
|
|
rm = rm->next;
|
|
}
|
|
}
|
|
|
|
struct refname_hash_entry {
|
|
struct hashmap_entry ent;
|
|
struct object_id oid;
|
|
int ignore;
|
|
char refname[FLEX_ARRAY];
|
|
};
|
|
|
|
static int refname_hash_entry_cmp(const void *hashmap_cmp_fn_data,
|
|
const struct hashmap_entry *eptr,
|
|
const struct hashmap_entry *entry_or_key,
|
|
const void *keydata)
|
|
{
|
|
const struct refname_hash_entry *e1, *e2;
|
|
|
|
e1 = container_of(eptr, const struct refname_hash_entry, ent);
|
|
e2 = container_of(entry_or_key, const struct refname_hash_entry, ent);
|
|
return strcmp(e1->refname, keydata ? keydata : e2->refname);
|
|
}
|
|
|
|
static struct refname_hash_entry *refname_hash_add(struct hashmap *map,
|
|
const char *refname,
|
|
const struct object_id *oid)
|
|
{
|
|
struct refname_hash_entry *ent;
|
|
size_t len = strlen(refname);
|
|
|
|
FLEX_ALLOC_MEM(ent, refname, refname, len);
|
|
hashmap_entry_init(&ent->ent, strhash(refname));
|
|
oidcpy(&ent->oid, oid);
|
|
hashmap_add(map, &ent->ent);
|
|
return ent;
|
|
}
|
|
|
|
static int add_one_refname(const char *refname,
|
|
const struct object_id *oid,
|
|
int flag, void *cbdata)
|
|
{
|
|
struct hashmap *refname_map = cbdata;
|
|
|
|
(void) refname_hash_add(refname_map, refname, oid);
|
|
return 0;
|
|
}
|
|
|
|
static void refname_hash_init(struct hashmap *map)
|
|
{
|
|
hashmap_init(map, refname_hash_entry_cmp, NULL, 0);
|
|
}
|
|
|
|
static int refname_hash_exists(struct hashmap *map, const char *refname)
|
|
{
|
|
return !!hashmap_get_from_hash(map, strhash(refname), refname);
|
|
}
|
|
|
|
static void clear_item(struct refname_hash_entry *item)
|
|
{
|
|
item->ignore = 1;
|
|
}
|
|
|
|
static void find_non_local_tags(const struct ref *refs,
|
|
struct ref **head,
|
|
struct ref ***tail)
|
|
{
|
|
struct hashmap existing_refs;
|
|
struct hashmap remote_refs;
|
|
struct oidset fetch_oids = OIDSET_INIT;
|
|
struct string_list remote_refs_list = STRING_LIST_INIT_NODUP;
|
|
struct string_list_item *remote_ref_item;
|
|
const struct ref *ref;
|
|
struct refname_hash_entry *item = NULL;
|
|
const int quick_flags = OBJECT_INFO_QUICK | OBJECT_INFO_SKIP_FETCH_OBJECT;
|
|
|
|
refname_hash_init(&existing_refs);
|
|
refname_hash_init(&remote_refs);
|
|
create_fetch_oidset(head, &fetch_oids);
|
|
|
|
for_each_ref(add_one_refname, &existing_refs);
|
|
for (ref = refs; ref; ref = ref->next) {
|
|
if (!starts_with(ref->name, "refs/tags/"))
|
|
continue;
|
|
|
|
/*
|
|
* The peeled ref always follows the matching base
|
|
* ref, so if we see a peeled ref that we don't want
|
|
* to fetch then we can mark the ref entry in the list
|
|
* as one to ignore by setting util to NULL.
|
|
*/
|
|
if (ends_with(ref->name, "^{}")) {
|
|
if (item &&
|
|
!has_object_file_with_flags(&ref->old_oid, quick_flags) &&
|
|
!oidset_contains(&fetch_oids, &ref->old_oid) &&
|
|
!has_object_file_with_flags(&item->oid, quick_flags) &&
|
|
!oidset_contains(&fetch_oids, &item->oid))
|
|
clear_item(item);
|
|
item = NULL;
|
|
continue;
|
|
}
|
|
|
|
/*
|
|
* If item is non-NULL here, then we previously saw a
|
|
* ref not followed by a peeled reference, so we need
|
|
* to check if it is a lightweight tag that we want to
|
|
* fetch.
|
|
*/
|
|
if (item &&
|
|
!has_object_file_with_flags(&item->oid, quick_flags) &&
|
|
!oidset_contains(&fetch_oids, &item->oid))
|
|
clear_item(item);
|
|
|
|
item = NULL;
|
|
|
|
/* skip duplicates and refs that we already have */
|
|
if (refname_hash_exists(&remote_refs, ref->name) ||
|
|
refname_hash_exists(&existing_refs, ref->name))
|
|
continue;
|
|
|
|
item = refname_hash_add(&remote_refs, ref->name, &ref->old_oid);
|
|
string_list_insert(&remote_refs_list, ref->name);
|
|
}
|
|
hashmap_clear_and_free(&existing_refs, struct refname_hash_entry, ent);
|
|
|
|
/*
|
|
* We may have a final lightweight tag that needs to be
|
|
* checked to see if it needs fetching.
|
|
*/
|
|
if (item &&
|
|
!has_object_file_with_flags(&item->oid, quick_flags) &&
|
|
!oidset_contains(&fetch_oids, &item->oid))
|
|
clear_item(item);
|
|
|
|
/*
|
|
* For all the tags in the remote_refs_list,
|
|
* add them to the list of refs to be fetched
|
|
*/
|
|
for_each_string_list_item(remote_ref_item, &remote_refs_list) {
|
|
const char *refname = remote_ref_item->string;
|
|
struct ref *rm;
|
|
unsigned int hash = strhash(refname);
|
|
|
|
item = hashmap_get_entry_from_hash(&remote_refs, hash, refname,
|
|
struct refname_hash_entry, ent);
|
|
if (!item)
|
|
BUG("unseen remote ref?");
|
|
|
|
/* Unless we have already decided to ignore this item... */
|
|
if (item->ignore)
|
|
continue;
|
|
|
|
rm = alloc_ref(item->refname);
|
|
rm->peer_ref = alloc_ref(item->refname);
|
|
oidcpy(&rm->old_oid, &item->oid);
|
|
**tail = rm;
|
|
*tail = &rm->next;
|
|
}
|
|
hashmap_clear_and_free(&remote_refs, struct refname_hash_entry, ent);
|
|
string_list_clear(&remote_refs_list, 0);
|
|
oidset_clear(&fetch_oids);
|
|
}
|
|
|
|
static void filter_prefetch_refspec(struct refspec *rs)
|
|
{
|
|
int i;
|
|
|
|
if (!prefetch)
|
|
return;
|
|
|
|
for (i = 0; i < rs->nr; i++) {
|
|
struct strbuf new_dst = STRBUF_INIT;
|
|
char *old_dst;
|
|
const char *sub = NULL;
|
|
|
|
if (rs->items[i].negative)
|
|
continue;
|
|
if (!rs->items[i].dst ||
|
|
(rs->items[i].src &&
|
|
!strncmp(rs->items[i].src, "refs/tags/", 10))) {
|
|
int j;
|
|
|
|
free(rs->items[i].src);
|
|
free(rs->items[i].dst);
|
|
|
|
for (j = i + 1; j < rs->nr; j++) {
|
|
rs->items[j - 1] = rs->items[j];
|
|
rs->raw[j - 1] = rs->raw[j];
|
|
}
|
|
rs->nr--;
|
|
i--;
|
|
continue;
|
|
}
|
|
|
|
old_dst = rs->items[i].dst;
|
|
strbuf_addstr(&new_dst, "refs/prefetch/");
|
|
|
|
/*
|
|
* If old_dst starts with "refs/", then place
|
|
* sub after that prefix. Otherwise, start at
|
|
* the beginning of the string.
|
|
*/
|
|
if (!skip_prefix(old_dst, "refs/", &sub))
|
|
sub = old_dst;
|
|
strbuf_addstr(&new_dst, sub);
|
|
|
|
rs->items[i].dst = strbuf_detach(&new_dst, NULL);
|
|
rs->items[i].force = 1;
|
|
|
|
free(old_dst);
|
|
}
|
|
}
|
|
|
|
static struct ref *get_ref_map(struct remote *remote,
|
|
const struct ref *remote_refs,
|
|
struct refspec *rs,
|
|
int tags, int *autotags)
|
|
{
|
|
int i;
|
|
struct ref *rm;
|
|
struct ref *ref_map = NULL;
|
|
struct ref **tail = &ref_map;
|
|
|
|
/* opportunistically-updated references: */
|
|
struct ref *orefs = NULL, **oref_tail = &orefs;
|
|
|
|
struct hashmap existing_refs;
|
|
int existing_refs_populated = 0;
|
|
|
|
filter_prefetch_refspec(rs);
|
|
if (remote)
|
|
filter_prefetch_refspec(&remote->fetch);
|
|
|
|
if (rs->nr) {
|
|
struct refspec *fetch_refspec;
|
|
|
|
for (i = 0; i < rs->nr; i++) {
|
|
get_fetch_map(remote_refs, &rs->items[i], &tail, 0);
|
|
if (rs->items[i].dst && rs->items[i].dst[0])
|
|
*autotags = 1;
|
|
}
|
|
/* Merge everything on the command line (but not --tags) */
|
|
for (rm = ref_map; rm; rm = rm->next)
|
|
rm->fetch_head_status = FETCH_HEAD_MERGE;
|
|
|
|
/*
|
|
* For any refs that we happen to be fetching via
|
|
* command-line arguments, the destination ref might
|
|
* have been missing or have been different than the
|
|
* remote-tracking ref that would be derived from the
|
|
* configured refspec. In these cases, we want to
|
|
* take the opportunity to update their configured
|
|
* remote-tracking reference. However, we do not want
|
|
* to mention these entries in FETCH_HEAD at all, as
|
|
* they would simply be duplicates of existing
|
|
* entries, so we set them FETCH_HEAD_IGNORE below.
|
|
*
|
|
* We compute these entries now, based only on the
|
|
* refspecs specified on the command line. But we add
|
|
* them to the list following the refspecs resulting
|
|
* from the tags option so that one of the latter,
|
|
* which has FETCH_HEAD_NOT_FOR_MERGE, is not removed
|
|
* by ref_remove_duplicates() in favor of one of these
|
|
* opportunistic entries with FETCH_HEAD_IGNORE.
|
|
*/
|
|
if (refmap.nr)
|
|
fetch_refspec = &refmap;
|
|
else
|
|
fetch_refspec = &remote->fetch;
|
|
|
|
for (i = 0; i < fetch_refspec->nr; i++)
|
|
get_fetch_map(ref_map, &fetch_refspec->items[i], &oref_tail, 1);
|
|
} else if (refmap.nr) {
|
|
die("--refmap option is only meaningful with command-line refspec(s).");
|
|
} else {
|
|
/* Use the defaults */
|
|
struct branch *branch = branch_get(NULL);
|
|
int has_merge = branch_has_merge_config(branch);
|
|
if (remote &&
|
|
(remote->fetch.nr ||
|
|
/* Note: has_merge implies non-NULL branch->remote_name */
|
|
(has_merge && !strcmp(branch->remote_name, remote->name)))) {
|
|
for (i = 0; i < remote->fetch.nr; i++) {
|
|
get_fetch_map(remote_refs, &remote->fetch.items[i], &tail, 0);
|
|
if (remote->fetch.items[i].dst &&
|
|
remote->fetch.items[i].dst[0])
|
|
*autotags = 1;
|
|
if (!i && !has_merge && ref_map &&
|
|
!remote->fetch.items[0].pattern)
|
|
ref_map->fetch_head_status = FETCH_HEAD_MERGE;
|
|
}
|
|
/*
|
|
* if the remote we're fetching from is the same
|
|
* as given in branch.<name>.remote, we add the
|
|
* ref given in branch.<name>.merge, too.
|
|
*
|
|
* Note: has_merge implies non-NULL branch->remote_name
|
|
*/
|
|
if (has_merge &&
|
|
!strcmp(branch->remote_name, remote->name))
|
|
add_merge_config(&ref_map, remote_refs, branch, &tail);
|
|
} else if (!prefetch) {
|
|
ref_map = get_remote_ref(remote_refs, "HEAD");
|
|
if (!ref_map)
|
|
die(_("Couldn't find remote ref HEAD"));
|
|
ref_map->fetch_head_status = FETCH_HEAD_MERGE;
|
|
tail = &ref_map->next;
|
|
}
|
|
}
|
|
|
|
if (tags == TAGS_SET)
|
|
/* also fetch all tags */
|
|
get_fetch_map(remote_refs, tag_refspec, &tail, 0);
|
|
else if (tags == TAGS_DEFAULT && *autotags)
|
|
find_non_local_tags(remote_refs, &ref_map, &tail);
|
|
|
|
/* Now append any refs to be updated opportunistically: */
|
|
*tail = orefs;
|
|
for (rm = orefs; rm; rm = rm->next) {
|
|
rm->fetch_head_status = FETCH_HEAD_IGNORE;
|
|
tail = &rm->next;
|
|
}
|
|
|
|
/*
|
|
* apply negative refspecs first, before we remove duplicates. This is
|
|
* necessary as negative refspecs might remove an otherwise conflicting
|
|
* duplicate.
|
|
*/
|
|
if (rs->nr)
|
|
ref_map = apply_negative_refspecs(ref_map, rs);
|
|
else
|
|
ref_map = apply_negative_refspecs(ref_map, &remote->fetch);
|
|
|
|
ref_map = ref_remove_duplicates(ref_map);
|
|
|
|
for (rm = ref_map; rm; rm = rm->next) {
|
|
if (rm->peer_ref) {
|
|
const char *refname = rm->peer_ref->name;
|
|
struct refname_hash_entry *peer_item;
|
|
unsigned int hash = strhash(refname);
|
|
|
|
if (!existing_refs_populated) {
|
|
refname_hash_init(&existing_refs);
|
|
for_each_ref(add_one_refname, &existing_refs);
|
|
existing_refs_populated = 1;
|
|
}
|
|
|
|
peer_item = hashmap_get_entry_from_hash(&existing_refs,
|
|
hash, refname,
|
|
struct refname_hash_entry, ent);
|
|
if (peer_item) {
|
|
struct object_id *old_oid = &peer_item->oid;
|
|
oidcpy(&rm->peer_ref->old_oid, old_oid);
|
|
}
|
|
}
|
|
}
|
|
if (existing_refs_populated)
|
|
hashmap_clear_and_free(&existing_refs, struct refname_hash_entry, ent);
|
|
|
|
return ref_map;
|
|
}
|
|
|
|
#define STORE_REF_ERROR_OTHER 1
|
|
#define STORE_REF_ERROR_DF_CONFLICT 2
|
|
|
|
static int s_update_ref(const char *action,
|
|
struct ref *ref,
|
|
struct ref_transaction *transaction,
|
|
int check_old)
|
|
{
|
|
char *msg;
|
|
char *rla = getenv("GIT_REFLOG_ACTION");
|
|
struct ref_transaction *our_transaction = NULL;
|
|
struct strbuf err = STRBUF_INIT;
|
|
int ret;
|
|
|
|
if (dry_run)
|
|
return 0;
|
|
if (!rla)
|
|
rla = default_rla.buf;
|
|
msg = xstrfmt("%s: %s", rla, action);
|
|
|
|
/*
|
|
* If no transaction was passed to us, we manage the transaction
|
|
* ourselves. Otherwise, we trust the caller to handle the transaction
|
|
* lifecycle.
|
|
*/
|
|
if (!transaction) {
|
|
transaction = our_transaction = ref_transaction_begin(&err);
|
|
if (!transaction) {
|
|
ret = STORE_REF_ERROR_OTHER;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
ret = ref_transaction_update(transaction, ref->name, &ref->new_oid,
|
|
check_old ? &ref->old_oid : NULL,
|
|
0, msg, &err);
|
|
if (ret) {
|
|
ret = STORE_REF_ERROR_OTHER;
|
|
goto out;
|
|
}
|
|
|
|
if (our_transaction) {
|
|
switch (ref_transaction_commit(our_transaction, &err)) {
|
|
case 0:
|
|
break;
|
|
case TRANSACTION_NAME_CONFLICT:
|
|
ret = STORE_REF_ERROR_DF_CONFLICT;
|
|
goto out;
|
|
default:
|
|
ret = STORE_REF_ERROR_OTHER;
|
|
goto out;
|
|
}
|
|
}
|
|
|
|
out:
|
|
ref_transaction_free(our_transaction);
|
|
if (ret)
|
|
error("%s", err.buf);
|
|
strbuf_release(&err);
|
|
free(msg);
|
|
return ret;
|
|
}
|
|
|
|
static int refcol_width = 10;
|
|
static int compact_format;
|
|
|
|
static void adjust_refcol_width(const struct ref *ref)
|
|
{
|
|
int max, rlen, llen, len;
|
|
|
|
/* uptodate lines are only shown on high verbosity level */
|
|
if (!verbosity && oideq(&ref->peer_ref->old_oid, &ref->old_oid))
|
|
return;
|
|
|
|
max = term_columns();
|
|
rlen = utf8_strwidth(prettify_refname(ref->name));
|
|
|
|
llen = utf8_strwidth(prettify_refname(ref->peer_ref->name));
|
|
|
|
/*
|
|
* rough estimation to see if the output line is too long and
|
|
* should not be counted (we can't do precise calculation
|
|
* anyway because we don't know if the error explanation part
|
|
* will be printed in update_local_ref)
|
|
*/
|
|
if (compact_format) {
|
|
llen = 0;
|
|
max = max * 2 / 3;
|
|
}
|
|
len = 21 /* flag and summary */ + rlen + 4 /* -> */ + llen;
|
|
if (len >= max)
|
|
return;
|
|
|
|
/*
|
|
* Not precise calculation for compact mode because '*' can
|
|
* appear on the left hand side of '->' and shrink the column
|
|
* back.
|
|
*/
|
|
if (refcol_width < rlen)
|
|
refcol_width = rlen;
|
|
}
|
|
|
|
static void prepare_format_display(struct ref *ref_map)
|
|
{
|
|
struct ref *rm;
|
|
const char *format = "full";
|
|
|
|
git_config_get_string_tmp("fetch.output", &format);
|
|
if (!strcasecmp(format, "full"))
|
|
compact_format = 0;
|
|
else if (!strcasecmp(format, "compact"))
|
|
compact_format = 1;
|
|
else
|
|
die(_("configuration fetch.output contains invalid value %s"),
|
|
format);
|
|
|
|
for (rm = ref_map; rm; rm = rm->next) {
|
|
if (rm->status == REF_STATUS_REJECT_SHALLOW ||
|
|
!rm->peer_ref ||
|
|
!strcmp(rm->name, "HEAD"))
|
|
continue;
|
|
|
|
adjust_refcol_width(rm);
|
|
}
|
|
}
|
|
|
|
static void print_remote_to_local(struct strbuf *display,
|
|
const char *remote, const char *local)
|
|
{
|
|
strbuf_addf(display, "%-*s -> %s", refcol_width, remote, local);
|
|
}
|
|
|
|
static int find_and_replace(struct strbuf *haystack,
|
|
const char *needle,
|
|
const char *placeholder)
|
|
{
|
|
const char *p = NULL;
|
|
int plen, nlen;
|
|
|
|
nlen = strlen(needle);
|
|
if (ends_with(haystack->buf, needle))
|
|
p = haystack->buf + haystack->len - nlen;
|
|
else
|
|
p = strstr(haystack->buf, needle);
|
|
if (!p)
|
|
return 0;
|
|
|
|
if (p > haystack->buf && p[-1] != '/')
|
|
return 0;
|
|
|
|
plen = strlen(p);
|
|
if (plen > nlen && p[nlen] != '/')
|
|
return 0;
|
|
|
|
strbuf_splice(haystack, p - haystack->buf, nlen,
|
|
placeholder, strlen(placeholder));
|
|
return 1;
|
|
}
|
|
|
|
static void print_compact(struct strbuf *display,
|
|
const char *remote, const char *local)
|
|
{
|
|
struct strbuf r = STRBUF_INIT;
|
|
struct strbuf l = STRBUF_INIT;
|
|
|
|
if (!strcmp(remote, local)) {
|
|
strbuf_addf(display, "%-*s -> *", refcol_width, remote);
|
|
return;
|
|
}
|
|
|
|
strbuf_addstr(&r, remote);
|
|
strbuf_addstr(&l, local);
|
|
|
|
if (!find_and_replace(&r, local, "*"))
|
|
find_and_replace(&l, remote, "*");
|
|
print_remote_to_local(display, r.buf, l.buf);
|
|
|
|
strbuf_release(&r);
|
|
strbuf_release(&l);
|
|
}
|
|
|
|
static void format_display(struct strbuf *display, char code,
|
|
const char *summary, const char *error,
|
|
const char *remote, const char *local,
|
|
int summary_width)
|
|
{
|
|
int width = (summary_width + strlen(summary) - gettext_width(summary));
|
|
|
|
strbuf_addf(display, "%c %-*s ", code, width, summary);
|
|
if (!compact_format)
|
|
print_remote_to_local(display, remote, local);
|
|
else
|
|
print_compact(display, remote, local);
|
|
if (error)
|
|
strbuf_addf(display, " (%s)", error);
|
|
}
|
|
|
|
static int update_local_ref(struct ref *ref,
|
|
struct ref_transaction *transaction,
|
|
const char *remote,
|
|
const struct ref *remote_ref,
|
|
struct strbuf *display,
|
|
int summary_width)
|
|
{
|
|
struct commit *current = NULL, *updated;
|
|
enum object_type type;
|
|
struct branch *current_branch = branch_get(NULL);
|
|
const char *pretty_ref = prettify_refname(ref->name);
|
|
int fast_forward = 0;
|
|
|
|
type = oid_object_info(the_repository, &ref->new_oid, NULL);
|
|
if (type < 0)
|
|
die(_("object %s not found"), oid_to_hex(&ref->new_oid));
|
|
|
|
if (oideq(&ref->old_oid, &ref->new_oid)) {
|
|
if (verbosity > 0)
|
|
format_display(display, '=', _("[up to date]"), NULL,
|
|
remote, pretty_ref, summary_width);
|
|
return 0;
|
|
}
|
|
|
|
if (current_branch &&
|
|
!strcmp(ref->name, current_branch->name) &&
|
|
!(update_head_ok || is_bare_repository()) &&
|
|
!is_null_oid(&ref->old_oid)) {
|
|
/*
|
|
* If this is the head, and it's not okay to update
|
|
* the head, and the old value of the head isn't empty...
|
|
*/
|
|
format_display(display, '!', _("[rejected]"),
|
|
_("can't fetch in current branch"),
|
|
remote, pretty_ref, summary_width);
|
|
return 1;
|
|
}
|
|
|
|
if (!is_null_oid(&ref->old_oid) &&
|
|
starts_with(ref->name, "refs/tags/")) {
|
|
if (force || ref->force) {
|
|
int r;
|
|
r = s_update_ref("updating tag", ref, transaction, 0);
|
|
format_display(display, r ? '!' : 't', _("[tag update]"),
|
|
r ? _("unable to update local ref") : NULL,
|
|
remote, pretty_ref, summary_width);
|
|
return r;
|
|
} else {
|
|
format_display(display, '!', _("[rejected]"), _("would clobber existing tag"),
|
|
remote, pretty_ref, summary_width);
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
current = lookup_commit_reference_gently(the_repository,
|
|
&ref->old_oid, 1);
|
|
updated = lookup_commit_reference_gently(the_repository,
|
|
&ref->new_oid, 1);
|
|
if (!current || !updated) {
|
|
const char *msg;
|
|
const char *what;
|
|
int r;
|
|
/*
|
|
* Nicely describe the new ref we're fetching.
|
|
* Base this on the remote's ref name, as it's
|
|
* more likely to follow a standard layout.
|
|
*/
|
|
const char *name = remote_ref ? remote_ref->name : "";
|
|
if (starts_with(name, "refs/tags/")) {
|
|
msg = "storing tag";
|
|
what = _("[new tag]");
|
|
} else if (starts_with(name, "refs/heads/")) {
|
|
msg = "storing head";
|
|
what = _("[new branch]");
|
|
} else {
|
|
msg = "storing ref";
|
|
what = _("[new ref]");
|
|
}
|
|
|
|
r = s_update_ref(msg, ref, transaction, 0);
|
|
format_display(display, r ? '!' : '*', what,
|
|
r ? _("unable to update local ref") : NULL,
|
|
remote, pretty_ref, summary_width);
|
|
return r;
|
|
}
|
|
|
|
if (fetch_show_forced_updates) {
|
|
uint64_t t_before = getnanotime();
|
|
fast_forward = in_merge_bases(current, updated);
|
|
forced_updates_ms += (getnanotime() - t_before) / 1000000;
|
|
} else {
|
|
fast_forward = 1;
|
|
}
|
|
|
|
if (fast_forward) {
|
|
struct strbuf quickref = STRBUF_INIT;
|
|
int r;
|
|
|
|
strbuf_add_unique_abbrev(&quickref, ¤t->object.oid, DEFAULT_ABBREV);
|
|
strbuf_addstr(&quickref, "..");
|
|
strbuf_add_unique_abbrev(&quickref, &ref->new_oid, DEFAULT_ABBREV);
|
|
r = s_update_ref("fast-forward", ref, transaction, 1);
|
|
format_display(display, r ? '!' : ' ', quickref.buf,
|
|
r ? _("unable to update local ref") : NULL,
|
|
remote, pretty_ref, summary_width);
|
|
strbuf_release(&quickref);
|
|
return r;
|
|
} else if (force || ref->force) {
|
|
struct strbuf quickref = STRBUF_INIT;
|
|
int r;
|
|
strbuf_add_unique_abbrev(&quickref, ¤t->object.oid, DEFAULT_ABBREV);
|
|
strbuf_addstr(&quickref, "...");
|
|
strbuf_add_unique_abbrev(&quickref, &ref->new_oid, DEFAULT_ABBREV);
|
|
r = s_update_ref("forced-update", ref, transaction, 1);
|
|
format_display(display, r ? '!' : '+', quickref.buf,
|
|
r ? _("unable to update local ref") : _("forced update"),
|
|
remote, pretty_ref, summary_width);
|
|
strbuf_release(&quickref);
|
|
return r;
|
|
} else {
|
|
format_display(display, '!', _("[rejected]"), _("non-fast-forward"),
|
|
remote, pretty_ref, summary_width);
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
static int iterate_ref_map(void *cb_data, struct object_id *oid)
|
|
{
|
|
struct ref **rm = cb_data;
|
|
struct ref *ref = *rm;
|
|
|
|
while (ref && ref->status == REF_STATUS_REJECT_SHALLOW)
|
|
ref = ref->next;
|
|
if (!ref)
|
|
return -1; /* end of the list */
|
|
*rm = ref->next;
|
|
oidcpy(oid, &ref->old_oid);
|
|
return 0;
|
|
}
|
|
|
|
struct fetch_head {
|
|
FILE *fp;
|
|
struct strbuf buf;
|
|
};
|
|
|
|
static int open_fetch_head(struct fetch_head *fetch_head)
|
|
{
|
|
const char *filename = git_path_fetch_head(the_repository);
|
|
|
|
if (write_fetch_head) {
|
|
fetch_head->fp = fopen(filename, "a");
|
|
if (!fetch_head->fp)
|
|
return error_errno(_("cannot open %s"), filename);
|
|
strbuf_init(&fetch_head->buf, 0);
|
|
} else {
|
|
fetch_head->fp = NULL;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void append_fetch_head(struct fetch_head *fetch_head,
|
|
const struct object_id *old_oid,
|
|
enum fetch_head_status fetch_head_status,
|
|
const char *note,
|
|
const char *url, size_t url_len)
|
|
{
|
|
char old_oid_hex[GIT_MAX_HEXSZ + 1];
|
|
const char *merge_status_marker;
|
|
size_t i;
|
|
|
|
if (!fetch_head->fp)
|
|
return;
|
|
|
|
switch (fetch_head_status) {
|
|
case FETCH_HEAD_NOT_FOR_MERGE:
|
|
merge_status_marker = "not-for-merge";
|
|
break;
|
|
case FETCH_HEAD_MERGE:
|
|
merge_status_marker = "";
|
|
break;
|
|
default:
|
|
/* do not write anything to FETCH_HEAD */
|
|
return;
|
|
}
|
|
|
|
strbuf_addf(&fetch_head->buf, "%s\t%s\t%s",
|
|
oid_to_hex_r(old_oid_hex, old_oid), merge_status_marker, note);
|
|
for (i = 0; i < url_len; ++i)
|
|
if ('\n' == url[i])
|
|
strbuf_addstr(&fetch_head->buf, "\\n");
|
|
else
|
|
strbuf_addch(&fetch_head->buf, url[i]);
|
|
strbuf_addch(&fetch_head->buf, '\n');
|
|
|
|
/*
|
|
* When using an atomic fetch, we do not want to update FETCH_HEAD if
|
|
* any of the reference updates fails. We thus have to write all
|
|
* updates to a buffer first and only commit it as soon as all
|
|
* references have been successfully updated.
|
|
*/
|
|
if (!atomic_fetch) {
|
|
strbuf_write(&fetch_head->buf, fetch_head->fp);
|
|
strbuf_reset(&fetch_head->buf);
|
|
}
|
|
}
|
|
|
|
static void commit_fetch_head(struct fetch_head *fetch_head)
|
|
{
|
|
if (!fetch_head->fp || !atomic_fetch)
|
|
return;
|
|
strbuf_write(&fetch_head->buf, fetch_head->fp);
|
|
}
|
|
|
|
static void close_fetch_head(struct fetch_head *fetch_head)
|
|
{
|
|
if (!fetch_head->fp)
|
|
return;
|
|
|
|
fclose(fetch_head->fp);
|
|
strbuf_release(&fetch_head->buf);
|
|
}
|
|
|
|
static const char warn_show_forced_updates[] =
|
|
N_("Fetch normally indicates which branches had a forced update,\n"
|
|
"but that check has been disabled. To re-enable, use '--show-forced-updates'\n"
|
|
"flag or run 'git config fetch.showForcedUpdates true'.");
|
|
static const char warn_time_show_forced_updates[] =
|
|
N_("It took %.2f seconds to check forced updates. You can use\n"
|
|
"'--no-show-forced-updates' or run 'git config fetch.showForcedUpdates false'\n"
|
|
" to avoid this check.\n");
|
|
|
|
static int store_updated_refs(const char *raw_url, const char *remote_name,
|
|
int connectivity_checked, struct ref *ref_map)
|
|
{
|
|
struct fetch_head fetch_head;
|
|
struct commit *commit;
|
|
int url_len, i, rc = 0;
|
|
struct strbuf note = STRBUF_INIT, err = STRBUF_INIT;
|
|
struct ref_transaction *transaction = NULL;
|
|
const char *what, *kind;
|
|
struct ref *rm;
|
|
char *url;
|
|
int want_status;
|
|
int summary_width = transport_summary_width(ref_map);
|
|
|
|
rc = open_fetch_head(&fetch_head);
|
|
if (rc)
|
|
return -1;
|
|
|
|
if (raw_url)
|
|
url = transport_anonymize_url(raw_url);
|
|
else
|
|
url = xstrdup("foreign");
|
|
|
|
if (!connectivity_checked) {
|
|
struct check_connected_options opt = CHECK_CONNECTED_INIT;
|
|
|
|
rm = ref_map;
|
|
if (check_connected(iterate_ref_map, &rm, &opt)) {
|
|
rc = error(_("%s did not send all necessary objects\n"), url);
|
|
goto abort;
|
|
}
|
|
}
|
|
|
|
if (atomic_fetch) {
|
|
transaction = ref_transaction_begin(&err);
|
|
if (!transaction) {
|
|
error("%s", err.buf);
|
|
goto abort;
|
|
}
|
|
}
|
|
|
|
prepare_format_display(ref_map);
|
|
|
|
/*
|
|
* We do a pass for each fetch_head_status type in their enum order, so
|
|
* merged entries are written before not-for-merge. That lets readers
|
|
* use FETCH_HEAD as a refname to refer to the ref to be merged.
|
|
*/
|
|
for (want_status = FETCH_HEAD_MERGE;
|
|
want_status <= FETCH_HEAD_IGNORE;
|
|
want_status++) {
|
|
for (rm = ref_map; rm; rm = rm->next) {
|
|
struct ref *ref = NULL;
|
|
|
|
if (rm->status == REF_STATUS_REJECT_SHALLOW) {
|
|
if (want_status == FETCH_HEAD_MERGE)
|
|
warning(_("reject %s because shallow roots are not allowed to be updated"),
|
|
rm->peer_ref ? rm->peer_ref->name : rm->name);
|
|
continue;
|
|
}
|
|
|
|
commit = lookup_commit_reference_gently(the_repository,
|
|
&rm->old_oid,
|
|
1);
|
|
if (!commit)
|
|
rm->fetch_head_status = FETCH_HEAD_NOT_FOR_MERGE;
|
|
|
|
if (rm->fetch_head_status != want_status)
|
|
continue;
|
|
|
|
if (rm->peer_ref) {
|
|
ref = alloc_ref(rm->peer_ref->name);
|
|
oidcpy(&ref->old_oid, &rm->peer_ref->old_oid);
|
|
oidcpy(&ref->new_oid, &rm->old_oid);
|
|
ref->force = rm->peer_ref->force;
|
|
}
|
|
|
|
if (recurse_submodules != RECURSE_SUBMODULES_OFF &&
|
|
(!rm->peer_ref || !oideq(&ref->old_oid, &ref->new_oid))) {
|
|
check_for_new_submodule_commits(&rm->old_oid);
|
|
}
|
|
|
|
if (!strcmp(rm->name, "HEAD")) {
|
|
kind = "";
|
|
what = "";
|
|
}
|
|
else if (skip_prefix(rm->name, "refs/heads/", &what))
|
|
kind = "branch";
|
|
else if (skip_prefix(rm->name, "refs/tags/", &what))
|
|
kind = "tag";
|
|
else if (skip_prefix(rm->name, "refs/remotes/", &what))
|
|
kind = "remote-tracking branch";
|
|
else {
|
|
kind = "";
|
|
what = rm->name;
|
|
}
|
|
|
|
url_len = strlen(url);
|
|
for (i = url_len - 1; url[i] == '/' && 0 <= i; i--)
|
|
;
|
|
url_len = i + 1;
|
|
if (4 < i && !strncmp(".git", url + i - 3, 4))
|
|
url_len = i - 3;
|
|
|
|
strbuf_reset(¬e);
|
|
if (*what) {
|
|
if (*kind)
|
|
strbuf_addf(¬e, "%s ", kind);
|
|
strbuf_addf(¬e, "'%s' of ", what);
|
|
}
|
|
|
|
append_fetch_head(&fetch_head, &rm->old_oid,
|
|
rm->fetch_head_status,
|
|
note.buf, url, url_len);
|
|
|
|
strbuf_reset(¬e);
|
|
if (ref) {
|
|
rc |= update_local_ref(ref, transaction, what,
|
|
rm, ¬e, summary_width);
|
|
free(ref);
|
|
} else if (write_fetch_head || dry_run) {
|
|
/*
|
|
* Display fetches written to FETCH_HEAD (or
|
|
* would be written to FETCH_HEAD, if --dry-run
|
|
* is set).
|
|
*/
|
|
format_display(¬e, '*',
|
|
*kind ? kind : "branch", NULL,
|
|
*what ? what : "HEAD",
|
|
"FETCH_HEAD", summary_width);
|
|
}
|
|
if (note.len) {
|
|
if (verbosity >= 0 && !shown_url) {
|
|
fprintf(stderr, _("From %.*s\n"),
|
|
url_len, url);
|
|
shown_url = 1;
|
|
}
|
|
if (verbosity >= 0)
|
|
fprintf(stderr, " %s\n", note.buf);
|
|
}
|
|
}
|
|
}
|
|
|
|
if (!rc && transaction) {
|
|
rc = ref_transaction_commit(transaction, &err);
|
|
if (rc) {
|
|
error("%s", err.buf);
|
|
goto abort;
|
|
}
|
|
}
|
|
|
|
if (!rc)
|
|
commit_fetch_head(&fetch_head);
|
|
|
|
if (rc & STORE_REF_ERROR_DF_CONFLICT)
|
|
error(_("some local refs could not be updated; try running\n"
|
|
" 'git remote prune %s' to remove any old, conflicting "
|
|
"branches"), remote_name);
|
|
|
|
if (advice_fetch_show_forced_updates) {
|
|
if (!fetch_show_forced_updates) {
|
|
warning(_(warn_show_forced_updates));
|
|
} else if (forced_updates_ms > FORCED_UPDATES_DELAY_WARNING_IN_MS) {
|
|
warning(_(warn_time_show_forced_updates),
|
|
forced_updates_ms / 1000.0);
|
|
}
|
|
}
|
|
|
|
abort:
|
|
strbuf_release(¬e);
|
|
strbuf_release(&err);
|
|
ref_transaction_free(transaction);
|
|
free(url);
|
|
close_fetch_head(&fetch_head);
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* We would want to bypass the object transfer altogether if
|
|
* everything we are going to fetch already exists and is connected
|
|
* locally.
|
|
*/
|
|
static int check_exist_and_connected(struct ref *ref_map)
|
|
{
|
|
struct ref *rm = ref_map;
|
|
struct check_connected_options opt = CHECK_CONNECTED_INIT;
|
|
struct ref *r;
|
|
|
|
/*
|
|
* If we are deepening a shallow clone we already have these
|
|
* objects reachable. Running rev-list here will return with
|
|
* a good (0) exit status and we'll bypass the fetch that we
|
|
* really need to perform. Claiming failure now will ensure
|
|
* we perform the network exchange to deepen our history.
|
|
*/
|
|
if (deepen)
|
|
return -1;
|
|
|
|
/*
|
|
* check_connected() allows objects to merely be promised, but
|
|
* we need all direct targets to exist.
|
|
*/
|
|
for (r = rm; r; r = r->next) {
|
|
if (!has_object_file_with_flags(&r->old_oid,
|
|
OBJECT_INFO_SKIP_FETCH_OBJECT))
|
|
return -1;
|
|
}
|
|
|
|
opt.quiet = 1;
|
|
return check_connected(iterate_ref_map, &rm, &opt);
|
|
}
|
|
|
|
static int fetch_refs(struct transport *transport, struct ref *ref_map)
|
|
{
|
|
int ret = check_exist_and_connected(ref_map);
|
|
if (ret) {
|
|
trace2_region_enter("fetch", "fetch_refs", the_repository);
|
|
ret = transport_fetch_refs(transport, ref_map);
|
|
trace2_region_leave("fetch", "fetch_refs", the_repository);
|
|
}
|
|
if (!ret)
|
|
/*
|
|
* Keep the new pack's ".keep" file around to allow the caller
|
|
* time to update refs to reference the new objects.
|
|
*/
|
|
return 0;
|
|
transport_unlock_pack(transport);
|
|
return ret;
|
|
}
|
|
|
|
/* Update local refs based on the ref values fetched from a remote */
|
|
static int consume_refs(struct transport *transport, struct ref *ref_map)
|
|
{
|
|
int connectivity_checked = transport->smart_options
|
|
? transport->smart_options->connectivity_checked : 0;
|
|
int ret;
|
|
trace2_region_enter("fetch", "consume_refs", the_repository);
|
|
ret = store_updated_refs(transport->url,
|
|
transport->remote->name,
|
|
connectivity_checked,
|
|
ref_map);
|
|
transport_unlock_pack(transport);
|
|
trace2_region_leave("fetch", "consume_refs", the_repository);
|
|
return ret;
|
|
}
|
|
|
|
static int prune_refs(struct refspec *rs, struct ref *ref_map,
|
|
const char *raw_url)
|
|
{
|
|
int url_len, i, result = 0;
|
|
struct ref *ref, *stale_refs = get_stale_heads(rs, ref_map);
|
|
char *url;
|
|
int summary_width = transport_summary_width(stale_refs);
|
|
const char *dangling_msg = dry_run
|
|
? _(" (%s will become dangling)")
|
|
: _(" (%s has become dangling)");
|
|
|
|
if (raw_url)
|
|
url = transport_anonymize_url(raw_url);
|
|
else
|
|
url = xstrdup("foreign");
|
|
|
|
url_len = strlen(url);
|
|
for (i = url_len - 1; url[i] == '/' && 0 <= i; i--)
|
|
;
|
|
|
|
url_len = i + 1;
|
|
if (4 < i && !strncmp(".git", url + i - 3, 4))
|
|
url_len = i - 3;
|
|
|
|
if (!dry_run) {
|
|
struct string_list refnames = STRING_LIST_INIT_NODUP;
|
|
|
|
for (ref = stale_refs; ref; ref = ref->next)
|
|
string_list_append(&refnames, ref->name);
|
|
|
|
result = delete_refs("fetch: prune", &refnames, 0);
|
|
string_list_clear(&refnames, 0);
|
|
}
|
|
|
|
if (verbosity >= 0) {
|
|
for (ref = stale_refs; ref; ref = ref->next) {
|
|
struct strbuf sb = STRBUF_INIT;
|
|
if (!shown_url) {
|
|
fprintf(stderr, _("From %.*s\n"), url_len, url);
|
|
shown_url = 1;
|
|
}
|
|
format_display(&sb, '-', _("[deleted]"), NULL,
|
|
_("(none)"), prettify_refname(ref->name),
|
|
summary_width);
|
|
fprintf(stderr, " %s\n",sb.buf);
|
|
strbuf_release(&sb);
|
|
warn_dangling_symref(stderr, dangling_msg, ref->name);
|
|
}
|
|
}
|
|
|
|
free(url);
|
|
free_refs(stale_refs);
|
|
return result;
|
|
}
|
|
|
|
static void check_not_current_branch(struct ref *ref_map)
|
|
{
|
|
struct branch *current_branch = branch_get(NULL);
|
|
|
|
if (is_bare_repository() || !current_branch)
|
|
return;
|
|
|
|
for (; ref_map; ref_map = ref_map->next)
|
|
if (ref_map->peer_ref && !strcmp(current_branch->refname,
|
|
ref_map->peer_ref->name))
|
|
die(_("Refusing to fetch into current branch %s "
|
|
"of non-bare repository"), current_branch->refname);
|
|
}
|
|
|
|
static int truncate_fetch_head(void)
|
|
{
|
|
const char *filename = git_path_fetch_head(the_repository);
|
|
FILE *fp = fopen_for_writing(filename);
|
|
|
|
if (!fp)
|
|
return error_errno(_("cannot open %s"), filename);
|
|
fclose(fp);
|
|
return 0;
|
|
}
|
|
|
|
static void set_option(struct transport *transport, const char *name, const char *value)
|
|
{
|
|
int r = transport_set_option(transport, name, value);
|
|
if (r < 0)
|
|
die(_("Option \"%s\" value \"%s\" is not valid for %s"),
|
|
name, value, transport->url);
|
|
if (r > 0)
|
|
warning(_("Option \"%s\" is ignored for %s\n"),
|
|
name, transport->url);
|
|
}
|
|
|
|
|
|
static int add_oid(const char *refname, const struct object_id *oid, int flags,
|
|
void *cb_data)
|
|
{
|
|
struct oid_array *oids = cb_data;
|
|
|
|
oid_array_append(oids, oid);
|
|
return 0;
|
|
}
|
|
|
|
static void add_negotiation_tips(struct git_transport_options *smart_options)
|
|
{
|
|
struct oid_array *oids = xcalloc(1, sizeof(*oids));
|
|
int i;
|
|
|
|
for (i = 0; i < negotiation_tip.nr; i++) {
|
|
const char *s = negotiation_tip.items[i].string;
|
|
int old_nr;
|
|
if (!has_glob_specials(s)) {
|
|
struct object_id oid;
|
|
if (get_oid(s, &oid))
|
|
die(_("%s is not a valid object"), s);
|
|
if (!has_object(the_repository, &oid, 0))
|
|
die(_("the object %s does not exist"), s);
|
|
oid_array_append(oids, &oid);
|
|
continue;
|
|
}
|
|
old_nr = oids->nr;
|
|
for_each_glob_ref(add_oid, s, oids);
|
|
if (old_nr == oids->nr)
|
|
warning("Ignoring --negotiation-tip=%s because it does not match any refs",
|
|
s);
|
|
}
|
|
smart_options->negotiation_tips = oids;
|
|
}
|
|
|
|
static struct transport *prepare_transport(struct remote *remote, int deepen)
|
|
{
|
|
struct transport *transport;
|
|
|
|
transport = transport_get(remote, NULL);
|
|
transport_set_verbosity(transport, verbosity, progress);
|
|
transport->family = family;
|
|
if (upload_pack)
|
|
set_option(transport, TRANS_OPT_UPLOADPACK, upload_pack);
|
|
if (keep)
|
|
set_option(transport, TRANS_OPT_KEEP, "yes");
|
|
if (depth)
|
|
set_option(transport, TRANS_OPT_DEPTH, depth);
|
|
if (deepen && deepen_since)
|
|
set_option(transport, TRANS_OPT_DEEPEN_SINCE, deepen_since);
|
|
if (deepen && deepen_not.nr)
|
|
set_option(transport, TRANS_OPT_DEEPEN_NOT,
|
|
(const char *)&deepen_not);
|
|
if (deepen_relative)
|
|
set_option(transport, TRANS_OPT_DEEPEN_RELATIVE, "yes");
|
|
if (update_shallow)
|
|
set_option(transport, TRANS_OPT_UPDATE_SHALLOW, "yes");
|
|
if (filter_options.choice) {
|
|
const char *spec =
|
|
expand_list_objects_filter_spec(&filter_options);
|
|
set_option(transport, TRANS_OPT_LIST_OBJECTS_FILTER, spec);
|
|
set_option(transport, TRANS_OPT_FROM_PROMISOR, "1");
|
|
}
|
|
if (negotiation_tip.nr) {
|
|
if (transport->smart_options)
|
|
add_negotiation_tips(transport->smart_options);
|
|
else
|
|
warning("Ignoring --negotiation-tip because the protocol does not support it.");
|
|
}
|
|
return transport;
|
|
}
|
|
|
|
static void backfill_tags(struct transport *transport, struct ref *ref_map)
|
|
{
|
|
int cannot_reuse;
|
|
|
|
/*
|
|
* Once we have set TRANS_OPT_DEEPEN_SINCE, we can't unset it
|
|
* when remote helper is used (setting it to an empty string
|
|
* is not unsetting). We could extend the remote helper
|
|
* protocol for that, but for now, just force a new connection
|
|
* without deepen-since. Similar story for deepen-not.
|
|
*/
|
|
cannot_reuse = transport->cannot_reuse ||
|
|
deepen_since || deepen_not.nr;
|
|
if (cannot_reuse) {
|
|
gsecondary = prepare_transport(transport->remote, 0);
|
|
transport = gsecondary;
|
|
}
|
|
|
|
transport_set_option(transport, TRANS_OPT_FOLLOWTAGS, NULL);
|
|
transport_set_option(transport, TRANS_OPT_DEPTH, "0");
|
|
transport_set_option(transport, TRANS_OPT_DEEPEN_RELATIVE, NULL);
|
|
if (!fetch_refs(transport, ref_map))
|
|
consume_refs(transport, ref_map);
|
|
|
|
if (gsecondary) {
|
|
transport_disconnect(gsecondary);
|
|
gsecondary = NULL;
|
|
}
|
|
}
|
|
|
|
static int do_fetch(struct transport *transport,
|
|
struct refspec *rs)
|
|
{
|
|
struct ref *ref_map;
|
|
int autotags = (transport->remote->fetch_tags == 1);
|
|
int retcode = 0;
|
|
const struct ref *remote_refs;
|
|
struct transport_ls_refs_options transport_ls_refs_options =
|
|
TRANSPORT_LS_REFS_OPTIONS_INIT;
|
|
int must_list_refs = 1;
|
|
|
|
if (tags == TAGS_DEFAULT) {
|
|
if (transport->remote->fetch_tags == 2)
|
|
tags = TAGS_SET;
|
|
if (transport->remote->fetch_tags == -1)
|
|
tags = TAGS_UNSET;
|
|
}
|
|
|
|
/* if not appending, truncate FETCH_HEAD */
|
|
if (!append && write_fetch_head) {
|
|
retcode = truncate_fetch_head();
|
|
if (retcode)
|
|
goto cleanup;
|
|
}
|
|
|
|
if (rs->nr) {
|
|
int i;
|
|
|
|
refspec_ref_prefixes(rs, &transport_ls_refs_options.ref_prefixes);
|
|
|
|
/*
|
|
* We can avoid listing refs if all of them are exact
|
|
* OIDs
|
|
*/
|
|
must_list_refs = 0;
|
|
for (i = 0; i < rs->nr; i++) {
|
|
if (!rs->items[i].exact_sha1) {
|
|
must_list_refs = 1;
|
|
break;
|
|
}
|
|
}
|
|
} else if (transport->remote && transport->remote->fetch.nr)
|
|
refspec_ref_prefixes(&transport->remote->fetch,
|
|
&transport_ls_refs_options.ref_prefixes);
|
|
|
|
if (tags == TAGS_SET || tags == TAGS_DEFAULT) {
|
|
must_list_refs = 1;
|
|
if (transport_ls_refs_options.ref_prefixes.nr)
|
|
strvec_push(&transport_ls_refs_options.ref_prefixes,
|
|
"refs/tags/");
|
|
}
|
|
|
|
if (must_list_refs) {
|
|
trace2_region_enter("fetch", "remote_refs", the_repository);
|
|
remote_refs = transport_get_remote_refs(transport,
|
|
&transport_ls_refs_options);
|
|
trace2_region_leave("fetch", "remote_refs", the_repository);
|
|
} else
|
|
remote_refs = NULL;
|
|
|
|
strvec_clear(&transport_ls_refs_options.ref_prefixes);
|
|
|
|
ref_map = get_ref_map(transport->remote, remote_refs, rs,
|
|
tags, &autotags);
|
|
if (!update_head_ok)
|
|
check_not_current_branch(ref_map);
|
|
|
|
if (tags == TAGS_DEFAULT && autotags)
|
|
transport_set_option(transport, TRANS_OPT_FOLLOWTAGS, "1");
|
|
if (prune) {
|
|
/*
|
|
* We only prune based on refspecs specified
|
|
* explicitly (via command line or configuration); we
|
|
* don't care whether --tags was specified.
|
|
*/
|
|
if (rs->nr) {
|
|
prune_refs(rs, ref_map, transport->url);
|
|
} else {
|
|
prune_refs(&transport->remote->fetch,
|
|
ref_map,
|
|
transport->url);
|
|
}
|
|
}
|
|
if (fetch_refs(transport, ref_map) || consume_refs(transport, ref_map)) {
|
|
free_refs(ref_map);
|
|
retcode = 1;
|
|
goto cleanup;
|
|
}
|
|
|
|
if (set_upstream) {
|
|
struct branch *branch = branch_get("HEAD");
|
|
struct ref *rm;
|
|
struct ref *source_ref = NULL;
|
|
|
|
/*
|
|
* We're setting the upstream configuration for the
|
|
* current branch. The relevant upstream is the
|
|
* fetched branch that is meant to be merged with the
|
|
* current one, i.e. the one fetched to FETCH_HEAD.
|
|
*
|
|
* When there are several such branches, consider the
|
|
* request ambiguous and err on the safe side by doing
|
|
* nothing and just emit a warning.
|
|
*/
|
|
for (rm = ref_map; rm; rm = rm->next) {
|
|
if (!rm->peer_ref) {
|
|
if (source_ref) {
|
|
warning(_("multiple branches detected, incompatible with --set-upstream"));
|
|
goto skip;
|
|
} else {
|
|
source_ref = rm;
|
|
}
|
|
}
|
|
}
|
|
if (source_ref) {
|
|
if (!strcmp(source_ref->name, "HEAD") ||
|
|
starts_with(source_ref->name, "refs/heads/"))
|
|
install_branch_config(0,
|
|
branch->name,
|
|
transport->remote->name,
|
|
source_ref->name);
|
|
else if (starts_with(source_ref->name, "refs/remotes/"))
|
|
warning(_("not setting upstream for a remote remote-tracking branch"));
|
|
else if (starts_with(source_ref->name, "refs/tags/"))
|
|
warning(_("not setting upstream for a remote tag"));
|
|
else
|
|
warning(_("unknown branch type"));
|
|
} else {
|
|
warning(_("no source branch found.\n"
|
|
"you need to specify exactly one branch with the --set-upstream option."));
|
|
}
|
|
}
|
|
skip:
|
|
free_refs(ref_map);
|
|
|
|
/* if neither --no-tags nor --tags was specified, do automated tag
|
|
* following ... */
|
|
if (tags == TAGS_DEFAULT && autotags) {
|
|
struct ref **tail = &ref_map;
|
|
ref_map = NULL;
|
|
find_non_local_tags(remote_refs, &ref_map, &tail);
|
|
if (ref_map)
|
|
backfill_tags(transport, ref_map);
|
|
free_refs(ref_map);
|
|
}
|
|
|
|
cleanup:
|
|
return retcode;
|
|
}
|
|
|
|
static int get_one_remote_for_fetch(struct remote *remote, void *priv)
|
|
{
|
|
struct string_list *list = priv;
|
|
if (!remote->skip_default_update)
|
|
string_list_append(list, remote->name);
|
|
return 0;
|
|
}
|
|
|
|
struct remote_group_data {
|
|
const char *name;
|
|
struct string_list *list;
|
|
};
|
|
|
|
static int get_remote_group(const char *key, const char *value, void *priv)
|
|
{
|
|
struct remote_group_data *g = priv;
|
|
|
|
if (skip_prefix(key, "remotes.", &key) && !strcmp(key, g->name)) {
|
|
/* split list by white space */
|
|
while (*value) {
|
|
size_t wordlen = strcspn(value, " \t\n");
|
|
|
|
if (wordlen >= 1)
|
|
string_list_append_nodup(g->list,
|
|
xstrndup(value, wordlen));
|
|
value += wordlen + (value[wordlen] != '\0');
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int add_remote_or_group(const char *name, struct string_list *list)
|
|
{
|
|
int prev_nr = list->nr;
|
|
struct remote_group_data g;
|
|
g.name = name; g.list = list;
|
|
|
|
git_config(get_remote_group, &g);
|
|
if (list->nr == prev_nr) {
|
|
struct remote *remote = remote_get(name);
|
|
if (!remote_is_configured(remote, 0))
|
|
return 0;
|
|
string_list_append(list, remote->name);
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
static void add_options_to_argv(struct strvec *argv)
|
|
{
|
|
if (dry_run)
|
|
strvec_push(argv, "--dry-run");
|
|
if (prune != -1)
|
|
strvec_push(argv, prune ? "--prune" : "--no-prune");
|
|
if (prune_tags != -1)
|
|
strvec_push(argv, prune_tags ? "--prune-tags" : "--no-prune-tags");
|
|
if (update_head_ok)
|
|
strvec_push(argv, "--update-head-ok");
|
|
if (force)
|
|
strvec_push(argv, "--force");
|
|
if (keep)
|
|
strvec_push(argv, "--keep");
|
|
if (recurse_submodules == RECURSE_SUBMODULES_ON)
|
|
strvec_push(argv, "--recurse-submodules");
|
|
else if (recurse_submodules == RECURSE_SUBMODULES_ON_DEMAND)
|
|
strvec_push(argv, "--recurse-submodules=on-demand");
|
|
if (tags == TAGS_SET)
|
|
strvec_push(argv, "--tags");
|
|
else if (tags == TAGS_UNSET)
|
|
strvec_push(argv, "--no-tags");
|
|
if (verbosity >= 2)
|
|
strvec_push(argv, "-v");
|
|
if (verbosity >= 1)
|
|
strvec_push(argv, "-v");
|
|
else if (verbosity < 0)
|
|
strvec_push(argv, "-q");
|
|
if (family == TRANSPORT_FAMILY_IPV4)
|
|
strvec_push(argv, "--ipv4");
|
|
else if (family == TRANSPORT_FAMILY_IPV6)
|
|
strvec_push(argv, "--ipv6");
|
|
}
|
|
|
|
/* Fetch multiple remotes in parallel */
|
|
|
|
struct parallel_fetch_state {
|
|
const char **argv;
|
|
struct string_list *remotes;
|
|
int next, result;
|
|
};
|
|
|
|
static int fetch_next_remote(struct child_process *cp, struct strbuf *out,
|
|
void *cb, void **task_cb)
|
|
{
|
|
struct parallel_fetch_state *state = cb;
|
|
char *remote;
|
|
|
|
if (state->next < 0 || state->next >= state->remotes->nr)
|
|
return 0;
|
|
|
|
remote = state->remotes->items[state->next++].string;
|
|
*task_cb = remote;
|
|
|
|
strvec_pushv(&cp->args, state->argv);
|
|
strvec_push(&cp->args, remote);
|
|
cp->git_cmd = 1;
|
|
|
|
if (verbosity >= 0)
|
|
printf(_("Fetching %s\n"), remote);
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int fetch_failed_to_start(struct strbuf *out, void *cb, void *task_cb)
|
|
{
|
|
struct parallel_fetch_state *state = cb;
|
|
const char *remote = task_cb;
|
|
|
|
state->result = error(_("Could not fetch %s"), remote);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fetch_finished(int result, struct strbuf *out,
|
|
void *cb, void *task_cb)
|
|
{
|
|
struct parallel_fetch_state *state = cb;
|
|
const char *remote = task_cb;
|
|
|
|
if (result) {
|
|
strbuf_addf(out, _("could not fetch '%s' (exit code: %d)\n"),
|
|
remote, result);
|
|
state->result = -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int fetch_multiple(struct string_list *list, int max_children)
|
|
{
|
|
int i, result = 0;
|
|
struct strvec argv = STRVEC_INIT;
|
|
|
|
if (!append && write_fetch_head) {
|
|
int errcode = truncate_fetch_head();
|
|
if (errcode)
|
|
return errcode;
|
|
}
|
|
|
|
strvec_pushl(&argv, "fetch", "--append", "--no-auto-gc",
|
|
"--no-write-commit-graph", NULL);
|
|
add_options_to_argv(&argv);
|
|
|
|
if (max_children != 1 && list->nr != 1) {
|
|
struct parallel_fetch_state state = { argv.v, list, 0, 0 };
|
|
|
|
strvec_push(&argv, "--end-of-options");
|
|
result = run_processes_parallel_tr2(max_children,
|
|
&fetch_next_remote,
|
|
&fetch_failed_to_start,
|
|
&fetch_finished,
|
|
&state,
|
|
"fetch", "parallel/fetch");
|
|
|
|
if (!result)
|
|
result = state.result;
|
|
} else
|
|
for (i = 0; i < list->nr; i++) {
|
|
const char *name = list->items[i].string;
|
|
strvec_push(&argv, name);
|
|
if (verbosity >= 0)
|
|
printf(_("Fetching %s\n"), name);
|
|
if (run_command_v_opt(argv.v, RUN_GIT_CMD)) {
|
|
error(_("Could not fetch %s"), name);
|
|
result = 1;
|
|
}
|
|
strvec_pop(&argv);
|
|
}
|
|
|
|
strvec_clear(&argv);
|
|
return !!result;
|
|
}
|
|
|
|
/*
|
|
* Fetching from the promisor remote should use the given filter-spec
|
|
* or inherit the default filter-spec from the config.
|
|
*/
|
|
static inline void fetch_one_setup_partial(struct remote *remote)
|
|
{
|
|
/*
|
|
* Explicit --no-filter argument overrides everything, regardless
|
|
* of any prior partial clones and fetches.
|
|
*/
|
|
if (filter_options.no_filter)
|
|
return;
|
|
|
|
/*
|
|
* If no prior partial clone/fetch and the current fetch DID NOT
|
|
* request a partial-fetch, do a normal fetch.
|
|
*/
|
|
if (!has_promisor_remote() && !filter_options.choice)
|
|
return;
|
|
|
|
/*
|
|
* If this is a partial-fetch request, we enable partial on
|
|
* this repo if not already enabled and remember the given
|
|
* filter-spec as the default for subsequent fetches to this
|
|
* remote if there is currently no default filter-spec.
|
|
*/
|
|
if (filter_options.choice) {
|
|
partial_clone_register(remote->name, &filter_options);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Do a partial-fetch from the promisor remote using either the
|
|
* explicitly given filter-spec or inherit the filter-spec from
|
|
* the config.
|
|
*/
|
|
if (!filter_options.choice)
|
|
partial_clone_get_default_filter_spec(&filter_options, remote->name);
|
|
return;
|
|
}
|
|
|
|
static int fetch_one(struct remote *remote, int argc, const char **argv,
|
|
int prune_tags_ok, int use_stdin_refspecs)
|
|
{
|
|
struct refspec rs = REFSPEC_INIT_FETCH;
|
|
int i;
|
|
int exit_code;
|
|
int maybe_prune_tags;
|
|
int remote_via_config = remote_is_configured(remote, 0);
|
|
|
|
if (!remote)
|
|
die(_("No remote repository specified. Please, specify either a URL or a\n"
|
|
"remote name from which new revisions should be fetched."));
|
|
|
|
gtransport = prepare_transport(remote, 1);
|
|
|
|
if (prune < 0) {
|
|
/* no command line request */
|
|
if (0 <= remote->prune)
|
|
prune = remote->prune;
|
|
else if (0 <= fetch_prune_config)
|
|
prune = fetch_prune_config;
|
|
else
|
|
prune = PRUNE_BY_DEFAULT;
|
|
}
|
|
|
|
if (prune_tags < 0) {
|
|
/* no command line request */
|
|
if (0 <= remote->prune_tags)
|
|
prune_tags = remote->prune_tags;
|
|
else if (0 <= fetch_prune_tags_config)
|
|
prune_tags = fetch_prune_tags_config;
|
|
else
|
|
prune_tags = PRUNE_TAGS_BY_DEFAULT;
|
|
}
|
|
|
|
maybe_prune_tags = prune_tags_ok && prune_tags;
|
|
if (maybe_prune_tags && remote_via_config)
|
|
refspec_append(&remote->fetch, TAG_REFSPEC);
|
|
|
|
if (maybe_prune_tags && (argc || !remote_via_config))
|
|
refspec_append(&rs, TAG_REFSPEC);
|
|
|
|
for (i = 0; i < argc; i++) {
|
|
if (!strcmp(argv[i], "tag")) {
|
|
i++;
|
|
if (i >= argc)
|
|
die(_("You need to specify a tag name."));
|
|
|
|
refspec_appendf(&rs, "refs/tags/%s:refs/tags/%s",
|
|
argv[i], argv[i]);
|
|
} else {
|
|
refspec_append(&rs, argv[i]);
|
|
}
|
|
}
|
|
|
|
if (use_stdin_refspecs) {
|
|
struct strbuf line = STRBUF_INIT;
|
|
while (strbuf_getline_lf(&line, stdin) != EOF)
|
|
refspec_append(&rs, line.buf);
|
|
strbuf_release(&line);
|
|
}
|
|
|
|
if (server_options.nr)
|
|
gtransport->server_options = &server_options;
|
|
|
|
sigchain_push_common(unlock_pack_on_signal);
|
|
atexit(unlock_pack);
|
|
sigchain_push(SIGPIPE, SIG_IGN);
|
|
exit_code = do_fetch(gtransport, &rs);
|
|
sigchain_pop(SIGPIPE);
|
|
refspec_clear(&rs);
|
|
transport_disconnect(gtransport);
|
|
gtransport = NULL;
|
|
return exit_code;
|
|
}
|
|
|
|
int cmd_fetch(int argc, const char **argv, const char *prefix)
|
|
{
|
|
int i;
|
|
struct string_list list = STRING_LIST_INIT_DUP;
|
|
struct remote *remote = NULL;
|
|
int result = 0;
|
|
int prune_tags_ok = 1;
|
|
|
|
packet_trace_identity("fetch");
|
|
|
|
/* Record the command line for the reflog */
|
|
strbuf_addstr(&default_rla, "fetch");
|
|
for (i = 1; i < argc; i++) {
|
|
/* This handles non-URLs gracefully */
|
|
char *anon = transport_anonymize_url(argv[i]);
|
|
|
|
strbuf_addf(&default_rla, " %s", anon);
|
|
free(anon);
|
|
}
|
|
|
|
git_config(git_fetch_config, NULL);
|
|
|
|
argc = parse_options(argc, argv, prefix,
|
|
builtin_fetch_options, builtin_fetch_usage, 0);
|
|
if (recurse_submodules != RECURSE_SUBMODULES_OFF) {
|
|
int *sfjc = submodule_fetch_jobs_config == -1
|
|
? &submodule_fetch_jobs_config : NULL;
|
|
int *rs = recurse_submodules == RECURSE_SUBMODULES_DEFAULT
|
|
? &recurse_submodules : NULL;
|
|
|
|
fetch_config_from_gitmodules(sfjc, rs);
|
|
}
|
|
|
|
if (deepen_relative) {
|
|
if (deepen_relative < 0)
|
|
die(_("Negative depth in --deepen is not supported"));
|
|
if (depth)
|
|
die(_("--deepen and --depth are mutually exclusive"));
|
|
depth = xstrfmt("%d", deepen_relative);
|
|
}
|
|
if (unshallow) {
|
|
if (depth)
|
|
die(_("--depth and --unshallow cannot be used together"));
|
|
else if (!is_repository_shallow(the_repository))
|
|
die(_("--unshallow on a complete repository does not make sense"));
|
|
else
|
|
depth = xstrfmt("%d", INFINITE_DEPTH);
|
|
}
|
|
|
|
/* no need to be strict, transport_set_option() will validate it again */
|
|
if (depth && atoi(depth) < 1)
|
|
die(_("depth %s is not a positive number"), depth);
|
|
if (depth || deepen_since || deepen_not.nr)
|
|
deepen = 1;
|
|
|
|
/* FETCH_HEAD never gets updated in --dry-run mode */
|
|
if (dry_run)
|
|
write_fetch_head = 0;
|
|
|
|
if (all) {
|
|
if (argc == 1)
|
|
die(_("fetch --all does not take a repository argument"));
|
|
else if (argc > 1)
|
|
die(_("fetch --all does not make sense with refspecs"));
|
|
(void) for_each_remote(get_one_remote_for_fetch, &list);
|
|
} else if (argc == 0) {
|
|
/* No arguments -- use default remote */
|
|
remote = remote_get(NULL);
|
|
} else if (multiple) {
|
|
/* All arguments are assumed to be remotes or groups */
|
|
for (i = 0; i < argc; i++)
|
|
if (!add_remote_or_group(argv[i], &list))
|
|
die(_("No such remote or remote group: %s"), argv[i]);
|
|
} else {
|
|
/* Single remote or group */
|
|
(void) add_remote_or_group(argv[0], &list);
|
|
if (list.nr > 1) {
|
|
/* More than one remote */
|
|
if (argc > 1)
|
|
die(_("Fetching a group and specifying refspecs does not make sense"));
|
|
} else {
|
|
/* Zero or one remotes */
|
|
remote = remote_get(argv[0]);
|
|
prune_tags_ok = (argc == 1);
|
|
argc--;
|
|
argv++;
|
|
}
|
|
}
|
|
|
|
if (negotiate_only) {
|
|
struct oidset acked_commits = OIDSET_INIT;
|
|
struct oidset_iter iter;
|
|
const struct object_id *oid;
|
|
|
|
if (!remote)
|
|
die(_("must supply remote when using --negotiate-only"));
|
|
gtransport = prepare_transport(remote, 1);
|
|
if (gtransport->smart_options) {
|
|
gtransport->smart_options->acked_commits = &acked_commits;
|
|
} else {
|
|
warning(_("Protocol does not support --negotiate-only, exiting."));
|
|
return 1;
|
|
}
|
|
if (server_options.nr)
|
|
gtransport->server_options = &server_options;
|
|
result = transport_fetch_refs(gtransport, NULL);
|
|
|
|
oidset_iter_init(&acked_commits, &iter);
|
|
while ((oid = oidset_iter_next(&iter)))
|
|
printf("%s\n", oid_to_hex(oid));
|
|
oidset_clear(&acked_commits);
|
|
} else if (remote) {
|
|
if (filter_options.choice || has_promisor_remote())
|
|
fetch_one_setup_partial(remote);
|
|
result = fetch_one(remote, argc, argv, prune_tags_ok, stdin_refspecs);
|
|
} else {
|
|
int max_children = max_jobs;
|
|
|
|
if (filter_options.choice)
|
|
die(_("--filter can only be used with the remote "
|
|
"configured in extensions.partialclone"));
|
|
|
|
if (atomic_fetch)
|
|
die(_("--atomic can only be used when fetching "
|
|
"from one remote"));
|
|
|
|
if (stdin_refspecs)
|
|
die(_("--stdin can only be used when fetching "
|
|
"from one remote"));
|
|
|
|
if (max_children < 0)
|
|
max_children = fetch_parallel_config;
|
|
|
|
/* TODO should this also die if we have a previous partial-clone? */
|
|
result = fetch_multiple(&list, max_children);
|
|
}
|
|
|
|
if (!result && (recurse_submodules != RECURSE_SUBMODULES_OFF)) {
|
|
struct strvec options = STRVEC_INIT;
|
|
int max_children = max_jobs;
|
|
|
|
if (max_children < 0)
|
|
max_children = submodule_fetch_jobs_config;
|
|
if (max_children < 0)
|
|
max_children = fetch_parallel_config;
|
|
|
|
add_options_to_argv(&options);
|
|
result = fetch_populated_submodules(the_repository,
|
|
&options,
|
|
submodule_prefix,
|
|
recurse_submodules,
|
|
recurse_submodules_default,
|
|
verbosity < 0,
|
|
max_children);
|
|
strvec_clear(&options);
|
|
}
|
|
|
|
string_list_clear(&list, 0);
|
|
|
|
prepare_repo_settings(the_repository);
|
|
if (fetch_write_commit_graph > 0 ||
|
|
(fetch_write_commit_graph < 0 &&
|
|
the_repository->settings.fetch_write_commit_graph)) {
|
|
int commit_graph_flags = COMMIT_GRAPH_WRITE_SPLIT;
|
|
|
|
if (progress)
|
|
commit_graph_flags |= COMMIT_GRAPH_WRITE_PROGRESS;
|
|
|
|
write_commit_graph_reachable(the_repository->objects->odb,
|
|
commit_graph_flags,
|
|
NULL);
|
|
}
|
|
|
|
close_object_store(the_repository->objects);
|
|
|
|
if (enable_auto_gc)
|
|
run_auto_maintenance(verbosity < 0);
|
|
|
|
return result;
|
|
}
|