2019-06-25 16:40:27 +03:00
|
|
|
#include "cache.h"
|
2019-06-25 16:40:28 +03:00
|
|
|
#include "object-store.h"
|
2019-06-25 16:40:27 +03:00
|
|
|
#include "promisor-remote.h"
|
|
|
|
#include "config.h"
|
2019-06-25 16:40:37 +03:00
|
|
|
#include "transport.h"
|
2020-08-18 07:01:36 +03:00
|
|
|
#include "strvec.h"
|
promisor-remote: die upon failing fetch
In a partial clone, an attempt to read a missing object results in an
attempt to fetch that single object. In order to avoid multiple
sequential fetches, which would occur when multiple objects are missing
(which is the typical case), some commands have been taught to prefetch
in a batch: such a command would, in a partial clone, notice that
several objects that it will eventually need are missing, and call
promisor_remote_get_direct() with all such objects at once.
When this batch prefetch fails, these commands fall back to the
sequential fetches. But at $DAYJOB we have noticed that this results in
a bad user experience: a command would take unexpectedly long to finish
(and possibly use up a lot of bandwidth) if the batch prefetch would
fail for some intermittent reason, but all subsequent fetches would
work. It would be a better user experience for such a command would
just fail.
Therefore, make it a fatal error if the prefetch fails and at least one
object being fetched is known to be a promisor object. (The latter
criterion is to make sure that we are not misleading the user that such
an object would be present from the promisor remote. For example, a
missing object may be a result of repository corruption and not because
it is expectedly missing due to the repository being a partial clone.)
Signed-off-by: Jonathan Tan <jonathantanmy@google.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2022-10-05 00:13:41 +03:00
|
|
|
#include "packfile.h"
|
2019-06-25 16:40:37 +03:00
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
struct promisor_remote_config {
|
|
|
|
struct promisor_remote *promisors;
|
|
|
|
struct promisor_remote **promisors_tail;
|
|
|
|
};
|
2019-06-25 16:40:38 +03:00
|
|
|
|
2021-06-16 01:41:42 +03:00
|
|
|
static int fetch_objects(struct repository *repo,
|
|
|
|
const char *remote_name,
|
2019-06-25 16:40:37 +03:00
|
|
|
const struct object_id *oids,
|
|
|
|
int oid_nr)
|
|
|
|
{
|
2020-08-18 07:01:36 +03:00
|
|
|
struct child_process child = CHILD_PROCESS_INIT;
|
2019-06-25 16:40:37 +03:00
|
|
|
int i;
|
2020-08-18 07:01:36 +03:00
|
|
|
FILE *child_in;
|
|
|
|
|
|
|
|
child.git_cmd = 1;
|
|
|
|
child.in = -1;
|
2021-06-17 20:13:26 +03:00
|
|
|
if (repo != the_repository)
|
2022-06-02 12:09:50 +03:00
|
|
|
prepare_other_repo_env(&child.env, repo->gitdir);
|
2020-08-18 07:01:36 +03:00
|
|
|
strvec_pushl(&child.args, "-c", "fetch.negotiationAlgorithm=noop",
|
|
|
|
"fetch", remote_name, "--no-tags",
|
|
|
|
"--no-write-fetch-head", "--recurse-submodules=no",
|
|
|
|
"--filter=blob:none", "--stdin", NULL);
|
|
|
|
if (start_command(&child))
|
|
|
|
die(_("promisor-remote: unable to fork off fetch subprocess"));
|
|
|
|
child_in = xfdopen(child.in, "w");
|
2019-06-25 16:40:37 +03:00
|
|
|
|
2021-06-16 01:41:42 +03:00
|
|
|
trace2_data_intmax("promisor", repo, "fetch_count", oid_nr);
|
|
|
|
|
2019-06-25 16:40:37 +03:00
|
|
|
for (i = 0; i < oid_nr; i++) {
|
2020-08-18 07:01:36 +03:00
|
|
|
if (fputs(oid_to_hex(&oids[i]), child_in) < 0)
|
|
|
|
die_errno(_("promisor-remote: could not write to fetch subprocess"));
|
|
|
|
if (fputc('\n', child_in) < 0)
|
|
|
|
die_errno(_("promisor-remote: could not write to fetch subprocess"));
|
2019-06-25 16:40:37 +03:00
|
|
|
}
|
2020-08-18 07:01:36 +03:00
|
|
|
|
|
|
|
if (fclose(child_in) < 0)
|
|
|
|
die_errno(_("promisor-remote: could not close stdin to fetch subprocess"));
|
|
|
|
return finish_command(&child) ? -1 : 0;
|
2019-06-25 16:40:37 +03:00
|
|
|
}
|
2019-06-25 16:40:27 +03:00
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
static struct promisor_remote *promisor_remote_new(struct promisor_remote_config *config,
|
|
|
|
const char *remote_name)
|
2019-06-25 16:40:27 +03:00
|
|
|
{
|
|
|
|
struct promisor_remote *r;
|
|
|
|
|
|
|
|
if (*remote_name == '/') {
|
|
|
|
warning(_("promisor remote name cannot begin with '/': %s"),
|
|
|
|
remote_name);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
FLEX_ALLOC_STR(r, name, remote_name);
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
*config->promisors_tail = r;
|
|
|
|
config->promisors_tail = &r->next;
|
2019-06-25 16:40:27 +03:00
|
|
|
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
static struct promisor_remote *promisor_remote_lookup(struct promisor_remote_config *config,
|
|
|
|
const char *remote_name,
|
2019-06-25 16:40:27 +03:00
|
|
|
struct promisor_remote **previous)
|
|
|
|
{
|
|
|
|
struct promisor_remote *r, *p;
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
for (p = NULL, r = config->promisors; r; p = r, r = r->next)
|
2019-06-25 16:40:27 +03:00
|
|
|
if (!strcmp(r->name, remote_name)) {
|
|
|
|
if (previous)
|
|
|
|
*previous = p;
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
static void promisor_remote_move_to_tail(struct promisor_remote_config *config,
|
|
|
|
struct promisor_remote *r,
|
2019-06-25 16:40:30 +03:00
|
|
|
struct promisor_remote *previous)
|
|
|
|
{
|
2022-05-02 19:50:37 +03:00
|
|
|
if (!r->next)
|
2019-10-01 01:03:55 +03:00
|
|
|
return;
|
|
|
|
|
2019-06-25 16:40:30 +03:00
|
|
|
if (previous)
|
|
|
|
previous->next = r->next;
|
|
|
|
else
|
2021-06-17 20:13:23 +03:00
|
|
|
config->promisors = r->next ? r->next : r;
|
2019-06-25 16:40:30 +03:00
|
|
|
r->next = NULL;
|
2021-06-17 20:13:23 +03:00
|
|
|
*config->promisors_tail = r;
|
|
|
|
config->promisors_tail = &r->next;
|
2019-06-25 16:40:30 +03:00
|
|
|
}
|
|
|
|
|
2019-06-25 16:40:27 +03:00
|
|
|
static int promisor_remote_config(const char *var, const char *value, void *data)
|
|
|
|
{
|
2021-06-17 20:13:23 +03:00
|
|
|
struct promisor_remote_config *config = data;
|
2019-06-25 16:40:27 +03:00
|
|
|
const char *name;
|
2020-04-10 22:44:28 +03:00
|
|
|
size_t namelen;
|
2019-06-25 16:40:27 +03:00
|
|
|
const char *subkey;
|
|
|
|
|
|
|
|
if (parse_config_key(var, "remote", &name, &namelen, &subkey) < 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!strcmp(subkey, "promisor")) {
|
|
|
|
char *remote_name;
|
|
|
|
|
|
|
|
if (!git_config_bool(var, value))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
remote_name = xmemdupz(name, namelen);
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
if (!promisor_remote_lookup(config, remote_name, NULL))
|
|
|
|
promisor_remote_new(config, remote_name);
|
2019-06-25 16:40:27 +03:00
|
|
|
|
|
|
|
free(remote_name);
|
|
|
|
return 0;
|
|
|
|
}
|
2019-06-25 16:40:32 +03:00
|
|
|
if (!strcmp(subkey, "partialclonefilter")) {
|
|
|
|
struct promisor_remote *r;
|
|
|
|
char *remote_name = xmemdupz(name, namelen);
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
r = promisor_remote_lookup(config, remote_name, NULL);
|
2019-06-25 16:40:32 +03:00
|
|
|
if (!r)
|
2021-06-17 20:13:23 +03:00
|
|
|
r = promisor_remote_new(config, remote_name);
|
2019-06-25 16:40:32 +03:00
|
|
|
|
|
|
|
free(remote_name);
|
|
|
|
|
|
|
|
if (!r)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
return git_config_string(&r->partial_clone_filter, var, value);
|
|
|
|
}
|
2019-06-25 16:40:27 +03:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
static void promisor_remote_init(struct repository *r)
|
2019-06-25 16:40:27 +03:00
|
|
|
{
|
2021-06-17 20:13:23 +03:00
|
|
|
struct promisor_remote_config *config;
|
|
|
|
|
|
|
|
if (r->promisor_remote_config)
|
2019-06-25 16:40:27 +03:00
|
|
|
return;
|
2021-06-17 20:13:23 +03:00
|
|
|
config = r->promisor_remote_config =
|
2022-08-23 12:57:33 +03:00
|
|
|
xcalloc(1, sizeof(*r->promisor_remote_config));
|
2021-06-17 20:13:23 +03:00
|
|
|
config->promisors_tail = &config->promisors;
|
2019-06-25 16:40:27 +03:00
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
repo_config(r, promisor_remote_config, config);
|
2019-06-25 16:40:30 +03:00
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
if (r->repository_format_partial_clone) {
|
2019-06-25 16:40:30 +03:00
|
|
|
struct promisor_remote *o, *previous;
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
o = promisor_remote_lookup(config,
|
|
|
|
r->repository_format_partial_clone,
|
2019-06-25 16:40:30 +03:00
|
|
|
&previous);
|
|
|
|
if (o)
|
2021-06-17 20:13:23 +03:00
|
|
|
promisor_remote_move_to_tail(config, o, previous);
|
2019-06-25 16:40:30 +03:00
|
|
|
else
|
2021-06-17 20:13:23 +03:00
|
|
|
promisor_remote_new(config, r->repository_format_partial_clone);
|
2019-06-25 16:40:30 +03:00
|
|
|
}
|
2019-06-25 16:40:27 +03:00
|
|
|
}
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
void promisor_remote_clear(struct promisor_remote_config *config)
|
2019-06-25 16:40:29 +03:00
|
|
|
{
|
2021-06-17 20:13:23 +03:00
|
|
|
while (config->promisors) {
|
|
|
|
struct promisor_remote *r = config->promisors;
|
|
|
|
config->promisors = config->promisors->next;
|
2019-06-25 16:40:29 +03:00
|
|
|
free(r);
|
|
|
|
}
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
config->promisors_tail = &config->promisors;
|
2019-06-25 16:40:29 +03:00
|
|
|
}
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
void repo_promisor_remote_reinit(struct repository *r)
|
2019-06-25 16:40:29 +03:00
|
|
|
{
|
2021-06-17 20:13:23 +03:00
|
|
|
promisor_remote_clear(r->promisor_remote_config);
|
|
|
|
FREE_AND_NULL(r->promisor_remote_config);
|
|
|
|
promisor_remote_init(r);
|
2019-06-25 16:40:29 +03:00
|
|
|
}
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
struct promisor_remote *repo_promisor_remote_find(struct repository *r,
|
|
|
|
const char *remote_name)
|
2019-06-25 16:40:27 +03:00
|
|
|
{
|
2021-06-17 20:13:23 +03:00
|
|
|
promisor_remote_init(r);
|
2019-06-25 16:40:27 +03:00
|
|
|
|
|
|
|
if (!remote_name)
|
2021-06-17 20:13:23 +03:00
|
|
|
return r->promisor_remote_config->promisors;
|
2019-06-25 16:40:27 +03:00
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
return promisor_remote_lookup(r->promisor_remote_config, remote_name, NULL);
|
2019-06-25 16:40:27 +03:00
|
|
|
}
|
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
int repo_has_promisor_remote(struct repository *r)
|
2019-06-25 16:40:27 +03:00
|
|
|
{
|
2021-06-17 20:13:23 +03:00
|
|
|
return !!repo_promisor_remote_find(r, NULL);
|
2019-06-25 16:40:27 +03:00
|
|
|
}
|
2019-06-25 16:40:28 +03:00
|
|
|
|
|
|
|
static int remove_fetched_oids(struct repository *repo,
|
|
|
|
struct object_id **oids,
|
|
|
|
int oid_nr, int to_free)
|
|
|
|
{
|
|
|
|
int i, remaining_nr = 0;
|
|
|
|
int *remaining = xcalloc(oid_nr, sizeof(*remaining));
|
|
|
|
struct object_id *old_oids = *oids;
|
|
|
|
struct object_id *new_oids;
|
|
|
|
|
|
|
|
for (i = 0; i < oid_nr; i++)
|
|
|
|
if (oid_object_info_extended(repo, &old_oids[i], NULL,
|
|
|
|
OBJECT_INFO_SKIP_FETCH_OBJECT)) {
|
|
|
|
remaining[i] = 1;
|
|
|
|
remaining_nr++;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (remaining_nr) {
|
|
|
|
int j = 0;
|
2021-03-13 19:17:22 +03:00
|
|
|
CALLOC_ARRAY(new_oids, remaining_nr);
|
2019-06-25 16:40:28 +03:00
|
|
|
for (i = 0; i < oid_nr; i++)
|
|
|
|
if (remaining[i])
|
|
|
|
oidcpy(&new_oids[j++], &old_oids[i]);
|
|
|
|
*oids = new_oids;
|
|
|
|
if (to_free)
|
|
|
|
free(old_oids);
|
|
|
|
}
|
|
|
|
|
|
|
|
free(remaining);
|
|
|
|
|
|
|
|
return remaining_nr;
|
|
|
|
}
|
|
|
|
|
2022-10-05 00:13:40 +03:00
|
|
|
void promisor_remote_get_direct(struct repository *repo,
|
|
|
|
const struct object_id *oids,
|
|
|
|
int oid_nr)
|
2019-06-25 16:40:28 +03:00
|
|
|
{
|
|
|
|
struct promisor_remote *r;
|
|
|
|
struct object_id *remaining_oids = (struct object_id *)oids;
|
|
|
|
int remaining_nr = oid_nr;
|
|
|
|
int to_free = 0;
|
promisor-remote: die upon failing fetch
In a partial clone, an attempt to read a missing object results in an
attempt to fetch that single object. In order to avoid multiple
sequential fetches, which would occur when multiple objects are missing
(which is the typical case), some commands have been taught to prefetch
in a batch: such a command would, in a partial clone, notice that
several objects that it will eventually need are missing, and call
promisor_remote_get_direct() with all such objects at once.
When this batch prefetch fails, these commands fall back to the
sequential fetches. But at $DAYJOB we have noticed that this results in
a bad user experience: a command would take unexpectedly long to finish
(and possibly use up a lot of bandwidth) if the batch prefetch would
fail for some intermittent reason, but all subsequent fetches would
work. It would be a better user experience for such a command would
just fail.
Therefore, make it a fatal error if the prefetch fails and at least one
object being fetched is known to be a promisor object. (The latter
criterion is to make sure that we are not misleading the user that such
an object would be present from the promisor remote. For example, a
missing object may be a result of repository corruption and not because
it is expectedly missing due to the repository being a partial clone.)
Signed-off-by: Jonathan Tan <jonathantanmy@google.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2022-10-05 00:13:41 +03:00
|
|
|
int i;
|
2019-06-25 16:40:28 +03:00
|
|
|
|
2020-04-02 22:19:16 +03:00
|
|
|
if (oid_nr == 0)
|
2022-10-05 00:13:40 +03:00
|
|
|
return;
|
2020-04-02 22:19:16 +03:00
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
promisor_remote_init(repo);
|
2019-06-25 16:40:28 +03:00
|
|
|
|
2021-06-17 20:13:23 +03:00
|
|
|
for (r = repo->promisor_remote_config->promisors; r; r = r->next) {
|
2021-06-16 01:41:42 +03:00
|
|
|
if (fetch_objects(repo, r->name, remaining_oids, remaining_nr) < 0) {
|
2019-06-25 16:40:28 +03:00
|
|
|
if (remaining_nr == 1)
|
|
|
|
continue;
|
|
|
|
remaining_nr = remove_fetched_oids(repo, &remaining_oids,
|
|
|
|
remaining_nr, to_free);
|
|
|
|
if (remaining_nr) {
|
|
|
|
to_free = 1;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
}
|
promisor-remote: die upon failing fetch
In a partial clone, an attempt to read a missing object results in an
attempt to fetch that single object. In order to avoid multiple
sequential fetches, which would occur when multiple objects are missing
(which is the typical case), some commands have been taught to prefetch
in a batch: such a command would, in a partial clone, notice that
several objects that it will eventually need are missing, and call
promisor_remote_get_direct() with all such objects at once.
When this batch prefetch fails, these commands fall back to the
sequential fetches. But at $DAYJOB we have noticed that this results in
a bad user experience: a command would take unexpectedly long to finish
(and possibly use up a lot of bandwidth) if the batch prefetch would
fail for some intermittent reason, but all subsequent fetches would
work. It would be a better user experience for such a command would
just fail.
Therefore, make it a fatal error if the prefetch fails and at least one
object being fetched is known to be a promisor object. (The latter
criterion is to make sure that we are not misleading the user that such
an object would be present from the promisor remote. For example, a
missing object may be a result of repository corruption and not because
it is expectedly missing due to the repository being a partial clone.)
Signed-off-by: Jonathan Tan <jonathantanmy@google.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2022-10-05 00:13:41 +03:00
|
|
|
goto all_fetched;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < remaining_nr; i++) {
|
|
|
|
if (is_promisor_object(&remaining_oids[i]))
|
|
|
|
die(_("could not fetch %s from promisor remote"),
|
|
|
|
oid_to_hex(&remaining_oids[i]));
|
2019-06-25 16:40:28 +03:00
|
|
|
}
|
|
|
|
|
promisor-remote: die upon failing fetch
In a partial clone, an attempt to read a missing object results in an
attempt to fetch that single object. In order to avoid multiple
sequential fetches, which would occur when multiple objects are missing
(which is the typical case), some commands have been taught to prefetch
in a batch: such a command would, in a partial clone, notice that
several objects that it will eventually need are missing, and call
promisor_remote_get_direct() with all such objects at once.
When this batch prefetch fails, these commands fall back to the
sequential fetches. But at $DAYJOB we have noticed that this results in
a bad user experience: a command would take unexpectedly long to finish
(and possibly use up a lot of bandwidth) if the batch prefetch would
fail for some intermittent reason, but all subsequent fetches would
work. It would be a better user experience for such a command would
just fail.
Therefore, make it a fatal error if the prefetch fails and at least one
object being fetched is known to be a promisor object. (The latter
criterion is to make sure that we are not misleading the user that such
an object would be present from the promisor remote. For example, a
missing object may be a result of repository corruption and not because
it is expectedly missing due to the repository being a partial clone.)
Signed-off-by: Jonathan Tan <jonathantanmy@google.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2022-10-05 00:13:41 +03:00
|
|
|
all_fetched:
|
2019-06-25 16:40:28 +03:00
|
|
|
if (to_free)
|
|
|
|
free(remaining_oids);
|
|
|
|
}
|