2017-08-19 01:20:16 +03:00
|
|
|
#include "cache.h"
|
2018-01-24 02:46:51 +03:00
|
|
|
#include "list.h"
|
2017-08-19 01:20:19 +03:00
|
|
|
#include "pack.h"
|
2018-03-23 20:20:57 +03:00
|
|
|
#include "repository.h"
|
2017-08-19 01:20:26 +03:00
|
|
|
#include "dir.h"
|
|
|
|
#include "mergesort.h"
|
|
|
|
#include "packfile.h"
|
2017-08-19 01:20:28 +03:00
|
|
|
#include "delta.h"
|
2017-08-19 01:20:30 +03:00
|
|
|
#include "streaming.h"
|
2017-08-19 01:20:33 +03:00
|
|
|
#include "sha1-lookup.h"
|
2017-12-05 19:58:44 +03:00
|
|
|
#include "commit.h"
|
|
|
|
#include "object.h"
|
|
|
|
#include "tag.h"
|
|
|
|
#include "tree-walk.h"
|
|
|
|
#include "tree.h"
|
2018-03-23 20:20:56 +03:00
|
|
|
#include "object-store.h"
|
2018-07-12 22:39:33 +03:00
|
|
|
#include "midx.h"
|
2019-05-17 21:41:48 +03:00
|
|
|
#include "commit-graph.h"
|
2019-06-25 16:40:31 +03:00
|
|
|
#include "promisor-remote.h"
|
2017-08-19 01:20:16 +03:00
|
|
|
|
|
|
|
char *odb_pack_name(struct strbuf *buf,
|
2019-08-18 23:04:22 +03:00
|
|
|
const unsigned char *hash,
|
2017-08-19 01:20:16 +03:00
|
|
|
const char *ext)
|
|
|
|
{
|
|
|
|
strbuf_reset(buf);
|
|
|
|
strbuf_addf(buf, "%s/pack/pack-%s.%s", get_object_directory(),
|
2019-08-18 23:04:22 +03:00
|
|
|
hash_to_hex(hash), ext);
|
2017-08-19 01:20:16 +03:00
|
|
|
return buf->buf;
|
|
|
|
}
|
|
|
|
|
|
|
|
char *sha1_pack_name(const unsigned char *sha1)
|
|
|
|
{
|
|
|
|
static struct strbuf buf = STRBUF_INIT;
|
|
|
|
return odb_pack_name(&buf, sha1, "pack");
|
|
|
|
}
|
|
|
|
|
|
|
|
char *sha1_pack_index_name(const unsigned char *sha1)
|
|
|
|
{
|
|
|
|
static struct strbuf buf = STRBUF_INIT;
|
|
|
|
return odb_pack_name(&buf, sha1, "idx");
|
|
|
|
}
|
2017-08-19 01:20:17 +03:00
|
|
|
|
2017-08-19 01:20:22 +03:00
|
|
|
static unsigned int pack_used_ctr;
|
|
|
|
static unsigned int pack_mmap_calls;
|
|
|
|
static unsigned int peak_pack_open_windows;
|
|
|
|
static unsigned int pack_open_windows;
|
2017-08-19 01:20:25 +03:00
|
|
|
static unsigned int pack_open_fds;
|
2017-08-19 01:20:22 +03:00
|
|
|
static unsigned int pack_max_fds;
|
|
|
|
static size_t peak_pack_mapped;
|
|
|
|
static size_t pack_mapped;
|
2017-08-19 01:20:18 +03:00
|
|
|
|
|
|
|
#define SZ_FMT PRIuMAX
|
|
|
|
static inline uintmax_t sz_fmt(size_t s) { return s; }
|
|
|
|
|
|
|
|
void pack_report(void)
|
|
|
|
{
|
|
|
|
fprintf(stderr,
|
|
|
|
"pack_report: getpagesize() = %10" SZ_FMT "\n"
|
|
|
|
"pack_report: core.packedGitWindowSize = %10" SZ_FMT "\n"
|
|
|
|
"pack_report: core.packedGitLimit = %10" SZ_FMT "\n",
|
|
|
|
sz_fmt(getpagesize()),
|
|
|
|
sz_fmt(packed_git_window_size),
|
|
|
|
sz_fmt(packed_git_limit));
|
|
|
|
fprintf(stderr,
|
|
|
|
"pack_report: pack_used_ctr = %10u\n"
|
|
|
|
"pack_report: pack_mmap_calls = %10u\n"
|
|
|
|
"pack_report: pack_open_windows = %10u / %10u\n"
|
|
|
|
"pack_report: pack_mapped = "
|
|
|
|
"%10" SZ_FMT " / %10" SZ_FMT "\n",
|
|
|
|
pack_used_ctr,
|
|
|
|
pack_mmap_calls,
|
|
|
|
pack_open_windows, peak_pack_open_windows,
|
|
|
|
sz_fmt(pack_mapped), sz_fmt(peak_pack_mapped));
|
|
|
|
}
|
2017-08-19 01:20:19 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Open and mmap the index file at path, perform a couple of
|
|
|
|
* consistency checks, then record its information to p. Return 0 on
|
|
|
|
* success.
|
|
|
|
*/
|
|
|
|
static int check_packed_git_idx(const char *path, struct packed_git *p)
|
|
|
|
{
|
|
|
|
void *idx_map;
|
|
|
|
size_t idx_size;
|
2018-10-13 03:58:41 +03:00
|
|
|
int fd = git_open(path), ret;
|
2017-08-19 01:20:19 +03:00
|
|
|
struct stat st;
|
2018-05-02 03:25:36 +03:00
|
|
|
const unsigned int hashsz = the_hash_algo->rawsz;
|
2017-08-19 01:20:19 +03:00
|
|
|
|
|
|
|
if (fd < 0)
|
|
|
|
return -1;
|
|
|
|
if (fstat(fd, &st)) {
|
|
|
|
close(fd);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
idx_size = xsize_t(st.st_size);
|
2018-05-02 03:25:36 +03:00
|
|
|
if (idx_size < 4 * 256 + hashsz + hashsz) {
|
2017-08-19 01:20:19 +03:00
|
|
|
close(fd);
|
|
|
|
return error("index file %s is too small", path);
|
|
|
|
}
|
|
|
|
idx_map = xmmap(NULL, idx_size, PROT_READ, MAP_PRIVATE, fd, 0);
|
|
|
|
close(fd);
|
|
|
|
|
2018-10-13 03:58:41 +03:00
|
|
|
ret = load_idx(path, hashsz, idx_map, idx_size, p);
|
|
|
|
|
|
|
|
if (ret)
|
|
|
|
munmap(idx_map, idx_size);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
int load_idx(const char *path, const unsigned int hashsz, void *idx_map,
|
|
|
|
size_t idx_size, struct packed_git *p)
|
|
|
|
{
|
|
|
|
struct pack_idx_header *hdr = idx_map;
|
|
|
|
uint32_t version, nr, i, *index;
|
|
|
|
|
|
|
|
if (idx_size < 4 * 256 + hashsz + hashsz)
|
|
|
|
return error("index file %s is too small", path);
|
|
|
|
if (idx_map == NULL)
|
|
|
|
return error("empty data");
|
|
|
|
|
2017-08-19 01:20:19 +03:00
|
|
|
if (hdr->idx_signature == htonl(PACK_IDX_SIGNATURE)) {
|
|
|
|
version = ntohl(hdr->idx_version);
|
2018-10-13 03:58:41 +03:00
|
|
|
if (version < 2 || version > 2)
|
2017-08-19 01:20:19 +03:00
|
|
|
return error("index file %s is version %"PRIu32
|
|
|
|
" and is not supported by this binary"
|
|
|
|
" (try upgrading GIT to a newer version)",
|
|
|
|
path, version);
|
|
|
|
} else
|
|
|
|
version = 1;
|
|
|
|
|
|
|
|
nr = 0;
|
|
|
|
index = idx_map;
|
|
|
|
if (version > 1)
|
|
|
|
index += 2; /* skip index header */
|
|
|
|
for (i = 0; i < 256; i++) {
|
|
|
|
uint32_t n = ntohl(index[i]);
|
2018-10-13 03:58:41 +03:00
|
|
|
if (n < nr)
|
2017-08-19 01:20:19 +03:00
|
|
|
return error("non-monotonic index %s", path);
|
|
|
|
nr = n;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (version == 1) {
|
|
|
|
/*
|
|
|
|
* Total size:
|
|
|
|
* - 256 index entries 4 bytes each
|
2018-05-02 03:25:36 +03:00
|
|
|
* - 24-byte entries * nr (object ID + 4-byte offset)
|
|
|
|
* - hash of the packfile
|
|
|
|
* - file checksum
|
2017-08-19 01:20:19 +03:00
|
|
|
*/
|
2018-10-13 03:58:41 +03:00
|
|
|
if (idx_size != 4 * 256 + nr * (hashsz + 4) + hashsz + hashsz)
|
2017-08-19 01:20:19 +03:00
|
|
|
return error("wrong index v1 file size in %s", path);
|
|
|
|
} else if (version == 2) {
|
|
|
|
/*
|
|
|
|
* Minimum size:
|
|
|
|
* - 8 bytes of header
|
|
|
|
* - 256 index entries 4 bytes each
|
2018-05-02 03:25:36 +03:00
|
|
|
* - object ID entry * nr
|
2017-08-19 01:20:19 +03:00
|
|
|
* - 4-byte crc entry * nr
|
|
|
|
* - 4-byte offset entry * nr
|
2018-05-02 03:25:36 +03:00
|
|
|
* - hash of the packfile
|
|
|
|
* - file checksum
|
2017-08-19 01:20:19 +03:00
|
|
|
* And after the 4-byte offset table might be a
|
|
|
|
* variable sized table containing 8-byte entries
|
|
|
|
* for offsets larger than 2^31.
|
|
|
|
*/
|
2018-05-02 03:25:36 +03:00
|
|
|
unsigned long min_size = 8 + 4*256 + nr*(hashsz + 4 + 4) + hashsz + hashsz;
|
2017-08-19 01:20:19 +03:00
|
|
|
unsigned long max_size = min_size;
|
|
|
|
if (nr)
|
|
|
|
max_size += (nr - 1)*8;
|
2018-10-13 03:58:41 +03:00
|
|
|
if (idx_size < min_size || idx_size > max_size)
|
2017-08-19 01:20:19 +03:00
|
|
|
return error("wrong index v2 file size in %s", path);
|
|
|
|
if (idx_size != min_size &&
|
|
|
|
/*
|
|
|
|
* make sure we can deal with large pack offsets.
|
|
|
|
* 31-bit signed offset won't be enough, neither
|
|
|
|
* 32-bit unsigned one will be.
|
|
|
|
*/
|
2018-10-13 03:58:41 +03:00
|
|
|
(sizeof(off_t) <= 4))
|
2017-08-19 01:20:19 +03:00
|
|
|
return error("pack too large for current definition of off_t in %s", path);
|
2020-05-25 22:59:10 +03:00
|
|
|
p->crc_offset = 8 + 4 * 256 + nr * hashsz;
|
2017-08-19 01:20:19 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
p->index_version = version;
|
|
|
|
p->index_data = idx_map;
|
|
|
|
p->index_size = idx_size;
|
|
|
|
p->num_objects = nr;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int open_pack_index(struct packed_git *p)
|
|
|
|
{
|
|
|
|
char *idx_name;
|
|
|
|
size_t len;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (p->index_data)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (!strip_suffix(p->pack_name, ".pack", &len))
|
2018-05-02 12:38:39 +03:00
|
|
|
BUG("pack_name does not end in .pack");
|
2017-08-19 01:20:19 +03:00
|
|
|
idx_name = xstrfmt("%.*s.idx", (int)len, p->pack_name);
|
|
|
|
ret = check_packed_git_idx(idx_name, p);
|
|
|
|
free(idx_name);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2018-07-12 22:39:29 +03:00
|
|
|
uint32_t get_pack_fanout(struct packed_git *p, uint32_t value)
|
|
|
|
{
|
|
|
|
const uint32_t *level1_ofs = p->index_data;
|
|
|
|
|
|
|
|
if (!level1_ofs) {
|
|
|
|
if (open_pack_index(p))
|
|
|
|
return 0;
|
|
|
|
level1_ofs = p->index_data;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (p->index_version > 1) {
|
|
|
|
level1_ofs += 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ntohl(level1_ofs[value]);
|
|
|
|
}
|
|
|
|
|
2017-08-19 01:20:19 +03:00
|
|
|
static struct packed_git *alloc_packed_git(int extra)
|
|
|
|
{
|
|
|
|
struct packed_git *p = xmalloc(st_add(sizeof(*p), extra));
|
|
|
|
memset(p, 0, sizeof(*p));
|
|
|
|
p->pack_fd = -1;
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct packed_git *parse_pack_index(unsigned char *sha1, const char *idx_path)
|
|
|
|
{
|
|
|
|
const char *path = sha1_pack_name(sha1);
|
|
|
|
size_t alloc = st_add(strlen(path), 1);
|
|
|
|
struct packed_git *p = alloc_packed_git(alloc);
|
|
|
|
|
|
|
|
memcpy(p->pack_name, path, alloc); /* includes NUL */
|
2019-02-19 03:05:03 +03:00
|
|
|
hashcpy(p->hash, sha1);
|
2017-08-19 01:20:19 +03:00
|
|
|
if (check_packed_git_idx(idx_path, p)) {
|
|
|
|
free(p);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
return p;
|
|
|
|
}
|
2017-08-19 01:20:20 +03:00
|
|
|
|
|
|
|
static void scan_windows(struct packed_git *p,
|
|
|
|
struct packed_git **lru_p,
|
|
|
|
struct pack_window **lru_w,
|
|
|
|
struct pack_window **lru_l)
|
|
|
|
{
|
|
|
|
struct pack_window *w, *w_l;
|
|
|
|
|
|
|
|
for (w_l = NULL, w = p->windows; w; w = w->next) {
|
|
|
|
if (!w->inuse_cnt) {
|
|
|
|
if (!*lru_w || w->last_used < (*lru_w)->last_used) {
|
|
|
|
*lru_p = p;
|
|
|
|
*lru_w = w;
|
|
|
|
*lru_l = w_l;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
w_l = w;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2017-08-19 01:20:22 +03:00
|
|
|
static int unuse_one_window(struct packed_git *current)
|
2017-08-19 01:20:20 +03:00
|
|
|
{
|
|
|
|
struct packed_git *p, *lru_p = NULL;
|
|
|
|
struct pack_window *lru_w = NULL, *lru_l = NULL;
|
|
|
|
|
|
|
|
if (current)
|
|
|
|
scan_windows(current, &lru_p, &lru_w, &lru_l);
|
2018-03-23 20:20:59 +03:00
|
|
|
for (p = the_repository->objects->packed_git; p; p = p->next)
|
2017-08-19 01:20:20 +03:00
|
|
|
scan_windows(p, &lru_p, &lru_w, &lru_l);
|
|
|
|
if (lru_p) {
|
|
|
|
munmap(lru_w->base, lru_w->len);
|
|
|
|
pack_mapped -= lru_w->len;
|
|
|
|
if (lru_l)
|
|
|
|
lru_l->next = lru_w->next;
|
|
|
|
else
|
|
|
|
lru_p->windows = lru_w->next;
|
|
|
|
free(lru_w);
|
|
|
|
pack_open_windows--;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-08-19 01:20:21 +03:00
|
|
|
void close_pack_windows(struct packed_git *p)
|
|
|
|
{
|
|
|
|
while (p->windows) {
|
|
|
|
struct pack_window *w = p->windows;
|
|
|
|
|
|
|
|
if (w->inuse_cnt)
|
|
|
|
die("pack '%s' still has open windows to it",
|
|
|
|
p->pack_name);
|
|
|
|
munmap(w->base, w->len);
|
|
|
|
pack_mapped -= w->len;
|
|
|
|
pack_open_windows--;
|
|
|
|
p->windows = w->next;
|
|
|
|
free(w);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-03-21 22:36:15 +03:00
|
|
|
int close_pack_fd(struct packed_git *p)
|
2017-08-19 01:20:21 +03:00
|
|
|
{
|
|
|
|
if (p->pack_fd < 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
close(p->pack_fd);
|
|
|
|
pack_open_fds--;
|
|
|
|
p->pack_fd = -1;
|
|
|
|
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
void close_pack_index(struct packed_git *p)
|
|
|
|
{
|
|
|
|
if (p->index_data) {
|
|
|
|
munmap((void *)p->index_data, p->index_size);
|
|
|
|
p->index_data = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-04-10 15:56:06 +03:00
|
|
|
void close_pack(struct packed_git *p)
|
2017-08-19 01:20:21 +03:00
|
|
|
{
|
|
|
|
close_pack_windows(p);
|
|
|
|
close_pack_fd(p);
|
|
|
|
close_pack_index(p);
|
|
|
|
}
|
|
|
|
|
2019-05-17 21:41:49 +03:00
|
|
|
void close_object_store(struct raw_object_store *o)
|
2017-08-19 01:20:21 +03:00
|
|
|
{
|
|
|
|
struct packed_git *p;
|
|
|
|
|
2018-03-23 20:21:00 +03:00
|
|
|
for (p = o->packed_git; p; p = p->next)
|
2017-08-19 01:20:21 +03:00
|
|
|
if (p->do_not_close)
|
2018-05-02 12:38:39 +03:00
|
|
|
BUG("want to close pack marked 'do-not-close'");
|
2017-08-19 01:20:21 +03:00
|
|
|
else
|
|
|
|
close_pack(p);
|
2018-10-25 15:54:05 +03:00
|
|
|
|
|
|
|
if (o->multi_pack_index) {
|
|
|
|
close_midx(o->multi_pack_index);
|
|
|
|
o->multi_pack_index = NULL;
|
|
|
|
}
|
2019-05-17 21:41:48 +03:00
|
|
|
|
|
|
|
close_commit_graph(o);
|
2017-08-19 01:20:21 +03:00
|
|
|
}
|
2017-08-19 01:20:22 +03:00
|
|
|
|
2019-06-11 02:35:22 +03:00
|
|
|
void unlink_pack_path(const char *pack_name, int force_delete)
|
|
|
|
{
|
|
|
|
static const char *exts[] = {".pack", ".idx", ".keep", ".bitmap", ".promisor"};
|
|
|
|
int i;
|
|
|
|
struct strbuf buf = STRBUF_INIT;
|
|
|
|
size_t plen;
|
|
|
|
|
|
|
|
strbuf_addstr(&buf, pack_name);
|
|
|
|
strip_suffix_mem(buf.buf, &buf.len, ".pack");
|
|
|
|
plen = buf.len;
|
|
|
|
|
|
|
|
if (!force_delete) {
|
|
|
|
strbuf_addstr(&buf, ".keep");
|
|
|
|
if (!access(buf.buf, F_OK)) {
|
|
|
|
strbuf_release(&buf);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(exts); i++) {
|
|
|
|
strbuf_setlen(&buf, plen);
|
|
|
|
strbuf_addstr(&buf, exts[i]);
|
|
|
|
unlink(buf.buf);
|
|
|
|
}
|
|
|
|
|
|
|
|
strbuf_release(&buf);
|
|
|
|
}
|
|
|
|
|
2017-08-19 01:20:22 +03:00
|
|
|
/*
|
|
|
|
* The LRU pack is the one with the oldest MRU window, preferring packs
|
|
|
|
* with no used windows, or the oldest mtime if it has no windows allocated.
|
|
|
|
*/
|
|
|
|
static void find_lru_pack(struct packed_git *p, struct packed_git **lru_p, struct pack_window **mru_w, int *accept_windows_inuse)
|
|
|
|
{
|
|
|
|
struct pack_window *w, *this_mru_w;
|
|
|
|
int has_windows_inuse = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Reject this pack if it has windows and the previously selected
|
|
|
|
* one does not. If this pack does not have windows, reject
|
|
|
|
* it if the pack file is newer than the previously selected one.
|
|
|
|
*/
|
|
|
|
if (*lru_p && !*mru_w && (p->windows || p->mtime > (*lru_p)->mtime))
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (w = this_mru_w = p->windows; w; w = w->next) {
|
|
|
|
/*
|
|
|
|
* Reject this pack if any of its windows are in use,
|
|
|
|
* but the previously selected pack did not have any
|
|
|
|
* inuse windows. Otherwise, record that this pack
|
|
|
|
* has windows in use.
|
|
|
|
*/
|
|
|
|
if (w->inuse_cnt) {
|
|
|
|
if (*accept_windows_inuse)
|
|
|
|
has_windows_inuse = 1;
|
|
|
|
else
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (w->last_used > this_mru_w->last_used)
|
|
|
|
this_mru_w = w;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Reject this pack if it has windows that have been
|
|
|
|
* used more recently than the previously selected pack.
|
|
|
|
* If the previously selected pack had windows inuse and
|
|
|
|
* we have not encountered a window in this pack that is
|
|
|
|
* inuse, skip this check since we prefer a pack with no
|
|
|
|
* inuse windows to one that has inuse windows.
|
|
|
|
*/
|
|
|
|
if (*mru_w && *accept_windows_inuse == has_windows_inuse &&
|
|
|
|
this_mru_w->last_used > (*mru_w)->last_used)
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Select this pack.
|
|
|
|
*/
|
|
|
|
*mru_w = this_mru_w;
|
|
|
|
*lru_p = p;
|
|
|
|
*accept_windows_inuse = has_windows_inuse;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int close_one_pack(void)
|
|
|
|
{
|
|
|
|
struct packed_git *p, *lru_p = NULL;
|
|
|
|
struct pack_window *mru_w = NULL;
|
|
|
|
int accept_windows_inuse = 1;
|
|
|
|
|
2018-03-23 20:20:59 +03:00
|
|
|
for (p = the_repository->objects->packed_git; p; p = p->next) {
|
2017-08-19 01:20:22 +03:00
|
|
|
if (p->pack_fd == -1)
|
|
|
|
continue;
|
|
|
|
find_lru_pack(p, &lru_p, &mru_w, &accept_windows_inuse);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (lru_p)
|
|
|
|
return close_pack_fd(lru_p);
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int get_max_fd_limit(void)
|
|
|
|
{
|
|
|
|
#ifdef RLIMIT_NOFILE
|
|
|
|
{
|
|
|
|
struct rlimit lim;
|
|
|
|
|
|
|
|
if (!getrlimit(RLIMIT_NOFILE, &lim))
|
|
|
|
return lim.rlim_cur;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef _SC_OPEN_MAX
|
|
|
|
{
|
|
|
|
long open_max = sysconf(_SC_OPEN_MAX);
|
|
|
|
if (0 < open_max)
|
|
|
|
return open_max;
|
|
|
|
/*
|
|
|
|
* Otherwise, we got -1 for one of the two
|
|
|
|
* reasons:
|
|
|
|
*
|
|
|
|
* (1) sysconf() did not understand _SC_OPEN_MAX
|
|
|
|
* and signaled an error with -1; or
|
|
|
|
* (2) sysconf() said there is no limit.
|
|
|
|
*
|
|
|
|
* We _could_ clear errno before calling sysconf() to
|
|
|
|
* tell these two cases apart and return a huge number
|
|
|
|
* in the latter case to let the caller cap it to a
|
|
|
|
* value that is not so selfish, but letting the
|
|
|
|
* fallback OPEN_MAX codepath take care of these cases
|
|
|
|
* is a lot simpler.
|
|
|
|
*/
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef OPEN_MAX
|
|
|
|
return OPEN_MAX;
|
|
|
|
#else
|
|
|
|
return 1; /* see the caller ;-) */
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2019-04-05 21:06:22 +03:00
|
|
|
const char *pack_basename(struct packed_git *p)
|
|
|
|
{
|
|
|
|
const char *ret = strrchr(p->pack_name, '/');
|
|
|
|
if (ret)
|
|
|
|
ret = ret + 1; /* skip past slash */
|
|
|
|
else
|
|
|
|
ret = p->pack_name; /* we only have a base */
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2017-08-19 01:20:22 +03:00
|
|
|
/*
|
|
|
|
* Do not call this directly as this leaks p->pack_fd on error return;
|
|
|
|
* call open_packed_git() instead.
|
|
|
|
*/
|
|
|
|
static int open_packed_git_1(struct packed_git *p)
|
|
|
|
{
|
|
|
|
struct stat st;
|
|
|
|
struct pack_header hdr;
|
2018-05-02 03:25:36 +03:00
|
|
|
unsigned char hash[GIT_MAX_RAWSZ];
|
|
|
|
unsigned char *idx_hash;
|
2017-09-27 09:02:11 +03:00
|
|
|
ssize_t read_result;
|
2018-05-02 03:25:36 +03:00
|
|
|
const unsigned hashsz = the_hash_algo->rawsz;
|
2017-08-19 01:20:22 +03:00
|
|
|
|
2018-07-12 22:39:39 +03:00
|
|
|
if (!p->index_data) {
|
|
|
|
struct multi_pack_index *m;
|
2019-04-05 21:06:22 +03:00
|
|
|
const char *pack_name = pack_basename(p);
|
2018-07-12 22:39:39 +03:00
|
|
|
|
|
|
|
for (m = the_repository->objects->multi_pack_index;
|
|
|
|
m; m = m->next) {
|
|
|
|
if (midx_contains_pack(m, pack_name))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!m && open_pack_index(p))
|
|
|
|
return error("packfile %s index unavailable", p->pack_name);
|
|
|
|
}
|
2017-08-19 01:20:22 +03:00
|
|
|
|
|
|
|
if (!pack_max_fds) {
|
|
|
|
unsigned int max_fds = get_max_fd_limit();
|
|
|
|
|
|
|
|
/* Save 3 for stdin/stdout/stderr, 22 for work */
|
|
|
|
if (25 < max_fds)
|
|
|
|
pack_max_fds = max_fds - 25;
|
|
|
|
else
|
|
|
|
pack_max_fds = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
while (pack_max_fds <= pack_open_fds && close_one_pack())
|
|
|
|
; /* nothing */
|
|
|
|
|
|
|
|
p->pack_fd = git_open(p->pack_name);
|
|
|
|
if (p->pack_fd < 0 || fstat(p->pack_fd, &st))
|
|
|
|
return -1;
|
|
|
|
pack_open_fds++;
|
|
|
|
|
|
|
|
/* If we created the struct before we had the pack we lack size. */
|
|
|
|
if (!p->pack_size) {
|
|
|
|
if (!S_ISREG(st.st_mode))
|
|
|
|
return error("packfile %s not a regular file", p->pack_name);
|
|
|
|
p->pack_size = st.st_size;
|
|
|
|
} else if (p->pack_size != st.st_size)
|
|
|
|
return error("packfile %s size changed", p->pack_name);
|
|
|
|
|
|
|
|
/* Verify we recognize this pack file format. */
|
2017-09-27 09:02:11 +03:00
|
|
|
read_result = read_in_full(p->pack_fd, &hdr, sizeof(hdr));
|
|
|
|
if (read_result < 0)
|
|
|
|
return error_errno("error reading from %s", p->pack_name);
|
|
|
|
if (read_result != sizeof(hdr))
|
2017-08-19 01:20:22 +03:00
|
|
|
return error("file %s is far too short to be a packfile", p->pack_name);
|
|
|
|
if (hdr.hdr_signature != htonl(PACK_SIGNATURE))
|
|
|
|
return error("file %s is not a GIT packfile", p->pack_name);
|
|
|
|
if (!pack_version_ok(hdr.hdr_version))
|
|
|
|
return error("packfile %s is version %"PRIu32" and not"
|
|
|
|
" supported (try upgrading GIT to a newer version)",
|
|
|
|
p->pack_name, ntohl(hdr.hdr_version));
|
|
|
|
|
2018-07-12 22:39:39 +03:00
|
|
|
/* Skip index checking if in multi-pack-index */
|
|
|
|
if (!p->index_data)
|
|
|
|
return 0;
|
|
|
|
|
2017-08-19 01:20:22 +03:00
|
|
|
/* Verify the pack matches its index. */
|
|
|
|
if (p->num_objects != ntohl(hdr.hdr_entries))
|
|
|
|
return error("packfile %s claims to have %"PRIu32" objects"
|
|
|
|
" while index indicates %"PRIu32" objects",
|
|
|
|
p->pack_name, ntohl(hdr.hdr_entries),
|
|
|
|
p->num_objects);
|
2019-12-26 13:42:20 +03:00
|
|
|
read_result = pread_in_full(p->pack_fd, hash, hashsz,
|
|
|
|
p->pack_size - hashsz);
|
2017-09-27 09:02:11 +03:00
|
|
|
if (read_result < 0)
|
|
|
|
return error_errno("error reading from %s", p->pack_name);
|
2018-05-02 03:25:36 +03:00
|
|
|
if (read_result != hashsz)
|
2017-08-19 01:20:22 +03:00
|
|
|
return error("packfile %s signature is unavailable", p->pack_name);
|
2018-05-02 03:25:36 +03:00
|
|
|
idx_hash = ((unsigned char *)p->index_data) + p->index_size - hashsz * 2;
|
2018-08-29 00:22:52 +03:00
|
|
|
if (!hasheq(hash, idx_hash))
|
2017-08-19 01:20:22 +03:00
|
|
|
return error("packfile %s does not match index", p->pack_name);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2017-08-19 01:20:33 +03:00
|
|
|
static int open_packed_git(struct packed_git *p)
|
2017-08-19 01:20:22 +03:00
|
|
|
{
|
|
|
|
if (!open_packed_git_1(p))
|
|
|
|
return 0;
|
|
|
|
close_pack_fd(p);
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int in_window(struct pack_window *win, off_t offset)
|
|
|
|
{
|
2018-05-02 03:25:36 +03:00
|
|
|
/* We must promise at least one full hash after the
|
2017-08-19 01:20:22 +03:00
|
|
|
* offset is available from this window, otherwise the offset
|
|
|
|
* is not actually in this window and a different window (which
|
|
|
|
* has that one hash excess) must be used. This is to support
|
|
|
|
* the object header and delta base parsing routines below.
|
|
|
|
*/
|
|
|
|
off_t win_off = win->offset;
|
|
|
|
return win_off <= offset
|
2018-05-02 03:25:36 +03:00
|
|
|
&& (offset + the_hash_algo->rawsz) <= (win_off + win->len);
|
2017-08-19 01:20:22 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
unsigned char *use_pack(struct packed_git *p,
|
|
|
|
struct pack_window **w_cursor,
|
|
|
|
off_t offset,
|
|
|
|
unsigned long *left)
|
|
|
|
{
|
|
|
|
struct pack_window *win = *w_cursor;
|
|
|
|
|
|
|
|
/* Since packfiles end in a hash of their content and it's
|
|
|
|
* pointless to ask for an offset into the middle of that
|
|
|
|
* hash, and the in_window function above wouldn't match
|
|
|
|
* don't allow an offset too close to the end of the file.
|
|
|
|
*/
|
|
|
|
if (!p->pack_size && p->pack_fd == -1 && open_packed_git(p))
|
|
|
|
die("packfile %s cannot be accessed", p->pack_name);
|
2018-05-02 03:25:36 +03:00
|
|
|
if (offset > (p->pack_size - the_hash_algo->rawsz))
|
2017-08-19 01:20:22 +03:00
|
|
|
die("offset beyond end of packfile (truncated pack?)");
|
|
|
|
if (offset < 0)
|
|
|
|
die(_("offset before end of packfile (broken .idx?)"));
|
|
|
|
|
|
|
|
if (!win || !in_window(win, offset)) {
|
|
|
|
if (win)
|
|
|
|
win->inuse_cnt--;
|
|
|
|
for (win = p->windows; win; win = win->next) {
|
|
|
|
if (in_window(win, offset))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (!win) {
|
|
|
|
size_t window_align = packed_git_window_size / 2;
|
|
|
|
off_t len;
|
|
|
|
|
|
|
|
if (p->pack_fd == -1 && open_packed_git(p))
|
|
|
|
die("packfile %s cannot be accessed", p->pack_name);
|
|
|
|
|
|
|
|
win = xcalloc(1, sizeof(*win));
|
|
|
|
win->offset = (offset / window_align) * window_align;
|
|
|
|
len = p->pack_size - win->offset;
|
|
|
|
if (len > packed_git_window_size)
|
|
|
|
len = packed_git_window_size;
|
|
|
|
win->len = (size_t)len;
|
|
|
|
pack_mapped += win->len;
|
|
|
|
while (packed_git_limit < pack_mapped
|
|
|
|
&& unuse_one_window(p))
|
|
|
|
; /* nothing */
|
2019-05-16 03:37:36 +03:00
|
|
|
win->base = xmmap_gently(NULL, win->len,
|
2017-08-19 01:20:22 +03:00
|
|
|
PROT_READ, MAP_PRIVATE,
|
|
|
|
p->pack_fd, win->offset);
|
|
|
|
if (win->base == MAP_FAILED)
|
|
|
|
die_errno("packfile %s cannot be mapped",
|
|
|
|
p->pack_name);
|
|
|
|
if (!win->offset && win->len == p->pack_size
|
|
|
|
&& !p->do_not_close)
|
|
|
|
close_pack_fd(p);
|
|
|
|
pack_mmap_calls++;
|
|
|
|
pack_open_windows++;
|
|
|
|
if (pack_mapped > peak_pack_mapped)
|
|
|
|
peak_pack_mapped = pack_mapped;
|
|
|
|
if (pack_open_windows > peak_pack_open_windows)
|
|
|
|
peak_pack_open_windows = pack_open_windows;
|
|
|
|
win->next = p->windows;
|
|
|
|
p->windows = win;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (win != *w_cursor) {
|
|
|
|
win->last_used = pack_used_ctr++;
|
|
|
|
win->inuse_cnt++;
|
|
|
|
*w_cursor = win;
|
|
|
|
}
|
|
|
|
offset -= win->offset;
|
|
|
|
if (left)
|
|
|
|
*left = win->len - xsize_t(offset);
|
|
|
|
return win->base + offset;
|
|
|
|
}
|
2017-08-19 01:20:23 +03:00
|
|
|
|
|
|
|
void unuse_pack(struct pack_window **w_cursor)
|
|
|
|
{
|
|
|
|
struct pack_window *w = *w_cursor;
|
|
|
|
if (w) {
|
|
|
|
w->inuse_cnt--;
|
|
|
|
*w_cursor = NULL;
|
|
|
|
}
|
|
|
|
}
|
2017-08-19 01:20:24 +03:00
|
|
|
|
|
|
|
struct packed_git *add_packed_git(const char *path, size_t path_len, int local)
|
|
|
|
{
|
|
|
|
struct stat st;
|
|
|
|
size_t alloc;
|
|
|
|
struct packed_git *p;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Make sure a corresponding .pack file exists and that
|
|
|
|
* the index looks sane.
|
|
|
|
*/
|
|
|
|
if (!strip_suffix_mem(path, &path_len, ".idx"))
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
/*
|
2017-12-05 19:58:44 +03:00
|
|
|
* ".promisor" is long enough to hold any suffix we're adding (and
|
2017-08-19 01:20:24 +03:00
|
|
|
* the use xsnprintf double-checks that)
|
|
|
|
*/
|
2017-12-05 19:58:44 +03:00
|
|
|
alloc = st_add3(path_len, strlen(".promisor"), 1);
|
2017-08-19 01:20:24 +03:00
|
|
|
p = alloc_packed_git(alloc);
|
|
|
|
memcpy(p->pack_name, path, path_len);
|
|
|
|
|
|
|
|
xsnprintf(p->pack_name + path_len, alloc - path_len, ".keep");
|
|
|
|
if (!access(p->pack_name, F_OK))
|
|
|
|
p->pack_keep = 1;
|
|
|
|
|
2017-12-05 19:58:44 +03:00
|
|
|
xsnprintf(p->pack_name + path_len, alloc - path_len, ".promisor");
|
|
|
|
if (!access(p->pack_name, F_OK))
|
|
|
|
p->pack_promisor = 1;
|
|
|
|
|
2017-08-19 01:20:24 +03:00
|
|
|
xsnprintf(p->pack_name + path_len, alloc - path_len, ".pack");
|
|
|
|
if (stat(p->pack_name, &st) || !S_ISREG(st.st_mode)) {
|
|
|
|
free(p);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* ok, it looks sane as far as we can check without
|
|
|
|
* actually mapping the pack file.
|
|
|
|
*/
|
|
|
|
p->pack_size = st.st_size;
|
|
|
|
p->pack_local = local;
|
|
|
|
p->mtime = st.st_mtime;
|
2018-05-02 03:25:36 +03:00
|
|
|
if (path_len < the_hash_algo->hexsz ||
|
2019-02-19 03:05:03 +03:00
|
|
|
get_sha1_hex(path + path_len - the_hash_algo->hexsz, p->hash))
|
|
|
|
hashclr(p->hash);
|
2017-08-19 01:20:24 +03:00
|
|
|
return p;
|
|
|
|
}
|
2017-08-19 01:20:25 +03:00
|
|
|
|
2018-03-23 20:45:18 +03:00
|
|
|
void install_packed_git(struct repository *r, struct packed_git *pack)
|
2017-08-19 01:20:25 +03:00
|
|
|
{
|
|
|
|
if (pack->pack_fd != -1)
|
|
|
|
pack_open_fds++;
|
|
|
|
|
2018-03-23 20:45:18 +03:00
|
|
|
pack->next = r->objects->packed_git;
|
|
|
|
r->objects->packed_git = pack;
|
2019-11-28 01:24:53 +03:00
|
|
|
|
|
|
|
hashmap_entry_init(&pack->packmap_ent, strhash(pack->pack_name));
|
|
|
|
hashmap_add(&r->objects->pack_map, &pack->packmap_ent);
|
2017-08-19 01:20:25 +03:00
|
|
|
}
|
2017-08-19 01:20:26 +03:00
|
|
|
|
|
|
|
void (*report_garbage)(unsigned seen_bits, const char *path);
|
|
|
|
|
|
|
|
static void report_helper(const struct string_list *list,
|
|
|
|
int seen_bits, int first, int last)
|
|
|
|
{
|
|
|
|
if (seen_bits == (PACKDIR_FILE_PACK|PACKDIR_FILE_IDX))
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (; first < last; first++)
|
|
|
|
report_garbage(seen_bits, list->items[first].string);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void report_pack_garbage(struct string_list *list)
|
|
|
|
{
|
|
|
|
int i, baselen = -1, first = 0, seen_bits = 0;
|
|
|
|
|
|
|
|
if (!report_garbage)
|
|
|
|
return;
|
|
|
|
|
|
|
|
string_list_sort(list);
|
|
|
|
|
|
|
|
for (i = 0; i < list->nr; i++) {
|
|
|
|
const char *path = list->items[i].string;
|
|
|
|
if (baselen != -1 &&
|
|
|
|
strncmp(path, list->items[first].string, baselen)) {
|
|
|
|
report_helper(list, seen_bits, first, i);
|
|
|
|
baselen = -1;
|
|
|
|
seen_bits = 0;
|
|
|
|
}
|
|
|
|
if (baselen == -1) {
|
|
|
|
const char *dot = strrchr(path, '.');
|
|
|
|
if (!dot) {
|
|
|
|
report_garbage(PACKDIR_FILE_GARBAGE, path);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
baselen = dot - path + 1;
|
|
|
|
first = i;
|
|
|
|
}
|
|
|
|
if (!strcmp(path + baselen, "pack"))
|
|
|
|
seen_bits |= 1;
|
|
|
|
else if (!strcmp(path + baselen, "idx"))
|
|
|
|
seen_bits |= 2;
|
|
|
|
}
|
|
|
|
report_helper(list, seen_bits, first, list->nr);
|
|
|
|
}
|
|
|
|
|
2018-07-12 22:39:25 +03:00
|
|
|
void for_each_file_in_pack_dir(const char *objdir,
|
|
|
|
each_file_in_pack_dir_fn fn,
|
|
|
|
void *data)
|
2017-08-19 01:20:26 +03:00
|
|
|
{
|
|
|
|
struct strbuf path = STRBUF_INIT;
|
|
|
|
size_t dirnamelen;
|
|
|
|
DIR *dir;
|
|
|
|
struct dirent *de;
|
|
|
|
|
|
|
|
strbuf_addstr(&path, objdir);
|
|
|
|
strbuf_addstr(&path, "/pack");
|
|
|
|
dir = opendir(path.buf);
|
|
|
|
if (!dir) {
|
|
|
|
if (errno != ENOENT)
|
|
|
|
error_errno("unable to open object pack directory: %s",
|
|
|
|
path.buf);
|
|
|
|
strbuf_release(&path);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
strbuf_addch(&path, '/');
|
|
|
|
dirnamelen = path.len;
|
|
|
|
while ((de = readdir(dir)) != NULL) {
|
|
|
|
if (is_dot_or_dotdot(de->d_name))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
strbuf_setlen(&path, dirnamelen);
|
|
|
|
strbuf_addstr(&path, de->d_name);
|
|
|
|
|
2018-07-12 22:39:25 +03:00
|
|
|
fn(path.buf, path.len, de->d_name, data);
|
2017-08-19 01:20:26 +03:00
|
|
|
}
|
2018-07-12 22:39:25 +03:00
|
|
|
|
2017-08-19 01:20:26 +03:00
|
|
|
closedir(dir);
|
|
|
|
strbuf_release(&path);
|
|
|
|
}
|
|
|
|
|
2018-07-12 22:39:25 +03:00
|
|
|
struct prepare_pack_data {
|
|
|
|
struct repository *r;
|
|
|
|
struct string_list *garbage;
|
|
|
|
int local;
|
2018-07-12 22:39:38 +03:00
|
|
|
struct multi_pack_index *m;
|
2018-07-12 22:39:25 +03:00
|
|
|
};
|
|
|
|
|
|
|
|
static void prepare_pack(const char *full_name, size_t full_name_len,
|
|
|
|
const char *file_name, void *_data)
|
|
|
|
{
|
|
|
|
struct prepare_pack_data *data = (struct prepare_pack_data *)_data;
|
|
|
|
struct packed_git *p;
|
|
|
|
size_t base_len = full_name_len;
|
|
|
|
|
2018-08-20 19:51:59 +03:00
|
|
|
if (strip_suffix_mem(full_name, &base_len, ".idx") &&
|
|
|
|
!(data->m && midx_contains_pack(data->m, file_name))) {
|
2019-11-28 01:24:53 +03:00
|
|
|
struct hashmap_entry hent;
|
|
|
|
char *pack_name = xstrfmt("%.*s.pack", (int)base_len, full_name);
|
|
|
|
unsigned int hash = strhash(pack_name);
|
|
|
|
hashmap_entry_init(&hent, hash);
|
2018-07-12 22:39:25 +03:00
|
|
|
|
2019-11-28 01:24:53 +03:00
|
|
|
/* Don't reopen a pack we already have. */
|
|
|
|
if (!hashmap_get(&data->r->objects->pack_map, &hent, pack_name)) {
|
2018-07-12 22:39:25 +03:00
|
|
|
p = add_packed_git(full_name, full_name_len, data->local);
|
|
|
|
if (p)
|
|
|
|
install_packed_git(data->r, p);
|
|
|
|
}
|
2019-11-28 01:24:53 +03:00
|
|
|
free(pack_name);
|
2018-07-12 22:39:25 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
if (!report_garbage)
|
|
|
|
return;
|
|
|
|
|
2018-08-20 19:51:59 +03:00
|
|
|
if (!strcmp(file_name, "multi-pack-index"))
|
|
|
|
return;
|
2018-07-12 22:39:25 +03:00
|
|
|
if (ends_with(file_name, ".idx") ||
|
|
|
|
ends_with(file_name, ".pack") ||
|
|
|
|
ends_with(file_name, ".bitmap") ||
|
|
|
|
ends_with(file_name, ".keep") ||
|
|
|
|
ends_with(file_name, ".promisor"))
|
|
|
|
string_list_append(data->garbage, full_name);
|
|
|
|
else
|
|
|
|
report_garbage(PACKDIR_FILE_GARBAGE, full_name);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void prepare_packed_git_one(struct repository *r, char *objdir, int local)
|
|
|
|
{
|
|
|
|
struct prepare_pack_data data;
|
|
|
|
struct string_list garbage = STRING_LIST_INIT_DUP;
|
|
|
|
|
2018-07-12 22:39:38 +03:00
|
|
|
data.m = r->objects->multi_pack_index;
|
|
|
|
|
|
|
|
/* look for the multi-pack-index for this object directory */
|
|
|
|
while (data.m && strcmp(data.m->object_dir, objdir))
|
|
|
|
data.m = data.m->next;
|
|
|
|
|
2018-07-12 22:39:25 +03:00
|
|
|
data.r = r;
|
|
|
|
data.garbage = &garbage;
|
|
|
|
data.local = local;
|
|
|
|
|
|
|
|
for_each_file_in_pack_dir(objdir, prepare_pack, &data);
|
|
|
|
|
|
|
|
report_pack_garbage(data.garbage);
|
|
|
|
string_list_clear(data.garbage, 0);
|
|
|
|
}
|
|
|
|
|
2018-03-23 20:45:27 +03:00
|
|
|
static void prepare_packed_git(struct repository *r);
|
2017-08-19 01:20:26 +03:00
|
|
|
/*
|
|
|
|
* Give a fast, rough count of the number of objects in the repository. This
|
|
|
|
* ignores loose objects completely. If you have a lot of them, then either
|
|
|
|
* you should repack because your performance will be awful, or they are
|
|
|
|
* all unreachable objects about to be pruned, in which case they're not really
|
|
|
|
* interesting as a measure of repo size in the first place.
|
|
|
|
*/
|
2019-04-06 14:34:23 +03:00
|
|
|
unsigned long repo_approximate_object_count(struct repository *r)
|
2017-08-19 01:20:26 +03:00
|
|
|
{
|
2019-04-06 14:34:23 +03:00
|
|
|
if (!r->objects->approximate_object_count_valid) {
|
2018-03-23 20:21:02 +03:00
|
|
|
unsigned long count;
|
2018-07-12 22:39:37 +03:00
|
|
|
struct multi_pack_index *m;
|
2017-08-19 01:20:26 +03:00
|
|
|
struct packed_git *p;
|
|
|
|
|
2019-04-06 14:34:23 +03:00
|
|
|
prepare_packed_git(r);
|
2017-08-19 01:20:26 +03:00
|
|
|
count = 0;
|
2019-04-06 14:34:23 +03:00
|
|
|
for (m = get_multi_pack_index(r); m; m = m->next)
|
2018-07-12 22:39:37 +03:00
|
|
|
count += m->num_objects;
|
2019-04-06 14:34:23 +03:00
|
|
|
for (p = r->objects->packed_git; p; p = p->next) {
|
2017-08-19 01:20:26 +03:00
|
|
|
if (open_pack_index(p))
|
|
|
|
continue;
|
|
|
|
count += p->num_objects;
|
|
|
|
}
|
2019-04-06 14:34:23 +03:00
|
|
|
r->objects->approximate_object_count = count;
|
packfile: actually set approximate_object_count_valid
The approximate_object_count() function tries to compute the count only
once per process. But ever since it was introduced in 8e3f52d778
(find_unique_abbrev: move logic out of get_short_sha1(), 2016-10-03), we
failed to actually set the "valid" flag, meaning we'd compute it fresh
on every call.
This turns out not to be _too_ bad, because we're only iterating through
the packed_git list, and not making any system calls. But since it may
get called for every abbreviated hash we output, even this can add up if
you have many packs.
Here are before-and-after timings for a new perf test which just asks
rev-list to abbreviate each commit hash (the test repo is linux.git,
with commit-graphs):
Test origin HEAD
----------------------------------------------------------------------------
5303.3: rev-list (1) 28.91(28.46+0.44) 29.03(28.65+0.38) +0.4%
5303.4: abbrev-commit (1) 1.18(1.06+0.11) 1.17(1.02+0.14) -0.8%
5303.7: rev-list (50) 28.95(28.56+0.38) 29.50(29.17+0.32) +1.9%
5303.8: abbrev-commit (50) 3.67(3.56+0.10) 3.57(3.42+0.15) -2.7%
5303.11: rev-list (1000) 30.34(29.89+0.43) 30.82(30.35+0.46) +1.6%
5303.12: abbrev-commit (1000) 86.82(86.52+0.29) 77.82(77.59+0.22) -10.4%
5303.15: load 10,000 packs 0.08(0.02+0.05) 0.08(0.02+0.06) +0.0%
It doesn't help at all when we have 1 pack (5303.4), but we get a 10%
speedup when there are 1000 packs (5303.12). That's a modest speedup for
a case that's already slow and we'd hope to avoid in general (note how
slow it is even after, because we have to look in each of those packs
for abbreviations). But it's a one-line change that clearly matches the
original intent, so it seems worth doing.
The included perf test may also be useful for keeping an eye on any
regressions in the overall abbreviation code.
Reported-by: Rasmus Villemoes <rv@rasmusvillemoes.dk>
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-09-17 19:47:43 +03:00
|
|
|
r->objects->approximate_object_count_valid = 1;
|
2017-08-19 01:20:26 +03:00
|
|
|
}
|
2019-04-06 14:34:23 +03:00
|
|
|
return r->objects->approximate_object_count;
|
2017-08-19 01:20:26 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
static void *get_next_packed_git(const void *p)
|
|
|
|
{
|
|
|
|
return ((const struct packed_git *)p)->next;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void set_next_packed_git(void *p, void *next)
|
|
|
|
{
|
|
|
|
((struct packed_git *)p)->next = next;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int sort_pack(const void *a_, const void *b_)
|
|
|
|
{
|
|
|
|
const struct packed_git *a = a_;
|
|
|
|
const struct packed_git *b = b_;
|
|
|
|
int st;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Local packs tend to contain objects specific to our
|
|
|
|
* variant of the project than remote ones. In addition,
|
|
|
|
* remote ones could be on a network mounted filesystem.
|
|
|
|
* Favor local ones for these reasons.
|
|
|
|
*/
|
|
|
|
st = a->pack_local - b->pack_local;
|
|
|
|
if (st)
|
|
|
|
return -st;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Younger packs tend to contain more recent objects,
|
|
|
|
* and more recent objects tend to get accessed more
|
|
|
|
* often.
|
|
|
|
*/
|
|
|
|
if (a->mtime < b->mtime)
|
|
|
|
return 1;
|
|
|
|
else if (a->mtime == b->mtime)
|
|
|
|
return 0;
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
2018-03-23 20:45:17 +03:00
|
|
|
static void rearrange_packed_git(struct repository *r)
|
2017-08-19 01:20:26 +03:00
|
|
|
{
|
2018-03-23 20:45:17 +03:00
|
|
|
r->objects->packed_git = llist_mergesort(
|
|
|
|
r->objects->packed_git, get_next_packed_git,
|
2018-03-23 20:20:59 +03:00
|
|
|
set_next_packed_git, sort_pack);
|
2017-08-19 01:20:26 +03:00
|
|
|
}
|
|
|
|
|
2018-03-23 20:45:16 +03:00
|
|
|
static void prepare_packed_git_mru(struct repository *r)
|
2017-08-19 01:20:26 +03:00
|
|
|
{
|
|
|
|
struct packed_git *p;
|
|
|
|
|
2018-03-23 20:45:16 +03:00
|
|
|
INIT_LIST_HEAD(&r->objects->packed_git_mru);
|
2018-01-24 02:46:51 +03:00
|
|
|
|
2018-03-23 20:45:16 +03:00
|
|
|
for (p = r->objects->packed_git; p; p = p->next)
|
|
|
|
list_add_tail(&p->mru, &r->objects->packed_git_mru);
|
2017-08-19 01:20:26 +03:00
|
|
|
}
|
|
|
|
|
2018-03-23 20:45:27 +03:00
|
|
|
static void prepare_packed_git(struct repository *r)
|
2017-08-19 01:20:26 +03:00
|
|
|
{
|
2018-11-12 17:48:47 +03:00
|
|
|
struct object_directory *odb;
|
2017-08-19 01:20:26 +03:00
|
|
|
|
2018-03-23 20:45:23 +03:00
|
|
|
if (r->objects->packed_git_initialized)
|
2017-08-19 01:20:26 +03:00
|
|
|
return;
|
sha1-file: use an object_directory for the main object dir
Our handling of alternate object directories is needlessly different
from the main object directory. As a result, many places in the code
basically look like this:
do_something(r->objects->objdir);
for (odb = r->objects->alt_odb_list; odb; odb = odb->next)
do_something(odb->path);
That gets annoying when do_something() is non-trivial, and we've
resorted to gross hacks like creating fake alternates (see
find_short_object_filename()).
Instead, let's give each raw_object_store a unified list of
object_directory structs. The first will be the main store, and
everything after is an alternate. Very few callers even care about the
distinction, and can just loop over the whole list (and those who care
can just treat the first element differently).
A few observations:
- we don't need r->objects->objectdir anymore, and can just
mechanically convert that to r->objects->odb->path
- object_directory's path field needs to become a real pointer rather
than a FLEX_ARRAY, in order to fill it with expand_base_dir()
- we'll call prepare_alt_odb() earlier in many functions (i.e.,
outside of the loop). This may result in us calling it even when our
function would be satisfied looking only at the main odb.
But this doesn't matter in practice. It's not a very expensive
operation in the first place, and in the majority of cases it will
be a noop. We call it already (and cache its results) in
prepare_packed_git(), and we'll generally check packs before loose
objects. So essentially every program is going to call it
immediately once per program.
Arguably we should just prepare_alt_odb() immediately upon setting
up the repository's object directory, which would save us sprinkling
calls throughout the code base (and forgetting to do so has been a
source of subtle bugs in the past). But I've stopped short of that
here, since there are already a lot of other moving parts in this
patch.
- Most call sites just get shorter. The check_and_freshen() functions
are an exception, because they have entry points to handle local and
nonlocal directories separately.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-11-12 17:50:39 +03:00
|
|
|
|
2018-03-23 20:45:23 +03:00
|
|
|
prepare_alt_odb(r);
|
sha1-file: use an object_directory for the main object dir
Our handling of alternate object directories is needlessly different
from the main object directory. As a result, many places in the code
basically look like this:
do_something(r->objects->objdir);
for (odb = r->objects->alt_odb_list; odb; odb = odb->next)
do_something(odb->path);
That gets annoying when do_something() is non-trivial, and we've
resorted to gross hacks like creating fake alternates (see
find_short_object_filename()).
Instead, let's give each raw_object_store a unified list of
object_directory structs. The first will be the main store, and
everything after is an alternate. Very few callers even care about the
distinction, and can just loop over the whole list (and those who care
can just treat the first element differently).
A few observations:
- we don't need r->objects->objectdir anymore, and can just
mechanically convert that to r->objects->odb->path
- object_directory's path field needs to become a real pointer rather
than a FLEX_ARRAY, in order to fill it with expand_base_dir()
- we'll call prepare_alt_odb() earlier in many functions (i.e.,
outside of the loop). This may result in us calling it even when our
function would be satisfied looking only at the main odb.
But this doesn't matter in practice. It's not a very expensive
operation in the first place, and in the majority of cases it will
be a noop. We call it already (and cache its results) in
prepare_packed_git(), and we'll generally check packs before loose
objects. So essentially every program is going to call it
immediately once per program.
Arguably we should just prepare_alt_odb() immediately upon setting
up the repository's object directory, which would save us sprinkling
calls throughout the code base (and forgetting to do so has been a
source of subtle bugs in the past). But I've stopped short of that
here, since there are already a lot of other moving parts in this
patch.
- Most call sites just get shorter. The check_and_freshen() functions
are an exception, because they have entry points to handle local and
nonlocal directories separately.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-11-12 17:50:39 +03:00
|
|
|
for (odb = r->objects->odb; odb; odb = odb->next) {
|
|
|
|
int local = (odb == r->objects->odb);
|
|
|
|
prepare_multi_pack_index_one(r, odb->path, local);
|
|
|
|
prepare_packed_git_one(r, odb->path, local);
|
2018-07-12 22:39:33 +03:00
|
|
|
}
|
2018-03-23 20:45:23 +03:00
|
|
|
rearrange_packed_git(r);
|
2018-08-20 19:52:02 +03:00
|
|
|
|
2018-03-23 20:45:23 +03:00
|
|
|
prepare_packed_git_mru(r);
|
|
|
|
r->objects->packed_git_initialized = 1;
|
2017-08-19 01:20:26 +03:00
|
|
|
}
|
|
|
|
|
2018-03-23 20:45:24 +03:00
|
|
|
void reprepare_packed_git(struct repository *r)
|
2017-08-19 01:20:26 +03:00
|
|
|
{
|
2018-11-12 17:50:56 +03:00
|
|
|
struct object_directory *odb;
|
|
|
|
|
2020-01-16 05:39:57 +03:00
|
|
|
obj_read_lock();
|
2019-01-06 19:45:39 +03:00
|
|
|
for (odb = r->objects->odb; odb; odb = odb->next)
|
|
|
|
odb_clear_loose_cache(odb);
|
2018-11-12 17:50:56 +03:00
|
|
|
|
2018-03-23 20:45:24 +03:00
|
|
|
r->objects->approximate_object_count_valid = 0;
|
|
|
|
r->objects->packed_git_initialized = 0;
|
|
|
|
prepare_packed_git(r);
|
2020-01-16 05:39:57 +03:00
|
|
|
obj_read_unlock();
|
2017-08-19 01:20:26 +03:00
|
|
|
}
|
2017-08-19 01:20:27 +03:00
|
|
|
|
2018-03-23 20:20:59 +03:00
|
|
|
struct packed_git *get_packed_git(struct repository *r)
|
|
|
|
{
|
2018-03-23 20:45:27 +03:00
|
|
|
prepare_packed_git(r);
|
2018-03-23 20:20:59 +03:00
|
|
|
return r->objects->packed_git;
|
|
|
|
}
|
|
|
|
|
2018-07-12 22:39:35 +03:00
|
|
|
struct multi_pack_index *get_multi_pack_index(struct repository *r)
|
|
|
|
{
|
|
|
|
prepare_packed_git(r);
|
|
|
|
return r->objects->multi_pack_index;
|
|
|
|
}
|
|
|
|
|
midx: traverse the local MIDX first
When a repository has an alternate object directory configured, callers
can traverse through each alternate's MIDX by walking the '->next'
pointer.
But, when 'prepare_multi_pack_index_one()' loads multiple MIDXs, it
places the new ones at the front of this pointer chain, not at the end.
This can be confusing for callers such as 'git repack -ad', causing test
failures like in t7700.6 with 'GIT_TEST_MULTI_PACK_INDEX=1'.
The occurs when dropping a pack known to the local MIDX with alternates
configured that have their own MIDX. Since the alternate's MIDX is
returned via 'get_multi_pack_index()', 'midx_contains_pack()' returns
true (which is correct, since it traverses through the '->next' pointer
to find the MIDX in the chain that does contain the requested object).
But, we call 'clear_midx_file()' on 'the_repository', which drops the
MIDX at the path of the first MIDX in the chain, which (in the case of
t7700.6 is the one in the alternate).
This patch addresses that by:
- placing the local MIDX first in the chain when calling
'prepare_multi_pack_index_one()', and
- introducing a new 'get_local_multi_pack_index()', which explicitly
returns the repository-local MIDX, if any.
Don't impose an additional order on the MIDX's '->next' pointer beyond
that the first item in the chain must be local if one exists so that we
avoid a quadratic insertion.
Likewise, use 'get_local_multi_pack_index()' in
'remove_redundant_pack()' to fix the formerly broken t7700.6 when run
with 'GIT_TEST_MULTI_PACK_INDEX=1'.
Finally, note that the MIDX ordering invariant is only preserved by the
insertion order in 'prepare_packed_git()', which traverses through the
ODB's '->next' pointer, meaning we visit the local object store first.
This fragility makes this an undesirable long-term solution if more
callers are added, but it is acceptable for now since this is the only
caller.
Helped-by: Jeff King <peff@peff.net>
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Taylor Blau <me@ttaylorr.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-08-28 23:22:13 +03:00
|
|
|
struct multi_pack_index *get_local_multi_pack_index(struct repository *r)
|
|
|
|
{
|
|
|
|
struct multi_pack_index *m = get_multi_pack_index(r);
|
|
|
|
|
|
|
|
/* no need to iterate; we always put the local one first (if any) */
|
|
|
|
if (m && m->local)
|
|
|
|
return m;
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2018-08-20 19:52:02 +03:00
|
|
|
struct packed_git *get_all_packs(struct repository *r)
|
|
|
|
{
|
midx: add packs to packed_git linked list
The multi-pack-index allows searching for objects across multiple
packs using one object list. The original design gains many of
these performance benefits by keeping the packs in the
multi-pack-index out of the packed_git list.
Unfortunately, this has one major drawback. If the multi-pack-index
covers thousands of packs, and a command loads many of those packs,
then we can hit the limit for open file descriptors. The
close_one_pack() method is used to limit this resource, but it
only looks at the packed_git list, and uses an LRU cache to prevent
thrashing.
Instead of complicating this close_one_pack() logic to include
direct references to the multi-pack-index, simply add the packs
opened by the multi-pack-index to the packed_git list. This
immediately solves the file-descriptor limit problem, but requires
some extra steps to avoid performance issues or other problems:
1. Create a multi_pack_index bit in the packed_git struct that is
one if and only if the pack was loaded from a multi-pack-index.
2. Skip packs with the multi_pack_index bit when doing object
lookups and abbreviations. These algorithms already check the
multi-pack-index before the packed_git struct. This has a very
small performance hit, as we need to walk more packed_git
structs. This is acceptable, since these operations run binary
search on the other packs, so this walk-and-ignore logic is
very fast by comparison.
3. When closing a multi-pack-index file, do not close its packs,
as those packs will be closed using close_all_packs(). In some
cases, such as 'git repack', we run 'close_midx()' without also
closing the packs, so we need to un-set the multi_pack_index bit
in those packs. This is necessary, and caught by running
t6501-freshen-objects.sh with GIT_TEST_MULTI_PACK_INDEX=1.
To manually test this change, I inserted trace2 logging into
close_pack_fd() and set pack_max_fds to 10, then ran 'git rev-list
--all --objects' on a copy of the Git repo with 300+ pack-files and
a multi-pack-index. The logs verified the packs are closed as
we read them beyond the file descriptor limit.
Signed-off-by: Derrick Stolee <dstolee@microsoft.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2019-04-29 19:18:56 +03:00
|
|
|
struct multi_pack_index *m;
|
2018-08-20 19:52:02 +03:00
|
|
|
|
midx: add packs to packed_git linked list
The multi-pack-index allows searching for objects across multiple
packs using one object list. The original design gains many of
these performance benefits by keeping the packs in the
multi-pack-index out of the packed_git list.
Unfortunately, this has one major drawback. If the multi-pack-index
covers thousands of packs, and a command loads many of those packs,
then we can hit the limit for open file descriptors. The
close_one_pack() method is used to limit this resource, but it
only looks at the packed_git list, and uses an LRU cache to prevent
thrashing.
Instead of complicating this close_one_pack() logic to include
direct references to the multi-pack-index, simply add the packs
opened by the multi-pack-index to the packed_git list. This
immediately solves the file-descriptor limit problem, but requires
some extra steps to avoid performance issues or other problems:
1. Create a multi_pack_index bit in the packed_git struct that is
one if and only if the pack was loaded from a multi-pack-index.
2. Skip packs with the multi_pack_index bit when doing object
lookups and abbreviations. These algorithms already check the
multi-pack-index before the packed_git struct. This has a very
small performance hit, as we need to walk more packed_git
structs. This is acceptable, since these operations run binary
search on the other packs, so this walk-and-ignore logic is
very fast by comparison.
3. When closing a multi-pack-index file, do not close its packs,
as those packs will be closed using close_all_packs(). In some
cases, such as 'git repack', we run 'close_midx()' without also
closing the packs, so we need to un-set the multi_pack_index bit
in those packs. This is necessary, and caught by running
t6501-freshen-objects.sh with GIT_TEST_MULTI_PACK_INDEX=1.
To manually test this change, I inserted trace2 logging into
close_pack_fd() and set pack_max_fds to 10, then ran 'git rev-list
--all --objects' on a copy of the Git repo with 300+ pack-files and
a multi-pack-index. The logs verified the packs are closed as
we read them beyond the file descriptor limit.
Signed-off-by: Derrick Stolee <dstolee@microsoft.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2019-04-29 19:18:56 +03:00
|
|
|
prepare_packed_git(r);
|
|
|
|
for (m = r->objects->multi_pack_index; m; m = m->next) {
|
|
|
|
uint32_t i;
|
|
|
|
for (i = 0; i < m->num_packs; i++)
|
|
|
|
prepare_midx_pack(r, m, i);
|
2018-08-20 19:52:02 +03:00
|
|
|
}
|
|
|
|
|
midx: add packs to packed_git linked list
The multi-pack-index allows searching for objects across multiple
packs using one object list. The original design gains many of
these performance benefits by keeping the packs in the
multi-pack-index out of the packed_git list.
Unfortunately, this has one major drawback. If the multi-pack-index
covers thousands of packs, and a command loads many of those packs,
then we can hit the limit for open file descriptors. The
close_one_pack() method is used to limit this resource, but it
only looks at the packed_git list, and uses an LRU cache to prevent
thrashing.
Instead of complicating this close_one_pack() logic to include
direct references to the multi-pack-index, simply add the packs
opened by the multi-pack-index to the packed_git list. This
immediately solves the file-descriptor limit problem, but requires
some extra steps to avoid performance issues or other problems:
1. Create a multi_pack_index bit in the packed_git struct that is
one if and only if the pack was loaded from a multi-pack-index.
2. Skip packs with the multi_pack_index bit when doing object
lookups and abbreviations. These algorithms already check the
multi-pack-index before the packed_git struct. This has a very
small performance hit, as we need to walk more packed_git
structs. This is acceptable, since these operations run binary
search on the other packs, so this walk-and-ignore logic is
very fast by comparison.
3. When closing a multi-pack-index file, do not close its packs,
as those packs will be closed using close_all_packs(). In some
cases, such as 'git repack', we run 'close_midx()' without also
closing the packs, so we need to un-set the multi_pack_index bit
in those packs. This is necessary, and caught by running
t6501-freshen-objects.sh with GIT_TEST_MULTI_PACK_INDEX=1.
To manually test this change, I inserted trace2 logging into
close_pack_fd() and set pack_max_fds to 10, then ran 'git rev-list
--all --objects' on a copy of the Git repo with 300+ pack-files and
a multi-pack-index. The logs verified the packs are closed as
we read them beyond the file descriptor limit.
Signed-off-by: Derrick Stolee <dstolee@microsoft.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2019-04-29 19:18:56 +03:00
|
|
|
return r->objects->packed_git;
|
2018-08-20 19:52:02 +03:00
|
|
|
}
|
|
|
|
|
2018-03-23 20:20:59 +03:00
|
|
|
struct list_head *get_packed_git_mru(struct repository *r)
|
|
|
|
{
|
2018-03-23 20:45:27 +03:00
|
|
|
prepare_packed_git(r);
|
2018-03-23 20:20:59 +03:00
|
|
|
return &r->objects->packed_git_mru;
|
|
|
|
}
|
|
|
|
|
2017-08-19 01:20:27 +03:00
|
|
|
unsigned long unpack_object_header_buffer(const unsigned char *buf,
|
|
|
|
unsigned long len, enum object_type *type, unsigned long *sizep)
|
|
|
|
{
|
|
|
|
unsigned shift;
|
|
|
|
unsigned long size, c;
|
|
|
|
unsigned long used = 0;
|
|
|
|
|
|
|
|
c = buf[used++];
|
|
|
|
*type = (c >> 4) & 7;
|
|
|
|
size = c & 15;
|
|
|
|
shift = 4;
|
|
|
|
while (c & 0x80) {
|
|
|
|
if (len <= used || bitsizeof(long) <= shift) {
|
|
|
|
error("bad object header");
|
|
|
|
size = used = 0;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
c = buf[used++];
|
|
|
|
size += (c & 0x7f) << shift;
|
|
|
|
shift += 7;
|
|
|
|
}
|
|
|
|
*sizep = size;
|
|
|
|
return used;
|
|
|
|
}
|
2017-08-19 01:20:28 +03:00
|
|
|
|
|
|
|
unsigned long get_size_from_delta(struct packed_git *p,
|
|
|
|
struct pack_window **w_curs,
|
|
|
|
off_t curpos)
|
|
|
|
{
|
|
|
|
const unsigned char *data;
|
|
|
|
unsigned char delta_head[20], *in;
|
|
|
|
git_zstream stream;
|
|
|
|
int st;
|
|
|
|
|
|
|
|
memset(&stream, 0, sizeof(stream));
|
|
|
|
stream.next_out = delta_head;
|
|
|
|
stream.avail_out = sizeof(delta_head);
|
|
|
|
|
|
|
|
git_inflate_init(&stream);
|
|
|
|
do {
|
|
|
|
in = use_pack(p, w_curs, curpos, &stream.avail_in);
|
|
|
|
stream.next_in = in;
|
object-store: allow threaded access to object reading
Allow object reading to be performed by multiple threads protecting it
with an internal lock, the obj_read_mutex. The lock usage can be toggled
with enable_obj_read_lock() and disable_obj_read_lock(). Currently, the
functions which can be safely called in parallel are:
read_object_file_extended(), repo_read_object_file(),
read_object_file(), read_object_with_reference(), read_object(),
oid_object_info() and oid_object_info_extended(). It's also possible
to use obj_read_lock() and obj_read_unlock() to protect other sections
that cannot execute in parallel with object reading.
Probably there are many spots in the functions listed above that could
be executed unlocked (and thus, in parallel). But, for now, we are most
interested in allowing parallel access to zlib inflation. This is one of
the sections where object reading spends most of the time in (e.g. up to
one-third of git-grep's execution time in the chromium repo corresponds
to inflation) and it's already thread-safe. So, to take advantage of
that, the obj_read_mutex is released when calling git_inflate() and
re-acquired right after, for every calling spot in
oid_object_info_extended()'s call chain. We may refine this lock to also
exploit other possible parallel spots in the future, but for now,
threaded zlib inflation should already give great speedups for threaded
object reading callers.
Note that add_delta_base_cache() was also modified to skip adding
already present entries to the cache. This wasn't possible before, but
it would be now, with the parallel inflation. Take for example the
following situation, where two threads - A and B - are executing the
code at unpack_entry():
1. Thread A is performing the decompression of a base O (which is not
yet in the cache) at PHASE II. Thread B is simultaneously trying to
unpack O, but just starting at PHASE I.
2. Since O is not yet in the cache, B will go to PHASE II to also
perform the decompression.
3. When they finish decompressing, one of them will get the object
reading mutex and go to PHASE III while the other waits for the
mutex. Let’s say A got the mutex first.
4. Thread A will add O to the cache, go throughout the rest of PHASE III
and return.
5. Thread B gets the mutex, also add O to the cache (if the check wasn't
there) and returns.
Finally, it is also important to highlight that the object reading lock
can only ensure thread-safety in the mentioned functions thanks to two
complementary mechanisms: the use of 'struct raw_object_store's
replace_mutex, which guards sections in the object reading machinery
that would otherwise be thread-unsafe; and the 'struct pack_window's
inuse_cnt, which protects window reading operations (such as the one
performed during the inflation of a packed object), allowing them to
execute without the acquisition of the obj_read_mutex.
Signed-off-by: Matheus Tavares <matheus.bernardino@usp.br>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-01-16 05:39:53 +03:00
|
|
|
/*
|
|
|
|
* Note: the window section returned by use_pack() must be
|
|
|
|
* available throughout git_inflate()'s unlocked execution. To
|
|
|
|
* ensure no other thread will modify the window in the
|
|
|
|
* meantime, we rely on the packed_window.inuse_cnt. This
|
|
|
|
* counter is incremented before window reading and checked
|
|
|
|
* before window disposal.
|
|
|
|
*
|
|
|
|
* Other worrying sections could be the call to close_pack_fd(),
|
|
|
|
* which can close packs even with in-use windows, and to
|
|
|
|
* reprepare_packed_git(). Regarding the former, mmap doc says:
|
|
|
|
* "closing the file descriptor does not unmap the region". And
|
|
|
|
* for the latter, it won't re-open already available packs.
|
|
|
|
*/
|
|
|
|
obj_read_unlock();
|
2017-08-19 01:20:28 +03:00
|
|
|
st = git_inflate(&stream, Z_FINISH);
|
object-store: allow threaded access to object reading
Allow object reading to be performed by multiple threads protecting it
with an internal lock, the obj_read_mutex. The lock usage can be toggled
with enable_obj_read_lock() and disable_obj_read_lock(). Currently, the
functions which can be safely called in parallel are:
read_object_file_extended(), repo_read_object_file(),
read_object_file(), read_object_with_reference(), read_object(),
oid_object_info() and oid_object_info_extended(). It's also possible
to use obj_read_lock() and obj_read_unlock() to protect other sections
that cannot execute in parallel with object reading.
Probably there are many spots in the functions listed above that could
be executed unlocked (and thus, in parallel). But, for now, we are most
interested in allowing parallel access to zlib inflation. This is one of
the sections where object reading spends most of the time in (e.g. up to
one-third of git-grep's execution time in the chromium repo corresponds
to inflation) and it's already thread-safe. So, to take advantage of
that, the obj_read_mutex is released when calling git_inflate() and
re-acquired right after, for every calling spot in
oid_object_info_extended()'s call chain. We may refine this lock to also
exploit other possible parallel spots in the future, but for now,
threaded zlib inflation should already give great speedups for threaded
object reading callers.
Note that add_delta_base_cache() was also modified to skip adding
already present entries to the cache. This wasn't possible before, but
it would be now, with the parallel inflation. Take for example the
following situation, where two threads - A and B - are executing the
code at unpack_entry():
1. Thread A is performing the decompression of a base O (which is not
yet in the cache) at PHASE II. Thread B is simultaneously trying to
unpack O, but just starting at PHASE I.
2. Since O is not yet in the cache, B will go to PHASE II to also
perform the decompression.
3. When they finish decompressing, one of them will get the object
reading mutex and go to PHASE III while the other waits for the
mutex. Let’s say A got the mutex first.
4. Thread A will add O to the cache, go throughout the rest of PHASE III
and return.
5. Thread B gets the mutex, also add O to the cache (if the check wasn't
there) and returns.
Finally, it is also important to highlight that the object reading lock
can only ensure thread-safety in the mentioned functions thanks to two
complementary mechanisms: the use of 'struct raw_object_store's
replace_mutex, which guards sections in the object reading machinery
that would otherwise be thread-unsafe; and the 'struct pack_window's
inuse_cnt, which protects window reading operations (such as the one
performed during the inflation of a packed object), allowing them to
execute without the acquisition of the obj_read_mutex.
Signed-off-by: Matheus Tavares <matheus.bernardino@usp.br>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-01-16 05:39:53 +03:00
|
|
|
obj_read_lock();
|
2017-08-19 01:20:28 +03:00
|
|
|
curpos += stream.next_in - in;
|
|
|
|
} while ((st == Z_OK || st == Z_BUF_ERROR) &&
|
|
|
|
stream.total_out < sizeof(delta_head));
|
|
|
|
git_inflate_end(&stream);
|
|
|
|
if ((st != Z_STREAM_END) && stream.total_out != sizeof(delta_head)) {
|
|
|
|
error("delta data unpack-initial failed");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Examine the initial part of the delta to figure out
|
|
|
|
* the result size.
|
|
|
|
*/
|
|
|
|
data = delta_head;
|
|
|
|
|
|
|
|
/* ignore base size */
|
|
|
|
get_delta_hdr_size(&data, delta_head+sizeof(delta_head));
|
|
|
|
|
|
|
|
/* Read the result size */
|
|
|
|
return get_delta_hdr_size(&data, delta_head+sizeof(delta_head));
|
|
|
|
}
|
2017-08-19 01:20:29 +03:00
|
|
|
|
|
|
|
int unpack_object_header(struct packed_git *p,
|
|
|
|
struct pack_window **w_curs,
|
|
|
|
off_t *curpos,
|
|
|
|
unsigned long *sizep)
|
|
|
|
{
|
|
|
|
unsigned char *base;
|
|
|
|
unsigned long left;
|
|
|
|
unsigned long used;
|
|
|
|
enum object_type type;
|
|
|
|
|
|
|
|
/* use_pack() assures us we have [base, base + 20) available
|
|
|
|
* as a range that we can look at. (Its actually the hash
|
|
|
|
* size that is assured.) With our object header encoding
|
|
|
|
* the maximum deflated object size is 2^137, which is just
|
|
|
|
* insane, so we know won't exceed what we have been given.
|
|
|
|
*/
|
|
|
|
base = use_pack(p, w_curs, *curpos, &left);
|
|
|
|
used = unpack_object_header_buffer(base, left, &type, sizep);
|
|
|
|
if (!used) {
|
|
|
|
type = OBJ_BAD;
|
|
|
|
} else
|
|
|
|
*curpos += used;
|
|
|
|
|
|
|
|
return type;
|
|
|
|
}
|
2017-08-19 01:20:30 +03:00
|
|
|
|
|
|
|
void mark_bad_packed_object(struct packed_git *p, const unsigned char *sha1)
|
|
|
|
{
|
|
|
|
unsigned i;
|
2018-10-15 03:01:54 +03:00
|
|
|
const unsigned hashsz = the_hash_algo->rawsz;
|
2017-08-19 01:20:30 +03:00
|
|
|
for (i = 0; i < p->num_bad_objects; i++)
|
2018-10-15 03:01:54 +03:00
|
|
|
if (hasheq(sha1, p->bad_object_sha1 + hashsz * i))
|
2017-08-19 01:20:30 +03:00
|
|
|
return;
|
|
|
|
p->bad_object_sha1 = xrealloc(p->bad_object_sha1,
|
|
|
|
st_mult(GIT_MAX_RAWSZ,
|
|
|
|
st_add(p->num_bad_objects, 1)));
|
2018-10-15 03:01:54 +03:00
|
|
|
hashcpy(p->bad_object_sha1 + hashsz * p->num_bad_objects, sha1);
|
2017-08-19 01:20:30 +03:00
|
|
|
p->num_bad_objects++;
|
|
|
|
}
|
|
|
|
|
2018-10-17 02:35:33 +03:00
|
|
|
const struct packed_git *has_packed_and_bad(struct repository *r,
|
|
|
|
const unsigned char *sha1)
|
2017-08-19 01:20:30 +03:00
|
|
|
{
|
|
|
|
struct packed_git *p;
|
|
|
|
unsigned i;
|
|
|
|
|
2018-10-17 02:35:33 +03:00
|
|
|
for (p = r->objects->packed_git; p; p = p->next)
|
2017-08-19 01:20:30 +03:00
|
|
|
for (i = 0; i < p->num_bad_objects; i++)
|
2018-08-29 00:22:44 +03:00
|
|
|
if (hasheq(sha1,
|
|
|
|
p->bad_object_sha1 + the_hash_algo->rawsz * i))
|
2017-08-19 01:20:30 +03:00
|
|
|
return p;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2019-09-13 16:02:18 +03:00
|
|
|
off_t get_delta_base(struct packed_git *p,
|
|
|
|
struct pack_window **w_curs,
|
|
|
|
off_t *curpos,
|
|
|
|
enum object_type type,
|
|
|
|
off_t delta_obj_offset)
|
2017-08-19 01:20:30 +03:00
|
|
|
{
|
|
|
|
unsigned char *base_info = use_pack(p, w_curs, *curpos, NULL);
|
|
|
|
off_t base_offset;
|
|
|
|
|
|
|
|
/* use_pack() assured us we have [base_info, base_info + 20)
|
|
|
|
* as a range that we can look at without walking off the
|
|
|
|
* end of the mapped window. Its actually the hash size
|
|
|
|
* that is assured. An OFS_DELTA longer than the hash size
|
|
|
|
* is stupid, as then a REF_DELTA would be smaller to store.
|
|
|
|
*/
|
|
|
|
if (type == OBJ_OFS_DELTA) {
|
|
|
|
unsigned used = 0;
|
|
|
|
unsigned char c = base_info[used++];
|
|
|
|
base_offset = c & 127;
|
|
|
|
while (c & 128) {
|
|
|
|
base_offset += 1;
|
|
|
|
if (!base_offset || MSB(base_offset, 7))
|
|
|
|
return 0; /* overflow */
|
|
|
|
c = base_info[used++];
|
|
|
|
base_offset = (base_offset << 7) + (c & 127);
|
|
|
|
}
|
|
|
|
base_offset = delta_obj_offset - base_offset;
|
|
|
|
if (base_offset <= 0 || base_offset >= delta_obj_offset)
|
|
|
|
return 0; /* out of bound */
|
|
|
|
*curpos += used;
|
|
|
|
} else if (type == OBJ_REF_DELTA) {
|
|
|
|
/* The base entry _must_ be in the same pack */
|
|
|
|
base_offset = find_pack_entry_one(base_info, p);
|
2018-05-02 03:25:36 +03:00
|
|
|
*curpos += the_hash_algo->rawsz;
|
2017-08-19 01:20:30 +03:00
|
|
|
} else
|
|
|
|
die("I am totally screwed");
|
|
|
|
return base_offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Like get_delta_base above, but we return the sha1 instead of the pack
|
|
|
|
* offset. This means it is cheaper for REF deltas (we do not have to do
|
|
|
|
* the final object lookup), but more expensive for OFS deltas (we
|
|
|
|
* have to load the revidx to convert the offset back into a sha1).
|
|
|
|
*/
|
2020-02-24 07:37:31 +03:00
|
|
|
static int get_delta_base_oid(struct packed_git *p,
|
|
|
|
struct pack_window **w_curs,
|
|
|
|
off_t curpos,
|
|
|
|
struct object_id *oid,
|
|
|
|
enum object_type type,
|
|
|
|
off_t delta_obj_offset)
|
2017-08-19 01:20:30 +03:00
|
|
|
{
|
|
|
|
if (type == OBJ_REF_DELTA) {
|
|
|
|
unsigned char *base = use_pack(p, w_curs, curpos, NULL);
|
2020-02-24 07:37:31 +03:00
|
|
|
oidread(oid, base);
|
|
|
|
return 0;
|
2017-08-19 01:20:30 +03:00
|
|
|
} else if (type == OBJ_OFS_DELTA) {
|
|
|
|
struct revindex_entry *revidx;
|
|
|
|
off_t base_offset = get_delta_base(p, w_curs, &curpos,
|
|
|
|
type, delta_obj_offset);
|
|
|
|
|
|
|
|
if (!base_offset)
|
2020-02-24 07:37:31 +03:00
|
|
|
return -1;
|
2017-08-19 01:20:30 +03:00
|
|
|
|
|
|
|
revidx = find_pack_revindex(p, base_offset);
|
|
|
|
if (!revidx)
|
2020-02-24 07:37:31 +03:00
|
|
|
return -1;
|
2017-08-19 01:20:30 +03:00
|
|
|
|
2020-02-24 07:37:31 +03:00
|
|
|
return nth_packed_object_id(oid, p, revidx->nr);
|
2017-08-19 01:20:30 +03:00
|
|
|
} else
|
2020-02-24 07:37:31 +03:00
|
|
|
return -1;
|
2017-08-19 01:20:30 +03:00
|
|
|
}
|
|
|
|
|
2018-04-25 21:21:06 +03:00
|
|
|
static int retry_bad_packed_offset(struct repository *r,
|
|
|
|
struct packed_git *p,
|
|
|
|
off_t obj_offset)
|
2017-08-19 01:20:30 +03:00
|
|
|
{
|
|
|
|
int type;
|
|
|
|
struct revindex_entry *revidx;
|
2018-03-12 05:27:43 +03:00
|
|
|
struct object_id oid;
|
2017-08-19 01:20:30 +03:00
|
|
|
revidx = find_pack_revindex(p, obj_offset);
|
|
|
|
if (!revidx)
|
|
|
|
return OBJ_BAD;
|
2020-02-24 07:27:36 +03:00
|
|
|
nth_packed_object_id(&oid, p, revidx->nr);
|
2018-03-12 05:27:43 +03:00
|
|
|
mark_bad_packed_object(p, oid.hash);
|
2018-04-25 21:21:06 +03:00
|
|
|
type = oid_object_info(r, &oid, NULL);
|
2017-08-19 01:20:30 +03:00
|
|
|
if (type <= OBJ_NONE)
|
|
|
|
return OBJ_BAD;
|
|
|
|
return type;
|
|
|
|
}
|
|
|
|
|
|
|
|
#define POI_STACK_PREALLOC 64
|
|
|
|
|
2018-04-25 21:21:06 +03:00
|
|
|
static enum object_type packed_to_object_type(struct repository *r,
|
|
|
|
struct packed_git *p,
|
2017-08-19 01:20:30 +03:00
|
|
|
off_t obj_offset,
|
|
|
|
enum object_type type,
|
|
|
|
struct pack_window **w_curs,
|
|
|
|
off_t curpos)
|
|
|
|
{
|
|
|
|
off_t small_poi_stack[POI_STACK_PREALLOC];
|
|
|
|
off_t *poi_stack = small_poi_stack;
|
|
|
|
int poi_stack_nr = 0, poi_stack_alloc = POI_STACK_PREALLOC;
|
|
|
|
|
|
|
|
while (type == OBJ_OFS_DELTA || type == OBJ_REF_DELTA) {
|
|
|
|
off_t base_offset;
|
|
|
|
unsigned long size;
|
|
|
|
/* Push the object we're going to leave behind */
|
|
|
|
if (poi_stack_nr >= poi_stack_alloc && poi_stack == small_poi_stack) {
|
|
|
|
poi_stack_alloc = alloc_nr(poi_stack_nr);
|
|
|
|
ALLOC_ARRAY(poi_stack, poi_stack_alloc);
|
2019-06-15 21:36:35 +03:00
|
|
|
COPY_ARRAY(poi_stack, small_poi_stack, poi_stack_nr);
|
2017-08-19 01:20:30 +03:00
|
|
|
} else {
|
|
|
|
ALLOC_GROW(poi_stack, poi_stack_nr+1, poi_stack_alloc);
|
|
|
|
}
|
|
|
|
poi_stack[poi_stack_nr++] = obj_offset;
|
|
|
|
/* If parsing the base offset fails, just unwind */
|
|
|
|
base_offset = get_delta_base(p, w_curs, &curpos, type, obj_offset);
|
|
|
|
if (!base_offset)
|
|
|
|
goto unwind;
|
|
|
|
curpos = obj_offset = base_offset;
|
|
|
|
type = unpack_object_header(p, w_curs, &curpos, &size);
|
|
|
|
if (type <= OBJ_NONE) {
|
|
|
|
/* If getting the base itself fails, we first
|
|
|
|
* retry the base, otherwise unwind */
|
2018-04-25 21:21:06 +03:00
|
|
|
type = retry_bad_packed_offset(r, p, base_offset);
|
2017-08-19 01:20:30 +03:00
|
|
|
if (type > OBJ_NONE)
|
|
|
|
goto out;
|
|
|
|
goto unwind;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
switch (type) {
|
|
|
|
case OBJ_BAD:
|
|
|
|
case OBJ_COMMIT:
|
|
|
|
case OBJ_TREE:
|
|
|
|
case OBJ_BLOB:
|
|
|
|
case OBJ_TAG:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
error("unknown object type %i at offset %"PRIuMAX" in %s",
|
|
|
|
type, (uintmax_t)obj_offset, p->pack_name);
|
|
|
|
type = OBJ_BAD;
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
if (poi_stack != small_poi_stack)
|
|
|
|
free(poi_stack);
|
|
|
|
return type;
|
|
|
|
|
|
|
|
unwind:
|
|
|
|
while (poi_stack_nr) {
|
|
|
|
obj_offset = poi_stack[--poi_stack_nr];
|
2018-04-25 21:21:06 +03:00
|
|
|
type = retry_bad_packed_offset(r, p, obj_offset);
|
2017-08-19 01:20:30 +03:00
|
|
|
if (type > OBJ_NONE)
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
type = OBJ_BAD;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct hashmap delta_base_cache;
|
|
|
|
static size_t delta_base_cached;
|
|
|
|
|
|
|
|
static LIST_HEAD(delta_base_cache_lru);
|
|
|
|
|
|
|
|
struct delta_base_cache_key {
|
|
|
|
struct packed_git *p;
|
|
|
|
off_t base_offset;
|
|
|
|
};
|
|
|
|
|
|
|
|
struct delta_base_cache_entry {
|
2019-10-07 02:30:26 +03:00
|
|
|
struct hashmap_entry ent;
|
2017-08-19 01:20:30 +03:00
|
|
|
struct delta_base_cache_key key;
|
|
|
|
struct list_head lru;
|
|
|
|
void *data;
|
|
|
|
unsigned long size;
|
|
|
|
enum object_type type;
|
|
|
|
};
|
|
|
|
|
|
|
|
static unsigned int pack_entry_hash(struct packed_git *p, off_t base_offset)
|
|
|
|
{
|
|
|
|
unsigned int hash;
|
|
|
|
|
|
|
|
hash = (unsigned int)(intptr_t)p + (unsigned int)base_offset;
|
|
|
|
hash += (hash >> 8) + (hash >> 16);
|
|
|
|
return hash;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct delta_base_cache_entry *
|
|
|
|
get_delta_base_cache_entry(struct packed_git *p, off_t base_offset)
|
|
|
|
{
|
2019-10-07 02:30:36 +03:00
|
|
|
struct hashmap_entry entry, *e;
|
2017-08-19 01:20:30 +03:00
|
|
|
struct delta_base_cache_key key;
|
|
|
|
|
|
|
|
if (!delta_base_cache.cmpfn)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
hashmap_entry_init(&entry, pack_entry_hash(p, base_offset));
|
|
|
|
key.p = p;
|
|
|
|
key.base_offset = base_offset;
|
2019-10-07 02:30:36 +03:00
|
|
|
e = hashmap_get(&delta_base_cache, &entry, &key);
|
|
|
|
return e ? container_of(e, struct delta_base_cache_entry, ent) : NULL;
|
2017-08-19 01:20:30 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
static int delta_base_cache_key_eq(const struct delta_base_cache_key *a,
|
|
|
|
const struct delta_base_cache_key *b)
|
|
|
|
{
|
|
|
|
return a->p == b->p && a->base_offset == b->base_offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int delta_base_cache_hash_cmp(const void *unused_cmp_data,
|
2019-10-07 02:30:37 +03:00
|
|
|
const struct hashmap_entry *va,
|
|
|
|
const struct hashmap_entry *vb,
|
2017-08-19 01:20:30 +03:00
|
|
|
const void *vkey)
|
|
|
|
{
|
2019-10-07 02:30:37 +03:00
|
|
|
const struct delta_base_cache_entry *a, *b;
|
2017-08-19 01:20:30 +03:00
|
|
|
const struct delta_base_cache_key *key = vkey;
|
2019-10-07 02:30:37 +03:00
|
|
|
|
|
|
|
a = container_of(va, const struct delta_base_cache_entry, ent);
|
|
|
|
b = container_of(vb, const struct delta_base_cache_entry, ent);
|
|
|
|
|
2017-08-19 01:20:30 +03:00
|
|
|
if (key)
|
|
|
|
return !delta_base_cache_key_eq(&a->key, key);
|
|
|
|
else
|
|
|
|
return !delta_base_cache_key_eq(&a->key, &b->key);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int in_delta_base_cache(struct packed_git *p, off_t base_offset)
|
|
|
|
{
|
|
|
|
return !!get_delta_base_cache_entry(p, base_offset);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Remove the entry from the cache, but do _not_ free the associated
|
|
|
|
* entry data. The caller takes ownership of the "data" buffer, and
|
|
|
|
* should copy out any fields it wants before detaching.
|
|
|
|
*/
|
|
|
|
static void detach_delta_base_cache_entry(struct delta_base_cache_entry *ent)
|
|
|
|
{
|
2019-10-07 02:30:31 +03:00
|
|
|
hashmap_remove(&delta_base_cache, &ent->ent, &ent->key);
|
2017-08-19 01:20:30 +03:00
|
|
|
list_del(&ent->lru);
|
|
|
|
delta_base_cached -= ent->size;
|
|
|
|
free(ent);
|
|
|
|
}
|
|
|
|
|
2018-04-25 21:21:06 +03:00
|
|
|
static void *cache_or_unpack_entry(struct repository *r, struct packed_git *p,
|
|
|
|
off_t base_offset, unsigned long *base_size,
|
|
|
|
enum object_type *type)
|
2017-08-19 01:20:30 +03:00
|
|
|
{
|
|
|
|
struct delta_base_cache_entry *ent;
|
|
|
|
|
|
|
|
ent = get_delta_base_cache_entry(p, base_offset);
|
|
|
|
if (!ent)
|
2018-04-25 21:21:06 +03:00
|
|
|
return unpack_entry(r, p, base_offset, type, base_size);
|
2017-08-19 01:20:30 +03:00
|
|
|
|
|
|
|
if (type)
|
|
|
|
*type = ent->type;
|
|
|
|
if (base_size)
|
|
|
|
*base_size = ent->size;
|
|
|
|
return xmemdupz(ent->data, ent->size);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void release_delta_base_cache(struct delta_base_cache_entry *ent)
|
|
|
|
{
|
|
|
|
free(ent->data);
|
|
|
|
detach_delta_base_cache_entry(ent);
|
|
|
|
}
|
|
|
|
|
|
|
|
void clear_delta_base_cache(void)
|
|
|
|
{
|
|
|
|
struct list_head *lru, *tmp;
|
|
|
|
list_for_each_safe(lru, tmp, &delta_base_cache_lru) {
|
|
|
|
struct delta_base_cache_entry *entry =
|
|
|
|
list_entry(lru, struct delta_base_cache_entry, lru);
|
|
|
|
release_delta_base_cache(entry);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void add_delta_base_cache(struct packed_git *p, off_t base_offset,
|
|
|
|
void *base, unsigned long base_size, enum object_type type)
|
|
|
|
{
|
2020-09-29 03:01:53 +03:00
|
|
|
struct delta_base_cache_entry *ent;
|
2017-08-19 01:20:30 +03:00
|
|
|
struct list_head *lru, *tmp;
|
|
|
|
|
object-store: allow threaded access to object reading
Allow object reading to be performed by multiple threads protecting it
with an internal lock, the obj_read_mutex. The lock usage can be toggled
with enable_obj_read_lock() and disable_obj_read_lock(). Currently, the
functions which can be safely called in parallel are:
read_object_file_extended(), repo_read_object_file(),
read_object_file(), read_object_with_reference(), read_object(),
oid_object_info() and oid_object_info_extended(). It's also possible
to use obj_read_lock() and obj_read_unlock() to protect other sections
that cannot execute in parallel with object reading.
Probably there are many spots in the functions listed above that could
be executed unlocked (and thus, in parallel). But, for now, we are most
interested in allowing parallel access to zlib inflation. This is one of
the sections where object reading spends most of the time in (e.g. up to
one-third of git-grep's execution time in the chromium repo corresponds
to inflation) and it's already thread-safe. So, to take advantage of
that, the obj_read_mutex is released when calling git_inflate() and
re-acquired right after, for every calling spot in
oid_object_info_extended()'s call chain. We may refine this lock to also
exploit other possible parallel spots in the future, but for now,
threaded zlib inflation should already give great speedups for threaded
object reading callers.
Note that add_delta_base_cache() was also modified to skip adding
already present entries to the cache. This wasn't possible before, but
it would be now, with the parallel inflation. Take for example the
following situation, where two threads - A and B - are executing the
code at unpack_entry():
1. Thread A is performing the decompression of a base O (which is not
yet in the cache) at PHASE II. Thread B is simultaneously trying to
unpack O, but just starting at PHASE I.
2. Since O is not yet in the cache, B will go to PHASE II to also
perform the decompression.
3. When they finish decompressing, one of them will get the object
reading mutex and go to PHASE III while the other waits for the
mutex. Let’s say A got the mutex first.
4. Thread A will add O to the cache, go throughout the rest of PHASE III
and return.
5. Thread B gets the mutex, also add O to the cache (if the check wasn't
there) and returns.
Finally, it is also important to highlight that the object reading lock
can only ensure thread-safety in the mentioned functions thanks to two
complementary mechanisms: the use of 'struct raw_object_store's
replace_mutex, which guards sections in the object reading machinery
that would otherwise be thread-unsafe; and the 'struct pack_window's
inuse_cnt, which protects window reading operations (such as the one
performed during the inflation of a packed object), allowing them to
execute without the acquisition of the obj_read_mutex.
Signed-off-by: Matheus Tavares <matheus.bernardino@usp.br>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-01-16 05:39:53 +03:00
|
|
|
/*
|
|
|
|
* Check required to avoid redundant entries when more than one thread
|
|
|
|
* is unpacking the same object, in unpack_entry() (since its phases I
|
|
|
|
* and III might run concurrently across multiple threads).
|
|
|
|
*/
|
2020-09-29 03:01:53 +03:00
|
|
|
if (in_delta_base_cache(p, base_offset)) {
|
|
|
|
free(base);
|
object-store: allow threaded access to object reading
Allow object reading to be performed by multiple threads protecting it
with an internal lock, the obj_read_mutex. The lock usage can be toggled
with enable_obj_read_lock() and disable_obj_read_lock(). Currently, the
functions which can be safely called in parallel are:
read_object_file_extended(), repo_read_object_file(),
read_object_file(), read_object_with_reference(), read_object(),
oid_object_info() and oid_object_info_extended(). It's also possible
to use obj_read_lock() and obj_read_unlock() to protect other sections
that cannot execute in parallel with object reading.
Probably there are many spots in the functions listed above that could
be executed unlocked (and thus, in parallel). But, for now, we are most
interested in allowing parallel access to zlib inflation. This is one of
the sections where object reading spends most of the time in (e.g. up to
one-third of git-grep's execution time in the chromium repo corresponds
to inflation) and it's already thread-safe. So, to take advantage of
that, the obj_read_mutex is released when calling git_inflate() and
re-acquired right after, for every calling spot in
oid_object_info_extended()'s call chain. We may refine this lock to also
exploit other possible parallel spots in the future, but for now,
threaded zlib inflation should already give great speedups for threaded
object reading callers.
Note that add_delta_base_cache() was also modified to skip adding
already present entries to the cache. This wasn't possible before, but
it would be now, with the parallel inflation. Take for example the
following situation, where two threads - A and B - are executing the
code at unpack_entry():
1. Thread A is performing the decompression of a base O (which is not
yet in the cache) at PHASE II. Thread B is simultaneously trying to
unpack O, but just starting at PHASE I.
2. Since O is not yet in the cache, B will go to PHASE II to also
perform the decompression.
3. When they finish decompressing, one of them will get the object
reading mutex and go to PHASE III while the other waits for the
mutex. Let’s say A got the mutex first.
4. Thread A will add O to the cache, go throughout the rest of PHASE III
and return.
5. Thread B gets the mutex, also add O to the cache (if the check wasn't
there) and returns.
Finally, it is also important to highlight that the object reading lock
can only ensure thread-safety in the mentioned functions thanks to two
complementary mechanisms: the use of 'struct raw_object_store's
replace_mutex, which guards sections in the object reading machinery
that would otherwise be thread-unsafe; and the 'struct pack_window's
inuse_cnt, which protects window reading operations (such as the one
performed during the inflation of a packed object), allowing them to
execute without the acquisition of the obj_read_mutex.
Signed-off-by: Matheus Tavares <matheus.bernardino@usp.br>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-01-16 05:39:53 +03:00
|
|
|
return;
|
2020-09-29 03:01:53 +03:00
|
|
|
}
|
object-store: allow threaded access to object reading
Allow object reading to be performed by multiple threads protecting it
with an internal lock, the obj_read_mutex. The lock usage can be toggled
with enable_obj_read_lock() and disable_obj_read_lock(). Currently, the
functions which can be safely called in parallel are:
read_object_file_extended(), repo_read_object_file(),
read_object_file(), read_object_with_reference(), read_object(),
oid_object_info() and oid_object_info_extended(). It's also possible
to use obj_read_lock() and obj_read_unlock() to protect other sections
that cannot execute in parallel with object reading.
Probably there are many spots in the functions listed above that could
be executed unlocked (and thus, in parallel). But, for now, we are most
interested in allowing parallel access to zlib inflation. This is one of
the sections where object reading spends most of the time in (e.g. up to
one-third of git-grep's execution time in the chromium repo corresponds
to inflation) and it's already thread-safe. So, to take advantage of
that, the obj_read_mutex is released when calling git_inflate() and
re-acquired right after, for every calling spot in
oid_object_info_extended()'s call chain. We may refine this lock to also
exploit other possible parallel spots in the future, but for now,
threaded zlib inflation should already give great speedups for threaded
object reading callers.
Note that add_delta_base_cache() was also modified to skip adding
already present entries to the cache. This wasn't possible before, but
it would be now, with the parallel inflation. Take for example the
following situation, where two threads - A and B - are executing the
code at unpack_entry():
1. Thread A is performing the decompression of a base O (which is not
yet in the cache) at PHASE II. Thread B is simultaneously trying to
unpack O, but just starting at PHASE I.
2. Since O is not yet in the cache, B will go to PHASE II to also
perform the decompression.
3. When they finish decompressing, one of them will get the object
reading mutex and go to PHASE III while the other waits for the
mutex. Let’s say A got the mutex first.
4. Thread A will add O to the cache, go throughout the rest of PHASE III
and return.
5. Thread B gets the mutex, also add O to the cache (if the check wasn't
there) and returns.
Finally, it is also important to highlight that the object reading lock
can only ensure thread-safety in the mentioned functions thanks to two
complementary mechanisms: the use of 'struct raw_object_store's
replace_mutex, which guards sections in the object reading machinery
that would otherwise be thread-unsafe; and the 'struct pack_window's
inuse_cnt, which protects window reading operations (such as the one
performed during the inflation of a packed object), allowing them to
execute without the acquisition of the obj_read_mutex.
Signed-off-by: Matheus Tavares <matheus.bernardino@usp.br>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-01-16 05:39:53 +03:00
|
|
|
|
2017-08-19 01:20:30 +03:00
|
|
|
delta_base_cached += base_size;
|
|
|
|
|
|
|
|
list_for_each_safe(lru, tmp, &delta_base_cache_lru) {
|
|
|
|
struct delta_base_cache_entry *f =
|
|
|
|
list_entry(lru, struct delta_base_cache_entry, lru);
|
|
|
|
if (delta_base_cached <= delta_base_cache_limit)
|
|
|
|
break;
|
|
|
|
release_delta_base_cache(f);
|
|
|
|
}
|
|
|
|
|
2020-09-29 03:01:53 +03:00
|
|
|
ent = xmalloc(sizeof(*ent));
|
2017-08-19 01:20:30 +03:00
|
|
|
ent->key.p = p;
|
|
|
|
ent->key.base_offset = base_offset;
|
|
|
|
ent->type = type;
|
|
|
|
ent->data = base;
|
|
|
|
ent->size = base_size;
|
|
|
|
list_add_tail(&ent->lru, &delta_base_cache_lru);
|
|
|
|
|
|
|
|
if (!delta_base_cache.cmpfn)
|
|
|
|
hashmap_init(&delta_base_cache, delta_base_cache_hash_cmp, NULL, 0);
|
2019-10-07 02:30:27 +03:00
|
|
|
hashmap_entry_init(&ent->ent, pack_entry_hash(p, base_offset));
|
2019-10-07 02:30:29 +03:00
|
|
|
hashmap_add(&delta_base_cache, &ent->ent);
|
2017-08-19 01:20:30 +03:00
|
|
|
}
|
|
|
|
|
2018-04-25 21:21:06 +03:00
|
|
|
int packed_object_info(struct repository *r, struct packed_git *p,
|
|
|
|
off_t obj_offset, struct object_info *oi)
|
2017-08-19 01:20:30 +03:00
|
|
|
{
|
|
|
|
struct pack_window *w_curs = NULL;
|
|
|
|
unsigned long size;
|
|
|
|
off_t curpos = obj_offset;
|
|
|
|
enum object_type type;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We always get the representation type, but only convert it to
|
|
|
|
* a "real" type later if the caller is interested.
|
|
|
|
*/
|
|
|
|
if (oi->contentp) {
|
2018-04-25 21:21:06 +03:00
|
|
|
*oi->contentp = cache_or_unpack_entry(r, p, obj_offset, oi->sizep,
|
2017-08-19 01:20:30 +03:00
|
|
|
&type);
|
|
|
|
if (!*oi->contentp)
|
|
|
|
type = OBJ_BAD;
|
|
|
|
} else {
|
|
|
|
type = unpack_object_header(p, &w_curs, &curpos, &size);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!oi->contentp && oi->sizep) {
|
|
|
|
if (type == OBJ_OFS_DELTA || type == OBJ_REF_DELTA) {
|
|
|
|
off_t tmp_pos = curpos;
|
|
|
|
off_t base_offset = get_delta_base(p, &w_curs, &tmp_pos,
|
|
|
|
type, obj_offset);
|
|
|
|
if (!base_offset) {
|
|
|
|
type = OBJ_BAD;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
*oi->sizep = get_size_from_delta(p, &w_curs, tmp_pos);
|
|
|
|
if (*oi->sizep == 0) {
|
|
|
|
type = OBJ_BAD;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
*oi->sizep = size;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (oi->disk_sizep) {
|
|
|
|
struct revindex_entry *revidx = find_pack_revindex(p, obj_offset);
|
|
|
|
*oi->disk_sizep = revidx[1].offset - obj_offset;
|
|
|
|
}
|
|
|
|
|
2018-02-14 21:59:23 +03:00
|
|
|
if (oi->typep || oi->type_name) {
|
2017-08-19 01:20:30 +03:00
|
|
|
enum object_type ptot;
|
2018-04-25 21:21:06 +03:00
|
|
|
ptot = packed_to_object_type(r, p, obj_offset,
|
2018-04-25 21:21:01 +03:00
|
|
|
type, &w_curs, curpos);
|
2017-08-19 01:20:30 +03:00
|
|
|
if (oi->typep)
|
|
|
|
*oi->typep = ptot;
|
2018-02-14 21:59:23 +03:00
|
|
|
if (oi->type_name) {
|
2018-02-14 21:59:24 +03:00
|
|
|
const char *tn = type_name(ptot);
|
2017-08-19 01:20:30 +03:00
|
|
|
if (tn)
|
2018-02-14 21:59:23 +03:00
|
|
|
strbuf_addstr(oi->type_name, tn);
|
2017-08-19 01:20:30 +03:00
|
|
|
}
|
|
|
|
if (ptot < 0) {
|
|
|
|
type = OBJ_BAD;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2020-02-24 07:36:56 +03:00
|
|
|
if (oi->delta_base_oid) {
|
2017-08-19 01:20:30 +03:00
|
|
|
if (type == OBJ_OFS_DELTA || type == OBJ_REF_DELTA) {
|
2020-02-24 07:37:31 +03:00
|
|
|
if (get_delta_base_oid(p, &w_curs, curpos,
|
|
|
|
oi->delta_base_oid,
|
|
|
|
type, obj_offset) < 0) {
|
2017-08-19 01:20:30 +03:00
|
|
|
type = OBJ_BAD;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
} else
|
2020-02-24 07:36:56 +03:00
|
|
|
oidclr(oi->delta_base_oid);
|
2017-08-19 01:20:30 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
oi->whence = in_delta_base_cache(p, obj_offset) ? OI_DBCACHED :
|
|
|
|
OI_PACKED;
|
|
|
|
|
|
|
|
out:
|
|
|
|
unuse_pack(&w_curs);
|
|
|
|
return type;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void *unpack_compressed_entry(struct packed_git *p,
|
|
|
|
struct pack_window **w_curs,
|
|
|
|
off_t curpos,
|
|
|
|
unsigned long size)
|
|
|
|
{
|
|
|
|
int st;
|
|
|
|
git_zstream stream;
|
|
|
|
unsigned char *buffer, *in;
|
|
|
|
|
|
|
|
buffer = xmallocz_gently(size);
|
|
|
|
if (!buffer)
|
|
|
|
return NULL;
|
|
|
|
memset(&stream, 0, sizeof(stream));
|
|
|
|
stream.next_out = buffer;
|
|
|
|
stream.avail_out = size + 1;
|
|
|
|
|
|
|
|
git_inflate_init(&stream);
|
|
|
|
do {
|
|
|
|
in = use_pack(p, w_curs, curpos, &stream.avail_in);
|
|
|
|
stream.next_in = in;
|
object-store: allow threaded access to object reading
Allow object reading to be performed by multiple threads protecting it
with an internal lock, the obj_read_mutex. The lock usage can be toggled
with enable_obj_read_lock() and disable_obj_read_lock(). Currently, the
functions which can be safely called in parallel are:
read_object_file_extended(), repo_read_object_file(),
read_object_file(), read_object_with_reference(), read_object(),
oid_object_info() and oid_object_info_extended(). It's also possible
to use obj_read_lock() and obj_read_unlock() to protect other sections
that cannot execute in parallel with object reading.
Probably there are many spots in the functions listed above that could
be executed unlocked (and thus, in parallel). But, for now, we are most
interested in allowing parallel access to zlib inflation. This is one of
the sections where object reading spends most of the time in (e.g. up to
one-third of git-grep's execution time in the chromium repo corresponds
to inflation) and it's already thread-safe. So, to take advantage of
that, the obj_read_mutex is released when calling git_inflate() and
re-acquired right after, for every calling spot in
oid_object_info_extended()'s call chain. We may refine this lock to also
exploit other possible parallel spots in the future, but for now,
threaded zlib inflation should already give great speedups for threaded
object reading callers.
Note that add_delta_base_cache() was also modified to skip adding
already present entries to the cache. This wasn't possible before, but
it would be now, with the parallel inflation. Take for example the
following situation, where two threads - A and B - are executing the
code at unpack_entry():
1. Thread A is performing the decompression of a base O (which is not
yet in the cache) at PHASE II. Thread B is simultaneously trying to
unpack O, but just starting at PHASE I.
2. Since O is not yet in the cache, B will go to PHASE II to also
perform the decompression.
3. When they finish decompressing, one of them will get the object
reading mutex and go to PHASE III while the other waits for the
mutex. Let’s say A got the mutex first.
4. Thread A will add O to the cache, go throughout the rest of PHASE III
and return.
5. Thread B gets the mutex, also add O to the cache (if the check wasn't
there) and returns.
Finally, it is also important to highlight that the object reading lock
can only ensure thread-safety in the mentioned functions thanks to two
complementary mechanisms: the use of 'struct raw_object_store's
replace_mutex, which guards sections in the object reading machinery
that would otherwise be thread-unsafe; and the 'struct pack_window's
inuse_cnt, which protects window reading operations (such as the one
performed during the inflation of a packed object), allowing them to
execute without the acquisition of the obj_read_mutex.
Signed-off-by: Matheus Tavares <matheus.bernardino@usp.br>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-01-16 05:39:53 +03:00
|
|
|
/*
|
|
|
|
* Note: we must ensure the window section returned by
|
|
|
|
* use_pack() will be available throughout git_inflate()'s
|
|
|
|
* unlocked execution. Please refer to the comment at
|
|
|
|
* get_size_from_delta() to see how this is done.
|
|
|
|
*/
|
|
|
|
obj_read_unlock();
|
2017-08-19 01:20:30 +03:00
|
|
|
st = git_inflate(&stream, Z_FINISH);
|
object-store: allow threaded access to object reading
Allow object reading to be performed by multiple threads protecting it
with an internal lock, the obj_read_mutex. The lock usage can be toggled
with enable_obj_read_lock() and disable_obj_read_lock(). Currently, the
functions which can be safely called in parallel are:
read_object_file_extended(), repo_read_object_file(),
read_object_file(), read_object_with_reference(), read_object(),
oid_object_info() and oid_object_info_extended(). It's also possible
to use obj_read_lock() and obj_read_unlock() to protect other sections
that cannot execute in parallel with object reading.
Probably there are many spots in the functions listed above that could
be executed unlocked (and thus, in parallel). But, for now, we are most
interested in allowing parallel access to zlib inflation. This is one of
the sections where object reading spends most of the time in (e.g. up to
one-third of git-grep's execution time in the chromium repo corresponds
to inflation) and it's already thread-safe. So, to take advantage of
that, the obj_read_mutex is released when calling git_inflate() and
re-acquired right after, for every calling spot in
oid_object_info_extended()'s call chain. We may refine this lock to also
exploit other possible parallel spots in the future, but for now,
threaded zlib inflation should already give great speedups for threaded
object reading callers.
Note that add_delta_base_cache() was also modified to skip adding
already present entries to the cache. This wasn't possible before, but
it would be now, with the parallel inflation. Take for example the
following situation, where two threads - A and B - are executing the
code at unpack_entry():
1. Thread A is performing the decompression of a base O (which is not
yet in the cache) at PHASE II. Thread B is simultaneously trying to
unpack O, but just starting at PHASE I.
2. Since O is not yet in the cache, B will go to PHASE II to also
perform the decompression.
3. When they finish decompressing, one of them will get the object
reading mutex and go to PHASE III while the other waits for the
mutex. Let’s say A got the mutex first.
4. Thread A will add O to the cache, go throughout the rest of PHASE III
and return.
5. Thread B gets the mutex, also add O to the cache (if the check wasn't
there) and returns.
Finally, it is also important to highlight that the object reading lock
can only ensure thread-safety in the mentioned functions thanks to two
complementary mechanisms: the use of 'struct raw_object_store's
replace_mutex, which guards sections in the object reading machinery
that would otherwise be thread-unsafe; and the 'struct pack_window's
inuse_cnt, which protects window reading operations (such as the one
performed during the inflation of a packed object), allowing them to
execute without the acquisition of the obj_read_mutex.
Signed-off-by: Matheus Tavares <matheus.bernardino@usp.br>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2020-01-16 05:39:53 +03:00
|
|
|
obj_read_lock();
|
2017-08-19 01:20:30 +03:00
|
|
|
if (!stream.avail_out)
|
|
|
|
break; /* the payload is larger than it should be */
|
|
|
|
curpos += stream.next_in - in;
|
|
|
|
} while (st == Z_OK || st == Z_BUF_ERROR);
|
|
|
|
git_inflate_end(&stream);
|
|
|
|
if ((st != Z_STREAM_END) || stream.total_out != size) {
|
|
|
|
free(buffer);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2018-06-13 17:22:07 +03:00
|
|
|
/* versions of zlib can clobber unconsumed portion of outbuf */
|
|
|
|
buffer[size] = '\0';
|
|
|
|
|
2017-08-19 01:20:30 +03:00
|
|
|
return buffer;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void write_pack_access_log(struct packed_git *p, off_t obj_offset)
|
|
|
|
{
|
|
|
|
static struct trace_key pack_access = TRACE_KEY_INIT(PACK_ACCESS);
|
|
|
|
trace_printf_key(&pack_access, "%s %"PRIuMAX"\n",
|
|
|
|
p->pack_name, (uintmax_t)obj_offset);
|
|
|
|
}
|
|
|
|
|
|
|
|
int do_check_packed_object_crc;
|
|
|
|
|
|
|
|
#define UNPACK_ENTRY_STACK_PREALLOC 64
|
|
|
|
struct unpack_entry_stack_ent {
|
|
|
|
off_t obj_offset;
|
|
|
|
off_t curpos;
|
|
|
|
unsigned long size;
|
|
|
|
};
|
|
|
|
|
2018-04-25 21:21:06 +03:00
|
|
|
static void *read_object(struct repository *r,
|
|
|
|
const struct object_id *oid,
|
|
|
|
enum object_type *type,
|
2017-08-19 01:20:30 +03:00
|
|
|
unsigned long *size)
|
|
|
|
{
|
|
|
|
struct object_info oi = OBJECT_INFO_INIT;
|
|
|
|
void *content;
|
|
|
|
oi.typep = type;
|
|
|
|
oi.sizep = size;
|
|
|
|
oi.contentp = &content;
|
|
|
|
|
2018-04-25 21:21:06 +03:00
|
|
|
if (oid_object_info_extended(r, oid, &oi, 0) < 0)
|
2017-08-19 01:20:30 +03:00
|
|
|
return NULL;
|
|
|
|
return content;
|
|
|
|
}
|
|
|
|
|
2018-04-25 21:21:06 +03:00
|
|
|
void *unpack_entry(struct repository *r, struct packed_git *p, off_t obj_offset,
|
2017-08-19 01:20:30 +03:00
|
|
|
enum object_type *final_type, unsigned long *final_size)
|
|
|
|
{
|
|
|
|
struct pack_window *w_curs = NULL;
|
|
|
|
off_t curpos = obj_offset;
|
|
|
|
void *data = NULL;
|
|
|
|
unsigned long size;
|
|
|
|
enum object_type type;
|
|
|
|
struct unpack_entry_stack_ent small_delta_stack[UNPACK_ENTRY_STACK_PREALLOC];
|
|
|
|
struct unpack_entry_stack_ent *delta_stack = small_delta_stack;
|
|
|
|
int delta_stack_nr = 0, delta_stack_alloc = UNPACK_ENTRY_STACK_PREALLOC;
|
|
|
|
int base_from_cache = 0;
|
|
|
|
|
|
|
|
write_pack_access_log(p, obj_offset);
|
|
|
|
|
|
|
|
/* PHASE 1: drill down to the innermost base object */
|
|
|
|
for (;;) {
|
|
|
|
off_t base_offset;
|
|
|
|
int i;
|
|
|
|
struct delta_base_cache_entry *ent;
|
|
|
|
|
|
|
|
ent = get_delta_base_cache_entry(p, curpos);
|
|
|
|
if (ent) {
|
|
|
|
type = ent->type;
|
|
|
|
data = ent->data;
|
|
|
|
size = ent->size;
|
|
|
|
detach_delta_base_cache_entry(ent);
|
|
|
|
base_from_cache = 1;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (do_check_packed_object_crc && p->index_version > 1) {
|
|
|
|
struct revindex_entry *revidx = find_pack_revindex(p, obj_offset);
|
|
|
|
off_t len = revidx[1].offset - obj_offset;
|
|
|
|
if (check_pack_crc(p, &w_curs, obj_offset, len, revidx->nr)) {
|
2018-03-12 05:27:44 +03:00
|
|
|
struct object_id oid;
|
2020-02-24 07:27:36 +03:00
|
|
|
nth_packed_object_id(&oid, p, revidx->nr);
|
2017-08-19 01:20:30 +03:00
|
|
|
error("bad packed object CRC for %s",
|
2018-03-12 05:27:44 +03:00
|
|
|
oid_to_hex(&oid));
|
|
|
|
mark_bad_packed_object(p, oid.hash);
|
2017-08-19 01:20:30 +03:00
|
|
|
data = NULL;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
type = unpack_object_header(p, &w_curs, &curpos, &size);
|
|
|
|
if (type != OBJ_OFS_DELTA && type != OBJ_REF_DELTA)
|
|
|
|
break;
|
|
|
|
|
|
|
|
base_offset = get_delta_base(p, &w_curs, &curpos, type, obj_offset);
|
|
|
|
if (!base_offset) {
|
|
|
|
error("failed to validate delta base reference "
|
|
|
|
"at offset %"PRIuMAX" from %s",
|
|
|
|
(uintmax_t)curpos, p->pack_name);
|
|
|
|
/* bail to phase 2, in hopes of recovery */
|
|
|
|
data = NULL;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* push object, proceed to base */
|
|
|
|
if (delta_stack_nr >= delta_stack_alloc
|
|
|
|
&& delta_stack == small_delta_stack) {
|
|
|
|
delta_stack_alloc = alloc_nr(delta_stack_nr);
|
|
|
|
ALLOC_ARRAY(delta_stack, delta_stack_alloc);
|
2019-06-15 21:36:35 +03:00
|
|
|
COPY_ARRAY(delta_stack, small_delta_stack,
|
|
|
|
delta_stack_nr);
|
2017-08-19 01:20:30 +03:00
|
|
|
} else {
|
|
|
|
ALLOC_GROW(delta_stack, delta_stack_nr+1, delta_stack_alloc);
|
|
|
|
}
|
|
|
|
i = delta_stack_nr++;
|
|
|
|
delta_stack[i].obj_offset = obj_offset;
|
|
|
|
delta_stack[i].curpos = curpos;
|
|
|
|
delta_stack[i].size = size;
|
|
|
|
|
|
|
|
curpos = obj_offset = base_offset;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* PHASE 2: handle the base */
|
|
|
|
switch (type) {
|
|
|
|
case OBJ_OFS_DELTA:
|
|
|
|
case OBJ_REF_DELTA:
|
|
|
|
if (data)
|
2018-05-02 12:38:39 +03:00
|
|
|
BUG("unpack_entry: left loop at a valid delta");
|
2017-08-19 01:20:30 +03:00
|
|
|
break;
|
|
|
|
case OBJ_COMMIT:
|
|
|
|
case OBJ_TREE:
|
|
|
|
case OBJ_BLOB:
|
|
|
|
case OBJ_TAG:
|
|
|
|
if (!base_from_cache)
|
|
|
|
data = unpack_compressed_entry(p, &w_curs, curpos, size);
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
data = NULL;
|
|
|
|
error("unknown object type %i at offset %"PRIuMAX" in %s",
|
|
|
|
type, (uintmax_t)obj_offset, p->pack_name);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* PHASE 3: apply deltas in order */
|
|
|
|
|
|
|
|
/* invariants:
|
|
|
|
* 'data' holds the base data, or NULL if there was corruption
|
|
|
|
*/
|
|
|
|
while (delta_stack_nr) {
|
|
|
|
void *delta_data;
|
|
|
|
void *base = data;
|
|
|
|
void *external_base = NULL;
|
|
|
|
unsigned long delta_size, base_size = size;
|
|
|
|
int i;
|
2020-09-29 03:01:52 +03:00
|
|
|
off_t base_obj_offset = obj_offset;
|
2017-08-19 01:20:30 +03:00
|
|
|
|
|
|
|
data = NULL;
|
|
|
|
|
|
|
|
if (!base) {
|
|
|
|
/*
|
|
|
|
* We're probably in deep shit, but let's try to fetch
|
|
|
|
* the required base anyway from another pack or loose.
|
|
|
|
* This is costly but should happen only in the presence
|
|
|
|
* of a corrupted pack, and is better than failing outright.
|
|
|
|
*/
|
|
|
|
struct revindex_entry *revidx;
|
2018-03-12 05:27:44 +03:00
|
|
|
struct object_id base_oid;
|
2017-08-19 01:20:30 +03:00
|
|
|
revidx = find_pack_revindex(p, obj_offset);
|
|
|
|
if (revidx) {
|
2020-02-24 07:27:36 +03:00
|
|
|
nth_packed_object_id(&base_oid, p, revidx->nr);
|
2017-08-19 01:20:30 +03:00
|
|
|
error("failed to read delta base object %s"
|
|
|
|
" at offset %"PRIuMAX" from %s",
|
2018-03-12 05:27:44 +03:00
|
|
|
oid_to_hex(&base_oid), (uintmax_t)obj_offset,
|
2017-08-19 01:20:30 +03:00
|
|
|
p->pack_name);
|
2018-03-12 05:27:44 +03:00
|
|
|
mark_bad_packed_object(p, base_oid.hash);
|
2018-04-25 21:21:06 +03:00
|
|
|
base = read_object(r, &base_oid, &type, &base_size);
|
2017-08-19 01:20:30 +03:00
|
|
|
external_base = base;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
i = --delta_stack_nr;
|
|
|
|
obj_offset = delta_stack[i].obj_offset;
|
|
|
|
curpos = delta_stack[i].curpos;
|
|
|
|
delta_size = delta_stack[i].size;
|
|
|
|
|
|
|
|
if (!base)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
delta_data = unpack_compressed_entry(p, &w_curs, curpos, delta_size);
|
|
|
|
|
|
|
|
if (!delta_data) {
|
|
|
|
error("failed to unpack compressed delta "
|
|
|
|
"at offset %"PRIuMAX" from %s",
|
|
|
|
(uintmax_t)curpos, p->pack_name);
|
|
|
|
data = NULL;
|
2020-09-29 03:01:52 +03:00
|
|
|
} else {
|
|
|
|
data = patch_delta(base, base_size, delta_data,
|
|
|
|
delta_size, &size);
|
2017-08-19 01:20:30 +03:00
|
|
|
|
2020-09-29 03:01:52 +03:00
|
|
|
/*
|
|
|
|
* We could not apply the delta; warn the user, but
|
|
|
|
* keep going. Our failure will be noticed either in
|
|
|
|
* the next iteration of the loop, or if this is the
|
|
|
|
* final delta, in the caller when we return NULL.
|
|
|
|
* Those code paths will take care of making a more
|
|
|
|
* explicit warning and retrying with another copy of
|
|
|
|
* the object.
|
|
|
|
*/
|
|
|
|
if (!data)
|
|
|
|
error("failed to apply delta");
|
|
|
|
}
|
2017-08-19 01:20:30 +03:00
|
|
|
|
|
|
|
/*
|
2020-09-29 03:01:52 +03:00
|
|
|
* We delay adding `base` to the cache until the end of the loop
|
|
|
|
* because unpack_compressed_entry() momentarily releases the
|
|
|
|
* obj_read_mutex, giving another thread the chance to access
|
|
|
|
* the cache. Therefore, if `base` was already there, this other
|
|
|
|
* thread could free() it (e.g. to make space for another entry)
|
|
|
|
* before we are done using it.
|
2017-08-19 01:20:30 +03:00
|
|
|
*/
|
2020-09-29 03:01:52 +03:00
|
|
|
if (!external_base)
|
|
|
|
add_delta_base_cache(p, base_obj_offset, base, base_size, type);
|
2017-08-19 01:20:30 +03:00
|
|
|
|
|
|
|
free(delta_data);
|
|
|
|
free(external_base);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (final_type)
|
|
|
|
*final_type = type;
|
|
|
|
if (final_size)
|
|
|
|
*final_size = size;
|
|
|
|
|
|
|
|
out:
|
|
|
|
unuse_pack(&w_curs);
|
|
|
|
|
|
|
|
if (delta_stack != small_delta_stack)
|
|
|
|
free(delta_stack);
|
|
|
|
|
|
|
|
return data;
|
|
|
|
}
|
2017-08-19 01:20:31 +03:00
|
|
|
|
2018-03-22 20:40:09 +03:00
|
|
|
int bsearch_pack(const struct object_id *oid, const struct packed_git *p, uint32_t *result)
|
|
|
|
{
|
|
|
|
const unsigned char *index_fanout = p->index_data;
|
|
|
|
const unsigned char *index_lookup;
|
2018-05-02 03:25:36 +03:00
|
|
|
const unsigned int hashsz = the_hash_algo->rawsz;
|
2018-03-22 20:40:09 +03:00
|
|
|
int index_lookup_width;
|
|
|
|
|
|
|
|
if (!index_fanout)
|
|
|
|
BUG("bsearch_pack called without a valid pack-index");
|
|
|
|
|
|
|
|
index_lookup = index_fanout + 4 * 256;
|
|
|
|
if (p->index_version == 1) {
|
2018-05-02 03:25:36 +03:00
|
|
|
index_lookup_width = hashsz + 4;
|
2018-03-22 20:40:09 +03:00
|
|
|
index_lookup += 4;
|
|
|
|
} else {
|
2018-05-02 03:25:36 +03:00
|
|
|
index_lookup_width = hashsz;
|
2018-03-22 20:40:09 +03:00
|
|
|
index_fanout += 8;
|
|
|
|
index_lookup += 8;
|
|
|
|
}
|
|
|
|
|
|
|
|
return bsearch_hash(oid->hash, (const uint32_t*)index_fanout,
|
|
|
|
index_lookup, index_lookup_width, result);
|
|
|
|
}
|
|
|
|
|
2020-02-24 07:37:54 +03:00
|
|
|
int nth_packed_object_id(struct object_id *oid,
|
|
|
|
struct packed_git *p,
|
|
|
|
uint32_t n)
|
2017-08-19 01:20:31 +03:00
|
|
|
{
|
|
|
|
const unsigned char *index = p->index_data;
|
2018-05-02 03:25:36 +03:00
|
|
|
const unsigned int hashsz = the_hash_algo->rawsz;
|
2017-08-19 01:20:31 +03:00
|
|
|
if (!index) {
|
|
|
|
if (open_pack_index(p))
|
2020-02-24 07:37:54 +03:00
|
|
|
return -1;
|
2017-08-19 01:20:31 +03:00
|
|
|
index = p->index_data;
|
|
|
|
}
|
|
|
|
if (n >= p->num_objects)
|
2020-02-24 07:37:54 +03:00
|
|
|
return -1;
|
2017-08-19 01:20:31 +03:00
|
|
|
index += 4 * 256;
|
|
|
|
if (p->index_version == 1) {
|
2020-02-24 07:37:54 +03:00
|
|
|
oidread(oid, index + (hashsz + 4) * n + 4);
|
2017-08-19 01:20:31 +03:00
|
|
|
} else {
|
|
|
|
index += 8;
|
2020-02-24 07:37:54 +03:00
|
|
|
oidread(oid, index + hashsz * n);
|
2017-08-19 01:20:31 +03:00
|
|
|
}
|
2020-02-24 07:27:36 +03:00
|
|
|
return 0;
|
2017-08-19 01:20:31 +03:00
|
|
|
}
|
2017-08-19 01:20:32 +03:00
|
|
|
|
|
|
|
void check_pack_index_ptr(const struct packed_git *p, const void *vptr)
|
|
|
|
{
|
|
|
|
const unsigned char *ptr = vptr;
|
|
|
|
const unsigned char *start = p->index_data;
|
|
|
|
const unsigned char *end = start + p->index_size;
|
|
|
|
if (ptr < start)
|
|
|
|
die(_("offset before start of pack index for %s (corrupt index?)"),
|
|
|
|
p->pack_name);
|
|
|
|
/* No need to check for underflow; .idx files must be at least 8 bytes */
|
|
|
|
if (ptr >= end - 8)
|
|
|
|
die(_("offset beyond end of pack index for %s (truncated index?)"),
|
|
|
|
p->pack_name);
|
|
|
|
}
|
|
|
|
|
|
|
|
off_t nth_packed_object_offset(const struct packed_git *p, uint32_t n)
|
|
|
|
{
|
|
|
|
const unsigned char *index = p->index_data;
|
2018-05-02 03:25:36 +03:00
|
|
|
const unsigned int hashsz = the_hash_algo->rawsz;
|
2017-08-19 01:20:32 +03:00
|
|
|
index += 4 * 256;
|
|
|
|
if (p->index_version == 1) {
|
2018-05-02 03:25:36 +03:00
|
|
|
return ntohl(*((uint32_t *)(index + (hashsz + 4) * n)));
|
2017-08-19 01:20:32 +03:00
|
|
|
} else {
|
|
|
|
uint32_t off;
|
2018-05-02 03:25:36 +03:00
|
|
|
index += 8 + p->num_objects * (hashsz + 4);
|
2017-08-19 01:20:32 +03:00
|
|
|
off = ntohl(*((uint32_t *)(index + 4 * n)));
|
|
|
|
if (!(off & 0x80000000))
|
|
|
|
return off;
|
|
|
|
index += p->num_objects * 4 + (off & 0x7fffffff) * 8;
|
|
|
|
check_pack_index_ptr(p, index);
|
2018-01-17 22:08:23 +03:00
|
|
|
return get_be64(index);
|
2017-08-19 01:20:32 +03:00
|
|
|
}
|
|
|
|
}
|
2017-08-19 01:20:33 +03:00
|
|
|
|
|
|
|
off_t find_pack_entry_one(const unsigned char *sha1,
|
|
|
|
struct packed_git *p)
|
|
|
|
{
|
|
|
|
const unsigned char *index = p->index_data;
|
2018-03-22 20:40:09 +03:00
|
|
|
struct object_id oid;
|
2018-02-13 21:39:39 +03:00
|
|
|
uint32_t result;
|
2017-08-19 01:20:33 +03:00
|
|
|
|
|
|
|
if (!index) {
|
|
|
|
if (open_pack_index(p))
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-03-22 20:40:09 +03:00
|
|
|
hashcpy(oid.hash, sha1);
|
|
|
|
if (bsearch_pack(&oid, p, &result))
|
2018-02-13 21:39:39 +03:00
|
|
|
return nth_packed_object_offset(p, result);
|
2017-08-19 01:20:33 +03:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int is_pack_valid(struct packed_git *p)
|
|
|
|
{
|
|
|
|
/* An already open pack is known to be valid. */
|
|
|
|
if (p->pack_fd != -1)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
/* If the pack has one window completely covering the
|
|
|
|
* file size, the pack is known to be valid even if
|
|
|
|
* the descriptor is not currently open.
|
|
|
|
*/
|
|
|
|
if (p->windows) {
|
|
|
|
struct pack_window *w = p->windows;
|
|
|
|
|
|
|
|
if (!w->offset && w->len == p->pack_size)
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Force the pack to open to prove its valid. */
|
|
|
|
return !open_packed_git(p);
|
|
|
|
}
|
2017-08-19 01:20:34 +03:00
|
|
|
|
|
|
|
struct packed_git *find_sha1_pack(const unsigned char *sha1,
|
|
|
|
struct packed_git *packs)
|
|
|
|
{
|
|
|
|
struct packed_git *p;
|
|
|
|
|
|
|
|
for (p = packs; p; p = p->next) {
|
|
|
|
if (find_pack_entry_one(sha1, p))
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
}
|
2017-08-19 01:20:35 +03:00
|
|
|
|
2018-05-02 03:25:35 +03:00
|
|
|
static int fill_pack_entry(const struct object_id *oid,
|
2017-08-19 01:20:35 +03:00
|
|
|
struct pack_entry *e,
|
|
|
|
struct packed_git *p)
|
|
|
|
{
|
|
|
|
off_t offset;
|
|
|
|
|
|
|
|
if (p->num_bad_objects) {
|
|
|
|
unsigned i;
|
|
|
|
for (i = 0; i < p->num_bad_objects; i++)
|
2018-08-29 00:22:44 +03:00
|
|
|
if (hasheq(oid->hash,
|
|
|
|
p->bad_object_sha1 + the_hash_algo->rawsz * i))
|
2017-08-19 01:20:35 +03:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-05-02 03:25:35 +03:00
|
|
|
offset = find_pack_entry_one(oid->hash, p);
|
2017-08-19 01:20:35 +03:00
|
|
|
if (!offset)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We are about to tell the caller where they can locate the
|
|
|
|
* requested object. We better make sure the packfile is
|
|
|
|
* still here and can be accessed before supplying that
|
|
|
|
* answer, as it may have been deleted since the index was
|
|
|
|
* loaded!
|
|
|
|
*/
|
|
|
|
if (!is_pack_valid(p))
|
|
|
|
return 0;
|
|
|
|
e->offset = offset;
|
|
|
|
e->p = p;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2018-05-02 03:25:35 +03:00
|
|
|
int find_pack_entry(struct repository *r, const struct object_id *oid, struct pack_entry *e)
|
2017-08-19 01:20:35 +03:00
|
|
|
{
|
2017-09-30 20:51:01 +03:00
|
|
|
struct list_head *pos;
|
2018-07-12 22:39:34 +03:00
|
|
|
struct multi_pack_index *m;
|
2017-08-19 01:20:35 +03:00
|
|
|
|
2018-03-23 20:45:26 +03:00
|
|
|
prepare_packed_git(r);
|
2018-07-12 22:39:34 +03:00
|
|
|
if (!r->objects->packed_git && !r->objects->multi_pack_index)
|
2017-08-19 01:20:35 +03:00
|
|
|
return 0;
|
|
|
|
|
2018-07-12 22:39:34 +03:00
|
|
|
for (m = r->objects->multi_pack_index; m; m = m->next) {
|
2019-04-29 19:18:55 +03:00
|
|
|
if (fill_midx_entry(r, oid, e, m))
|
2018-07-12 22:39:34 +03:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2018-03-23 20:45:26 +03:00
|
|
|
list_for_each(pos, &r->objects->packed_git_mru) {
|
2018-01-24 02:46:51 +03:00
|
|
|
struct packed_git *p = list_entry(pos, struct packed_git, mru);
|
midx: add packs to packed_git linked list
The multi-pack-index allows searching for objects across multiple
packs using one object list. The original design gains many of
these performance benefits by keeping the packs in the
multi-pack-index out of the packed_git list.
Unfortunately, this has one major drawback. If the multi-pack-index
covers thousands of packs, and a command loads many of those packs,
then we can hit the limit for open file descriptors. The
close_one_pack() method is used to limit this resource, but it
only looks at the packed_git list, and uses an LRU cache to prevent
thrashing.
Instead of complicating this close_one_pack() logic to include
direct references to the multi-pack-index, simply add the packs
opened by the multi-pack-index to the packed_git list. This
immediately solves the file-descriptor limit problem, but requires
some extra steps to avoid performance issues or other problems:
1. Create a multi_pack_index bit in the packed_git struct that is
one if and only if the pack was loaded from a multi-pack-index.
2. Skip packs with the multi_pack_index bit when doing object
lookups and abbreviations. These algorithms already check the
multi-pack-index before the packed_git struct. This has a very
small performance hit, as we need to walk more packed_git
structs. This is acceptable, since these operations run binary
search on the other packs, so this walk-and-ignore logic is
very fast by comparison.
3. When closing a multi-pack-index file, do not close its packs,
as those packs will be closed using close_all_packs(). In some
cases, such as 'git repack', we run 'close_midx()' without also
closing the packs, so we need to un-set the multi_pack_index bit
in those packs. This is necessary, and caught by running
t6501-freshen-objects.sh with GIT_TEST_MULTI_PACK_INDEX=1.
To manually test this change, I inserted trace2 logging into
close_pack_fd() and set pack_max_fds to 10, then ran 'git rev-list
--all --objects' on a copy of the Git repo with 300+ pack-files and
a multi-pack-index. The logs verified the packs are closed as
we read them beyond the file descriptor limit.
Signed-off-by: Derrick Stolee <dstolee@microsoft.com>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2019-04-29 19:18:56 +03:00
|
|
|
if (!p->multi_pack_index && fill_pack_entry(oid, e, p)) {
|
2018-03-23 20:45:26 +03:00
|
|
|
list_move(&p->mru, &r->objects->packed_git_mru);
|
2017-08-19 01:20:35 +03:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
2017-08-19 01:20:36 +03:00
|
|
|
|
2018-05-02 03:25:33 +03:00
|
|
|
int has_object_pack(const struct object_id *oid)
|
2017-08-19 01:20:36 +03:00
|
|
|
{
|
|
|
|
struct pack_entry e;
|
2018-05-02 03:25:35 +03:00
|
|
|
return find_pack_entry(the_repository, oid, &e);
|
2017-08-19 01:20:36 +03:00
|
|
|
}
|
2017-08-19 01:20:37 +03:00
|
|
|
|
|
|
|
int has_pack_index(const unsigned char *sha1)
|
|
|
|
{
|
|
|
|
struct stat st;
|
|
|
|
if (stat(sha1_pack_index_name(sha1), &st))
|
|
|
|
return 0;
|
|
|
|
return 1;
|
|
|
|
}
|
2017-08-19 01:20:38 +03:00
|
|
|
|
for_each_packed_object: support iterating in pack-order
We currently iterate over objects within a pack in .idx
order, which uses the object hashes. That means that it
is effectively random with respect to the location of the
object within the pack. If you're going to access the actual
object data, there are two reasons to move linearly through
the pack itself:
1. It improves the locality of access in the packfile. In
the cold-cache case, this may mean fewer disk seeks, or
better usage of disk cache.
2. We store related deltas together in the packfile. Which
means that the delta base cache can operate much more
efficiently if we visit all of those related deltas in
sequence, as the earlier items are likely to still be
in the cache. Whereas if we visit the objects in
random order, our cache entries are much more likely to
have been evicted by unrelated deltas in the meantime.
So in general, if you're going to access the object contents
pack order is generally going to end up more efficient.
But if you're simply generating a list of object names, or
if you're going to end up sorting the result anyway, you're
better off just using the .idx order, as finding the pack
order means generating the in-memory pack-revindex.
According to the numbers in 8b8dfd5132 (pack-revindex:
radix-sort the revindex, 2013-07-11), that takes about 200ms
for linux.git, and 20ms for git.git (those numbers are a few
years old but are still a good ballpark).
That makes it a good optimization for some cases (we can
save tens of seconds in git.git by having good locality of
delta access, for a 20ms cost), but a bad one for others
(e.g., right now "cat-file --batch-all-objects
--batch-check="%(objectname)" is 170ms in git.git, so adding
20ms to that is noticeable).
Hence this patch makes it an optional flag. You can't
actually do any interesting timings yet, as it's not plumbed
through to any user-facing tools like cat-file. That will
come in a later patch.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-08-11 02:15:49 +03:00
|
|
|
int for_each_object_in_pack(struct packed_git *p,
|
|
|
|
each_packed_object_fn cb, void *data,
|
|
|
|
enum for_each_object_flags flags)
|
2017-08-19 01:20:38 +03:00
|
|
|
{
|
|
|
|
uint32_t i;
|
|
|
|
int r = 0;
|
|
|
|
|
2019-04-05 21:04:24 +03:00
|
|
|
if (flags & FOR_EACH_OBJECT_PACK_ORDER) {
|
|
|
|
if (load_pack_revindex(p))
|
|
|
|
return -1;
|
|
|
|
}
|
for_each_packed_object: support iterating in pack-order
We currently iterate over objects within a pack in .idx
order, which uses the object hashes. That means that it
is effectively random with respect to the location of the
object within the pack. If you're going to access the actual
object data, there are two reasons to move linearly through
the pack itself:
1. It improves the locality of access in the packfile. In
the cold-cache case, this may mean fewer disk seeks, or
better usage of disk cache.
2. We store related deltas together in the packfile. Which
means that the delta base cache can operate much more
efficiently if we visit all of those related deltas in
sequence, as the earlier items are likely to still be
in the cache. Whereas if we visit the objects in
random order, our cache entries are much more likely to
have been evicted by unrelated deltas in the meantime.
So in general, if you're going to access the object contents
pack order is generally going to end up more efficient.
But if you're simply generating a list of object names, or
if you're going to end up sorting the result anyway, you're
better off just using the .idx order, as finding the pack
order means generating the in-memory pack-revindex.
According to the numbers in 8b8dfd5132 (pack-revindex:
radix-sort the revindex, 2013-07-11), that takes about 200ms
for linux.git, and 20ms for git.git (those numbers are a few
years old but are still a good ballpark).
That makes it a good optimization for some cases (we can
save tens of seconds in git.git by having good locality of
delta access, for a 20ms cost), but a bad one for others
(e.g., right now "cat-file --batch-all-objects
--batch-check="%(objectname)" is 170ms in git.git, so adding
20ms to that is noticeable).
Hence this patch makes it an optional flag. You can't
actually do any interesting timings yet, as it's not plumbed
through to any user-facing tools like cat-file. That will
come in a later patch.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-08-11 02:15:49 +03:00
|
|
|
|
2017-08-19 01:20:38 +03:00
|
|
|
for (i = 0; i < p->num_objects; i++) {
|
for_each_packed_object: support iterating in pack-order
We currently iterate over objects within a pack in .idx
order, which uses the object hashes. That means that it
is effectively random with respect to the location of the
object within the pack. If you're going to access the actual
object data, there are two reasons to move linearly through
the pack itself:
1. It improves the locality of access in the packfile. In
the cold-cache case, this may mean fewer disk seeks, or
better usage of disk cache.
2. We store related deltas together in the packfile. Which
means that the delta base cache can operate much more
efficiently if we visit all of those related deltas in
sequence, as the earlier items are likely to still be
in the cache. Whereas if we visit the objects in
random order, our cache entries are much more likely to
have been evicted by unrelated deltas in the meantime.
So in general, if you're going to access the object contents
pack order is generally going to end up more efficient.
But if you're simply generating a list of object names, or
if you're going to end up sorting the result anyway, you're
better off just using the .idx order, as finding the pack
order means generating the in-memory pack-revindex.
According to the numbers in 8b8dfd5132 (pack-revindex:
radix-sort the revindex, 2013-07-11), that takes about 200ms
for linux.git, and 20ms for git.git (those numbers are a few
years old but are still a good ballpark).
That makes it a good optimization for some cases (we can
save tens of seconds in git.git by having good locality of
delta access, for a 20ms cost), but a bad one for others
(e.g., right now "cat-file --batch-all-objects
--batch-check="%(objectname)" is 170ms in git.git, so adding
20ms to that is noticeable).
Hence this patch makes it an optional flag. You can't
actually do any interesting timings yet, as it's not plumbed
through to any user-facing tools like cat-file. That will
come in a later patch.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-08-11 02:15:49 +03:00
|
|
|
uint32_t pos;
|
2017-08-19 01:20:38 +03:00
|
|
|
struct object_id oid;
|
|
|
|
|
for_each_packed_object: support iterating in pack-order
We currently iterate over objects within a pack in .idx
order, which uses the object hashes. That means that it
is effectively random with respect to the location of the
object within the pack. If you're going to access the actual
object data, there are two reasons to move linearly through
the pack itself:
1. It improves the locality of access in the packfile. In
the cold-cache case, this may mean fewer disk seeks, or
better usage of disk cache.
2. We store related deltas together in the packfile. Which
means that the delta base cache can operate much more
efficiently if we visit all of those related deltas in
sequence, as the earlier items are likely to still be
in the cache. Whereas if we visit the objects in
random order, our cache entries are much more likely to
have been evicted by unrelated deltas in the meantime.
So in general, if you're going to access the object contents
pack order is generally going to end up more efficient.
But if you're simply generating a list of object names, or
if you're going to end up sorting the result anyway, you're
better off just using the .idx order, as finding the pack
order means generating the in-memory pack-revindex.
According to the numbers in 8b8dfd5132 (pack-revindex:
radix-sort the revindex, 2013-07-11), that takes about 200ms
for linux.git, and 20ms for git.git (those numbers are a few
years old but are still a good ballpark).
That makes it a good optimization for some cases (we can
save tens of seconds in git.git by having good locality of
delta access, for a 20ms cost), but a bad one for others
(e.g., right now "cat-file --batch-all-objects
--batch-check="%(objectname)" is 170ms in git.git, so adding
20ms to that is noticeable).
Hence this patch makes it an optional flag. You can't
actually do any interesting timings yet, as it's not plumbed
through to any user-facing tools like cat-file. That will
come in a later patch.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-08-11 02:15:49 +03:00
|
|
|
if (flags & FOR_EACH_OBJECT_PACK_ORDER)
|
|
|
|
pos = p->revindex[i].nr;
|
|
|
|
else
|
|
|
|
pos = i;
|
|
|
|
|
2020-02-24 07:27:36 +03:00
|
|
|
if (nth_packed_object_id(&oid, p, pos) < 0)
|
2017-08-19 01:20:38 +03:00
|
|
|
return error("unable to get sha1 of object %u in %s",
|
for_each_packed_object: support iterating in pack-order
We currently iterate over objects within a pack in .idx
order, which uses the object hashes. That means that it
is effectively random with respect to the location of the
object within the pack. If you're going to access the actual
object data, there are two reasons to move linearly through
the pack itself:
1. It improves the locality of access in the packfile. In
the cold-cache case, this may mean fewer disk seeks, or
better usage of disk cache.
2. We store related deltas together in the packfile. Which
means that the delta base cache can operate much more
efficiently if we visit all of those related deltas in
sequence, as the earlier items are likely to still be
in the cache. Whereas if we visit the objects in
random order, our cache entries are much more likely to
have been evicted by unrelated deltas in the meantime.
So in general, if you're going to access the object contents
pack order is generally going to end up more efficient.
But if you're simply generating a list of object names, or
if you're going to end up sorting the result anyway, you're
better off just using the .idx order, as finding the pack
order means generating the in-memory pack-revindex.
According to the numbers in 8b8dfd5132 (pack-revindex:
radix-sort the revindex, 2013-07-11), that takes about 200ms
for linux.git, and 20ms for git.git (those numbers are a few
years old but are still a good ballpark).
That makes it a good optimization for some cases (we can
save tens of seconds in git.git by having good locality of
delta access, for a 20ms cost), but a bad one for others
(e.g., right now "cat-file --batch-all-objects
--batch-check="%(objectname)" is 170ms in git.git, so adding
20ms to that is noticeable).
Hence this patch makes it an optional flag. You can't
actually do any interesting timings yet, as it's not plumbed
through to any user-facing tools like cat-file. That will
come in a later patch.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-08-11 02:15:49 +03:00
|
|
|
pos, p->pack_name);
|
2017-08-19 01:20:38 +03:00
|
|
|
|
for_each_packed_object: support iterating in pack-order
We currently iterate over objects within a pack in .idx
order, which uses the object hashes. That means that it
is effectively random with respect to the location of the
object within the pack. If you're going to access the actual
object data, there are two reasons to move linearly through
the pack itself:
1. It improves the locality of access in the packfile. In
the cold-cache case, this may mean fewer disk seeks, or
better usage of disk cache.
2. We store related deltas together in the packfile. Which
means that the delta base cache can operate much more
efficiently if we visit all of those related deltas in
sequence, as the earlier items are likely to still be
in the cache. Whereas if we visit the objects in
random order, our cache entries are much more likely to
have been evicted by unrelated deltas in the meantime.
So in general, if you're going to access the object contents
pack order is generally going to end up more efficient.
But if you're simply generating a list of object names, or
if you're going to end up sorting the result anyway, you're
better off just using the .idx order, as finding the pack
order means generating the in-memory pack-revindex.
According to the numbers in 8b8dfd5132 (pack-revindex:
radix-sort the revindex, 2013-07-11), that takes about 200ms
for linux.git, and 20ms for git.git (those numbers are a few
years old but are still a good ballpark).
That makes it a good optimization for some cases (we can
save tens of seconds in git.git by having good locality of
delta access, for a 20ms cost), but a bad one for others
(e.g., right now "cat-file --batch-all-objects
--batch-check="%(objectname)" is 170ms in git.git, so adding
20ms to that is noticeable).
Hence this patch makes it an optional flag. You can't
actually do any interesting timings yet, as it's not plumbed
through to any user-facing tools like cat-file. That will
come in a later patch.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-08-11 02:15:49 +03:00
|
|
|
r = cb(&oid, p, pos, data);
|
2017-08-19 01:20:38 +03:00
|
|
|
if (r)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return r;
|
|
|
|
}
|
|
|
|
|
2018-08-11 02:09:44 +03:00
|
|
|
int for_each_packed_object(each_packed_object_fn cb, void *data,
|
|
|
|
enum for_each_object_flags flags)
|
2017-08-19 01:20:38 +03:00
|
|
|
{
|
|
|
|
struct packed_git *p;
|
|
|
|
int r = 0;
|
|
|
|
int pack_errors = 0;
|
|
|
|
|
2018-03-23 20:45:20 +03:00
|
|
|
prepare_packed_git(the_repository);
|
2018-08-20 19:52:04 +03:00
|
|
|
for (p = get_all_packs(the_repository); p; p = p->next) {
|
2017-08-19 01:20:38 +03:00
|
|
|
if ((flags & FOR_EACH_OBJECT_LOCAL_ONLY) && !p->pack_local)
|
|
|
|
continue;
|
2017-12-05 19:58:44 +03:00
|
|
|
if ((flags & FOR_EACH_OBJECT_PROMISOR_ONLY) &&
|
|
|
|
!p->pack_promisor)
|
|
|
|
continue;
|
2017-08-19 01:20:38 +03:00
|
|
|
if (open_pack_index(p)) {
|
|
|
|
pack_errors = 1;
|
|
|
|
continue;
|
|
|
|
}
|
for_each_packed_object: support iterating in pack-order
We currently iterate over objects within a pack in .idx
order, which uses the object hashes. That means that it
is effectively random with respect to the location of the
object within the pack. If you're going to access the actual
object data, there are two reasons to move linearly through
the pack itself:
1. It improves the locality of access in the packfile. In
the cold-cache case, this may mean fewer disk seeks, or
better usage of disk cache.
2. We store related deltas together in the packfile. Which
means that the delta base cache can operate much more
efficiently if we visit all of those related deltas in
sequence, as the earlier items are likely to still be
in the cache. Whereas if we visit the objects in
random order, our cache entries are much more likely to
have been evicted by unrelated deltas in the meantime.
So in general, if you're going to access the object contents
pack order is generally going to end up more efficient.
But if you're simply generating a list of object names, or
if you're going to end up sorting the result anyway, you're
better off just using the .idx order, as finding the pack
order means generating the in-memory pack-revindex.
According to the numbers in 8b8dfd5132 (pack-revindex:
radix-sort the revindex, 2013-07-11), that takes about 200ms
for linux.git, and 20ms for git.git (those numbers are a few
years old but are still a good ballpark).
That makes it a good optimization for some cases (we can
save tens of seconds in git.git by having good locality of
delta access, for a 20ms cost), but a bad one for others
(e.g., right now "cat-file --batch-all-objects
--batch-check="%(objectname)" is 170ms in git.git, so adding
20ms to that is noticeable).
Hence this patch makes it an optional flag. You can't
actually do any interesting timings yet, as it's not plumbed
through to any user-facing tools like cat-file. That will
come in a later patch.
Signed-off-by: Jeff King <peff@peff.net>
Signed-off-by: Junio C Hamano <gitster@pobox.com>
2018-08-11 02:15:49 +03:00
|
|
|
r = for_each_object_in_pack(p, cb, data, flags);
|
2017-08-19 01:20:38 +03:00
|
|
|
if (r)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return r ? r : pack_errors;
|
|
|
|
}
|
2017-12-05 19:58:44 +03:00
|
|
|
|
|
|
|
static int add_promisor_object(const struct object_id *oid,
|
|
|
|
struct packed_git *pack,
|
|
|
|
uint32_t pos,
|
|
|
|
void *set_)
|
|
|
|
{
|
|
|
|
struct oidset *set = set_;
|
2018-06-29 04:21:51 +03:00
|
|
|
struct object *obj = parse_object(the_repository, oid);
|
2017-12-05 19:58:44 +03:00
|
|
|
if (!obj)
|
|
|
|
return 1;
|
|
|
|
|
|
|
|
oidset_insert(set, oid);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If this is a tree, commit, or tag, the objects it refers
|
2018-03-23 20:20:59 +03:00
|
|
|
* to are also promisor objects. (Blobs refer to no objects->)
|
2017-12-05 19:58:44 +03:00
|
|
|
*/
|
|
|
|
if (obj->type == OBJ_TREE) {
|
|
|
|
struct tree *tree = (struct tree *)obj;
|
|
|
|
struct tree_desc desc;
|
|
|
|
struct name_entry entry;
|
|
|
|
if (init_tree_desc_gently(&desc, tree->buffer, tree->size))
|
|
|
|
/*
|
|
|
|
* Error messages are given when packs are
|
|
|
|
* verified, so do not print any here.
|
|
|
|
*/
|
|
|
|
return 0;
|
|
|
|
while (tree_entry_gently(&desc, &entry))
|
2019-01-15 03:39:44 +03:00
|
|
|
oidset_insert(set, &entry.oid);
|
2017-12-05 19:58:44 +03:00
|
|
|
} else if (obj->type == OBJ_COMMIT) {
|
|
|
|
struct commit *commit = (struct commit *) obj;
|
|
|
|
struct commit_list *parents = commit->parents;
|
|
|
|
|
2018-04-06 22:09:38 +03:00
|
|
|
oidset_insert(set, get_commit_tree_oid(commit));
|
2017-12-05 19:58:44 +03:00
|
|
|
for (; parents; parents = parents->next)
|
|
|
|
oidset_insert(set, &parents->item->object.oid);
|
|
|
|
} else if (obj->type == OBJ_TAG) {
|
|
|
|
struct tag *tag = (struct tag *) obj;
|
2019-09-05 22:59:42 +03:00
|
|
|
oidset_insert(set, get_tagged_oid(tag));
|
2017-12-05 19:58:44 +03:00
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
int is_promisor_object(const struct object_id *oid)
|
|
|
|
{
|
|
|
|
static struct oidset promisor_objects;
|
|
|
|
static int promisor_objects_prepared;
|
|
|
|
|
|
|
|
if (!promisor_objects_prepared) {
|
2019-06-25 16:40:31 +03:00
|
|
|
if (has_promisor_remote()) {
|
2017-12-05 19:58:44 +03:00
|
|
|
for_each_packed_object(add_promisor_object,
|
|
|
|
&promisor_objects,
|
|
|
|
FOR_EACH_OBJECT_PROMISOR_ONLY);
|
|
|
|
}
|
|
|
|
promisor_objects_prepared = 1;
|
|
|
|
}
|
|
|
|
return oidset_contains(&promisor_objects, oid);
|
|
|
|
}
|