mirror of
https://github.com/git/git
synced 2024-10-30 14:03:28 +00:00
bf972896d7
When "cat-file --batch-all-objects" iterates over each object, it knows where to find each one. But when we look up details of the object, we don't use that information at all. This patch teaches it to use the pack/offset pair when we're iterating over objects in a pack. This yields a measurable speed improvement (timings on a fully packed clone of linux.git): Benchmark #1: ./git.old cat-file --batch-all-objects --unordered --batch-check="%(objecttype) %(objectname)" Time (mean ± σ): 8.128 s ± 0.118 s [User: 7.968 s, System: 0.156 s] Range (min … max): 8.007 s … 8.301 s 10 runs Benchmark #2: ./git.new cat-file --batch-all-objects --unordered --batch-check="%(objecttype) %(objectname)" Time (mean ± σ): 4.294 s ± 0.064 s [User: 4.167 s, System: 0.125 s] Range (min … max): 4.227 s … 4.457 s 10 runs Summary './git.new cat-file --batch-all-objects --unordered --batch-check="%(objecttype) %(objectname)"' ran 1.89 ± 0.04 times faster than './git.old cat-file --batch-all-objects --unordered --batch-check="%(objecttype) %(objectname)" The implementation is pretty simple: we just call packed_object_info() instead of oid_object_info_extended() when we can. Most of the changes are just plumbing the pack/offset pair through the callstack. There is one subtlety: replace lookups are not handled by packed_object_info(). But since those are disabled for --batch-all-objects, and since we'll only have pack info when that option is in effect, we don't have to worry about that. There are a few limitations to this optimization which we could address with further work: - I didn't bother recording when we found an object loose. Technically this could save us doing a fruitless lookup in the pack index. But opening and mmap-ing a loose object is so expensive in the first place that this doesn't matter much. And if your repository is large enough to care about per-object performance, most objects are going to be packed anyway. - This works only in --unordered mode. For the sorted mode, we'd have to record the pack/offset pair as part of our oid-collection. That's more code, plus at least 16 extra bytes of heap per object. It would probably still be a net win in runtime, but we'd need to measure. - For --batch, this still helps us with getting the object metadata, but we still do a from-scratch lookup for the object contents. This probably doesn't matter that much, because the lookup cost will be much smaller relative to the cost of actually unpacking and printing the objects. For small objects, we could probably swap out read_object_file() for using packed_object_info() with a "object_info.contentp" to get the contents. But we'd still need to deal with streaming for larger objects. A better path forward here is to teach the initial oid_object_info_extended() / packed_object_info() calls to retrieve the contents of smaller objects while they are already being accessed. That would save the extra lookup entirely. But it's a non-trivial feature to add to the object_info code, so I left it for now. Signed-off-by: Jeff King <peff@peff.net> Signed-off-by: Junio C Hamano <gitster@pobox.com>
745 lines
19 KiB
C
745 lines
19 KiB
C
/*
|
|
* GIT - The information manager from hell
|
|
*
|
|
* Copyright (C) Linus Torvalds, 2005
|
|
*/
|
|
#define USE_THE_INDEX_COMPATIBILITY_MACROS
|
|
#include "cache.h"
|
|
#include "config.h"
|
|
#include "builtin.h"
|
|
#include "diff.h"
|
|
#include "parse-options.h"
|
|
#include "userdiff.h"
|
|
#include "streaming.h"
|
|
#include "tree-walk.h"
|
|
#include "oid-array.h"
|
|
#include "packfile.h"
|
|
#include "object-store.h"
|
|
#include "promisor-remote.h"
|
|
|
|
struct batch_options {
|
|
int enabled;
|
|
int follow_symlinks;
|
|
int print_contents;
|
|
int buffer_output;
|
|
int all_objects;
|
|
int unordered;
|
|
int cmdmode; /* may be 'w' or 'c' for --filters or --textconv */
|
|
const char *format;
|
|
};
|
|
|
|
static const char *force_path;
|
|
|
|
static int filter_object(const char *path, unsigned mode,
|
|
const struct object_id *oid,
|
|
char **buf, unsigned long *size)
|
|
{
|
|
enum object_type type;
|
|
|
|
*buf = read_object_file(oid, &type, size);
|
|
if (!*buf)
|
|
return error(_("cannot read object %s '%s'"),
|
|
oid_to_hex(oid), path);
|
|
if ((type == OBJ_BLOB) && S_ISREG(mode)) {
|
|
struct strbuf strbuf = STRBUF_INIT;
|
|
struct checkout_metadata meta;
|
|
|
|
init_checkout_metadata(&meta, NULL, NULL, oid);
|
|
if (convert_to_working_tree(&the_index, path, *buf, *size, &strbuf, &meta)) {
|
|
free(*buf);
|
|
*size = strbuf.len;
|
|
*buf = strbuf_detach(&strbuf, NULL);
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int stream_blob(const struct object_id *oid)
|
|
{
|
|
if (stream_blob_to_fd(1, oid, NULL, 0))
|
|
die("unable to stream %s to stdout", oid_to_hex(oid));
|
|
return 0;
|
|
}
|
|
|
|
static int cat_one_file(int opt, const char *exp_type, const char *obj_name,
|
|
int unknown_type)
|
|
{
|
|
struct object_id oid;
|
|
enum object_type type;
|
|
char *buf;
|
|
unsigned long size;
|
|
struct object_context obj_context;
|
|
struct object_info oi = OBJECT_INFO_INIT;
|
|
struct strbuf sb = STRBUF_INIT;
|
|
unsigned flags = OBJECT_INFO_LOOKUP_REPLACE;
|
|
const char *path = force_path;
|
|
|
|
if (unknown_type)
|
|
flags |= OBJECT_INFO_ALLOW_UNKNOWN_TYPE;
|
|
|
|
if (get_oid_with_context(the_repository, obj_name,
|
|
GET_OID_RECORD_PATH,
|
|
&oid, &obj_context))
|
|
die("Not a valid object name %s", obj_name);
|
|
|
|
if (!path)
|
|
path = obj_context.path;
|
|
if (obj_context.mode == S_IFINVALID)
|
|
obj_context.mode = 0100644;
|
|
|
|
buf = NULL;
|
|
switch (opt) {
|
|
case 't':
|
|
oi.type_name = &sb;
|
|
if (oid_object_info_extended(the_repository, &oid, &oi, flags) < 0)
|
|
die("git cat-file: could not get object info");
|
|
if (sb.len) {
|
|
printf("%s\n", sb.buf);
|
|
strbuf_release(&sb);
|
|
return 0;
|
|
}
|
|
break;
|
|
|
|
case 's':
|
|
oi.sizep = &size;
|
|
if (oid_object_info_extended(the_repository, &oid, &oi, flags) < 0)
|
|
die("git cat-file: could not get object info");
|
|
printf("%"PRIuMAX"\n", (uintmax_t)size);
|
|
return 0;
|
|
|
|
case 'e':
|
|
return !has_object_file(&oid);
|
|
|
|
case 'w':
|
|
if (!path)
|
|
die("git cat-file --filters %s: <object> must be "
|
|
"<sha1:path>", obj_name);
|
|
|
|
if (filter_object(path, obj_context.mode,
|
|
&oid, &buf, &size))
|
|
return -1;
|
|
break;
|
|
|
|
case 'c':
|
|
if (!path)
|
|
die("git cat-file --textconv %s: <object> must be <sha1:path>",
|
|
obj_name);
|
|
|
|
if (textconv_object(the_repository, path, obj_context.mode,
|
|
&oid, 1, &buf, &size))
|
|
break;
|
|
/* else fallthrough */
|
|
|
|
case 'p':
|
|
type = oid_object_info(the_repository, &oid, NULL);
|
|
if (type < 0)
|
|
die("Not a valid object name %s", obj_name);
|
|
|
|
/* custom pretty-print here */
|
|
if (type == OBJ_TREE) {
|
|
const char *ls_args[3] = { NULL };
|
|
ls_args[0] = "ls-tree";
|
|
ls_args[1] = obj_name;
|
|
return cmd_ls_tree(2, ls_args, NULL);
|
|
}
|
|
|
|
if (type == OBJ_BLOB)
|
|
return stream_blob(&oid);
|
|
buf = read_object_file(&oid, &type, &size);
|
|
if (!buf)
|
|
die("Cannot read object %s", obj_name);
|
|
|
|
/* otherwise just spit out the data */
|
|
break;
|
|
|
|
case 0:
|
|
if (type_from_string(exp_type) == OBJ_BLOB) {
|
|
struct object_id blob_oid;
|
|
if (oid_object_info(the_repository, &oid, NULL) == OBJ_TAG) {
|
|
char *buffer = read_object_file(&oid, &type,
|
|
&size);
|
|
const char *target;
|
|
if (!skip_prefix(buffer, "object ", &target) ||
|
|
get_oid_hex(target, &blob_oid))
|
|
die("%s not a valid tag", oid_to_hex(&oid));
|
|
free(buffer);
|
|
} else
|
|
oidcpy(&blob_oid, &oid);
|
|
|
|
if (oid_object_info(the_repository, &blob_oid, NULL) == OBJ_BLOB)
|
|
return stream_blob(&blob_oid);
|
|
/*
|
|
* we attempted to dereference a tag to a blob
|
|
* and failed; there may be new dereference
|
|
* mechanisms this code is not aware of.
|
|
* fall-back to the usual case.
|
|
*/
|
|
}
|
|
buf = read_object_with_reference(the_repository,
|
|
&oid, exp_type, &size, NULL);
|
|
break;
|
|
|
|
default:
|
|
die("git cat-file: unknown option: %s", exp_type);
|
|
}
|
|
|
|
if (!buf)
|
|
die("git cat-file %s: bad file", obj_name);
|
|
|
|
write_or_die(1, buf, size);
|
|
free(buf);
|
|
free(obj_context.path);
|
|
return 0;
|
|
}
|
|
|
|
struct expand_data {
|
|
struct object_id oid;
|
|
enum object_type type;
|
|
unsigned long size;
|
|
off_t disk_size;
|
|
const char *rest;
|
|
struct object_id delta_base_oid;
|
|
|
|
/*
|
|
* If mark_query is true, we do not expand anything, but rather
|
|
* just mark the object_info with items we wish to query.
|
|
*/
|
|
int mark_query;
|
|
|
|
/*
|
|
* Whether to split the input on whitespace before feeding it to
|
|
* get_sha1; this is decided during the mark_query phase based on
|
|
* whether we have a %(rest) token in our format.
|
|
*/
|
|
int split_on_whitespace;
|
|
|
|
/*
|
|
* After a mark_query run, this object_info is set up to be
|
|
* passed to oid_object_info_extended. It will point to the data
|
|
* elements above, so you can retrieve the response from there.
|
|
*/
|
|
struct object_info info;
|
|
|
|
/*
|
|
* This flag will be true if the requested batch format and options
|
|
* don't require us to call oid_object_info, which can then be
|
|
* optimized out.
|
|
*/
|
|
unsigned skip_object_info : 1;
|
|
};
|
|
|
|
static int is_atom(const char *atom, const char *s, int slen)
|
|
{
|
|
int alen = strlen(atom);
|
|
return alen == slen && !memcmp(atom, s, alen);
|
|
}
|
|
|
|
static void expand_atom(struct strbuf *sb, const char *atom, int len,
|
|
void *vdata)
|
|
{
|
|
struct expand_data *data = vdata;
|
|
|
|
if (is_atom("objectname", atom, len)) {
|
|
if (!data->mark_query)
|
|
strbuf_addstr(sb, oid_to_hex(&data->oid));
|
|
} else if (is_atom("objecttype", atom, len)) {
|
|
if (data->mark_query)
|
|
data->info.typep = &data->type;
|
|
else
|
|
strbuf_addstr(sb, type_name(data->type));
|
|
} else if (is_atom("objectsize", atom, len)) {
|
|
if (data->mark_query)
|
|
data->info.sizep = &data->size;
|
|
else
|
|
strbuf_addf(sb, "%"PRIuMAX , (uintmax_t)data->size);
|
|
} else if (is_atom("objectsize:disk", atom, len)) {
|
|
if (data->mark_query)
|
|
data->info.disk_sizep = &data->disk_size;
|
|
else
|
|
strbuf_addf(sb, "%"PRIuMAX, (uintmax_t)data->disk_size);
|
|
} else if (is_atom("rest", atom, len)) {
|
|
if (data->mark_query)
|
|
data->split_on_whitespace = 1;
|
|
else if (data->rest)
|
|
strbuf_addstr(sb, data->rest);
|
|
} else if (is_atom("deltabase", atom, len)) {
|
|
if (data->mark_query)
|
|
data->info.delta_base_oid = &data->delta_base_oid;
|
|
else
|
|
strbuf_addstr(sb,
|
|
oid_to_hex(&data->delta_base_oid));
|
|
} else
|
|
die("unknown format element: %.*s", len, atom);
|
|
}
|
|
|
|
static size_t expand_format(struct strbuf *sb, const char *start, void *data)
|
|
{
|
|
const char *end;
|
|
|
|
if (*start != '(')
|
|
return 0;
|
|
end = strchr(start + 1, ')');
|
|
if (!end)
|
|
die("format element '%s' does not end in ')'", start);
|
|
|
|
expand_atom(sb, start + 1, end - start - 1, data);
|
|
|
|
return end - start + 1;
|
|
}
|
|
|
|
static void batch_write(struct batch_options *opt, const void *data, int len)
|
|
{
|
|
if (opt->buffer_output) {
|
|
if (fwrite(data, 1, len, stdout) != len)
|
|
die_errno("unable to write to stdout");
|
|
} else
|
|
write_or_die(1, data, len);
|
|
}
|
|
|
|
static void print_object_or_die(struct batch_options *opt, struct expand_data *data)
|
|
{
|
|
const struct object_id *oid = &data->oid;
|
|
|
|
assert(data->info.typep);
|
|
|
|
if (data->type == OBJ_BLOB) {
|
|
if (opt->buffer_output)
|
|
fflush(stdout);
|
|
if (opt->cmdmode) {
|
|
char *contents;
|
|
unsigned long size;
|
|
|
|
if (!data->rest)
|
|
die("missing path for '%s'", oid_to_hex(oid));
|
|
|
|
if (opt->cmdmode == 'w') {
|
|
if (filter_object(data->rest, 0100644, oid,
|
|
&contents, &size))
|
|
die("could not convert '%s' %s",
|
|
oid_to_hex(oid), data->rest);
|
|
} else if (opt->cmdmode == 'c') {
|
|
enum object_type type;
|
|
if (!textconv_object(the_repository,
|
|
data->rest, 0100644, oid,
|
|
1, &contents, &size))
|
|
contents = read_object_file(oid,
|
|
&type,
|
|
&size);
|
|
if (!contents)
|
|
die("could not convert '%s' %s",
|
|
oid_to_hex(oid), data->rest);
|
|
} else
|
|
BUG("invalid cmdmode: %c", opt->cmdmode);
|
|
batch_write(opt, contents, size);
|
|
free(contents);
|
|
} else {
|
|
stream_blob(oid);
|
|
}
|
|
}
|
|
else {
|
|
enum object_type type;
|
|
unsigned long size;
|
|
void *contents;
|
|
|
|
contents = read_object_file(oid, &type, &size);
|
|
if (!contents)
|
|
die("object %s disappeared", oid_to_hex(oid));
|
|
if (type != data->type)
|
|
die("object %s changed type!?", oid_to_hex(oid));
|
|
if (data->info.sizep && size != data->size)
|
|
die("object %s changed size!?", oid_to_hex(oid));
|
|
|
|
batch_write(opt, contents, size);
|
|
free(contents);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If "pack" is non-NULL, then "offset" is the byte offset within the pack from
|
|
* which the object may be accessed (though note that we may also rely on
|
|
* data->oid, too). If "pack" is NULL, then offset is ignored.
|
|
*/
|
|
static void batch_object_write(const char *obj_name,
|
|
struct strbuf *scratch,
|
|
struct batch_options *opt,
|
|
struct expand_data *data,
|
|
struct packed_git *pack,
|
|
off_t offset)
|
|
{
|
|
if (!data->skip_object_info) {
|
|
int ret;
|
|
|
|
if (pack)
|
|
ret = packed_object_info(the_repository, pack, offset,
|
|
&data->info);
|
|
else
|
|
ret = oid_object_info_extended(the_repository,
|
|
&data->oid, &data->info,
|
|
OBJECT_INFO_LOOKUP_REPLACE);
|
|
if (ret < 0) {
|
|
printf("%s missing\n",
|
|
obj_name ? obj_name : oid_to_hex(&data->oid));
|
|
fflush(stdout);
|
|
return;
|
|
}
|
|
}
|
|
|
|
strbuf_reset(scratch);
|
|
strbuf_expand(scratch, opt->format, expand_format, data);
|
|
strbuf_addch(scratch, '\n');
|
|
batch_write(opt, scratch->buf, scratch->len);
|
|
|
|
if (opt->print_contents) {
|
|
print_object_or_die(opt, data);
|
|
batch_write(opt, "\n", 1);
|
|
}
|
|
}
|
|
|
|
static void batch_one_object(const char *obj_name,
|
|
struct strbuf *scratch,
|
|
struct batch_options *opt,
|
|
struct expand_data *data)
|
|
{
|
|
struct object_context ctx;
|
|
int flags = opt->follow_symlinks ? GET_OID_FOLLOW_SYMLINKS : 0;
|
|
enum get_oid_result result;
|
|
|
|
result = get_oid_with_context(the_repository, obj_name,
|
|
flags, &data->oid, &ctx);
|
|
if (result != FOUND) {
|
|
switch (result) {
|
|
case MISSING_OBJECT:
|
|
printf("%s missing\n", obj_name);
|
|
break;
|
|
case SHORT_NAME_AMBIGUOUS:
|
|
printf("%s ambiguous\n", obj_name);
|
|
break;
|
|
case DANGLING_SYMLINK:
|
|
printf("dangling %"PRIuMAX"\n%s\n",
|
|
(uintmax_t)strlen(obj_name), obj_name);
|
|
break;
|
|
case SYMLINK_LOOP:
|
|
printf("loop %"PRIuMAX"\n%s\n",
|
|
(uintmax_t)strlen(obj_name), obj_name);
|
|
break;
|
|
case NOT_DIR:
|
|
printf("notdir %"PRIuMAX"\n%s\n",
|
|
(uintmax_t)strlen(obj_name), obj_name);
|
|
break;
|
|
default:
|
|
BUG("unknown get_sha1_with_context result %d\n",
|
|
result);
|
|
break;
|
|
}
|
|
fflush(stdout);
|
|
return;
|
|
}
|
|
|
|
if (ctx.mode == 0) {
|
|
printf("symlink %"PRIuMAX"\n%s\n",
|
|
(uintmax_t)ctx.symlink_path.len,
|
|
ctx.symlink_path.buf);
|
|
fflush(stdout);
|
|
return;
|
|
}
|
|
|
|
batch_object_write(obj_name, scratch, opt, data, NULL, 0);
|
|
}
|
|
|
|
struct object_cb_data {
|
|
struct batch_options *opt;
|
|
struct expand_data *expand;
|
|
struct oidset *seen;
|
|
struct strbuf *scratch;
|
|
};
|
|
|
|
static int batch_object_cb(const struct object_id *oid, void *vdata)
|
|
{
|
|
struct object_cb_data *data = vdata;
|
|
oidcpy(&data->expand->oid, oid);
|
|
batch_object_write(NULL, data->scratch, data->opt, data->expand,
|
|
NULL, 0);
|
|
return 0;
|
|
}
|
|
|
|
static int collect_loose_object(const struct object_id *oid,
|
|
const char *path,
|
|
void *data)
|
|
{
|
|
oid_array_append(data, oid);
|
|
return 0;
|
|
}
|
|
|
|
static int collect_packed_object(const struct object_id *oid,
|
|
struct packed_git *pack,
|
|
uint32_t pos,
|
|
void *data)
|
|
{
|
|
oid_array_append(data, oid);
|
|
return 0;
|
|
}
|
|
|
|
static int batch_unordered_object(const struct object_id *oid,
|
|
struct packed_git *pack, off_t offset,
|
|
void *vdata)
|
|
{
|
|
struct object_cb_data *data = vdata;
|
|
|
|
if (oidset_insert(data->seen, oid))
|
|
return 0;
|
|
|
|
oidcpy(&data->expand->oid, oid);
|
|
batch_object_write(NULL, data->scratch, data->opt, data->expand,
|
|
pack, offset);
|
|
return 0;
|
|
}
|
|
|
|
static int batch_unordered_loose(const struct object_id *oid,
|
|
const char *path,
|
|
void *data)
|
|
{
|
|
return batch_unordered_object(oid, NULL, 0, data);
|
|
}
|
|
|
|
static int batch_unordered_packed(const struct object_id *oid,
|
|
struct packed_git *pack,
|
|
uint32_t pos,
|
|
void *data)
|
|
{
|
|
return batch_unordered_object(oid, pack,
|
|
nth_packed_object_offset(pack, pos),
|
|
data);
|
|
}
|
|
|
|
static int batch_objects(struct batch_options *opt)
|
|
{
|
|
struct strbuf input = STRBUF_INIT;
|
|
struct strbuf output = STRBUF_INIT;
|
|
struct expand_data data;
|
|
int save_warning;
|
|
int retval = 0;
|
|
|
|
if (!opt->format)
|
|
opt->format = "%(objectname) %(objecttype) %(objectsize)";
|
|
|
|
/*
|
|
* Expand once with our special mark_query flag, which will prime the
|
|
* object_info to be handed to oid_object_info_extended for each
|
|
* object.
|
|
*/
|
|
memset(&data, 0, sizeof(data));
|
|
data.mark_query = 1;
|
|
strbuf_expand(&output, opt->format, expand_format, &data);
|
|
data.mark_query = 0;
|
|
strbuf_release(&output);
|
|
if (opt->cmdmode)
|
|
data.split_on_whitespace = 1;
|
|
|
|
/*
|
|
* If we are printing out the object, then always fill in the type,
|
|
* since we will want to decide whether or not to stream.
|
|
*/
|
|
if (opt->print_contents)
|
|
data.info.typep = &data.type;
|
|
|
|
if (opt->all_objects) {
|
|
struct object_cb_data cb;
|
|
struct object_info empty = OBJECT_INFO_INIT;
|
|
|
|
if (!memcmp(&data.info, &empty, sizeof(empty)))
|
|
data.skip_object_info = 1;
|
|
|
|
if (has_promisor_remote())
|
|
warning("This repository uses promisor remotes. Some objects may not be loaded.");
|
|
|
|
read_replace_refs = 0;
|
|
|
|
cb.opt = opt;
|
|
cb.expand = &data;
|
|
cb.scratch = &output;
|
|
|
|
if (opt->unordered) {
|
|
struct oidset seen = OIDSET_INIT;
|
|
|
|
cb.seen = &seen;
|
|
|
|
for_each_loose_object(batch_unordered_loose, &cb, 0);
|
|
for_each_packed_object(batch_unordered_packed, &cb,
|
|
FOR_EACH_OBJECT_PACK_ORDER);
|
|
|
|
oidset_clear(&seen);
|
|
} else {
|
|
struct oid_array sa = OID_ARRAY_INIT;
|
|
|
|
for_each_loose_object(collect_loose_object, &sa, 0);
|
|
for_each_packed_object(collect_packed_object, &sa, 0);
|
|
|
|
oid_array_for_each_unique(&sa, batch_object_cb, &cb);
|
|
|
|
oid_array_clear(&sa);
|
|
}
|
|
|
|
strbuf_release(&output);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* We are going to call get_sha1 on a potentially very large number of
|
|
* objects. In most large cases, these will be actual object sha1s. The
|
|
* cost to double-check that each one is not also a ref (just so we can
|
|
* warn) ends up dwarfing the actual cost of the object lookups
|
|
* themselves. We can work around it by just turning off the warning.
|
|
*/
|
|
save_warning = warn_on_object_refname_ambiguity;
|
|
warn_on_object_refname_ambiguity = 0;
|
|
|
|
while (strbuf_getline(&input, stdin) != EOF) {
|
|
if (data.split_on_whitespace) {
|
|
/*
|
|
* Split at first whitespace, tying off the beginning
|
|
* of the string and saving the remainder (or NULL) in
|
|
* data.rest.
|
|
*/
|
|
char *p = strpbrk(input.buf, " \t");
|
|
if (p) {
|
|
while (*p && strchr(" \t", *p))
|
|
*p++ = '\0';
|
|
}
|
|
data.rest = p;
|
|
}
|
|
|
|
batch_one_object(input.buf, &output, opt, &data);
|
|
}
|
|
|
|
strbuf_release(&input);
|
|
strbuf_release(&output);
|
|
warn_on_object_refname_ambiguity = save_warning;
|
|
return retval;
|
|
}
|
|
|
|
static const char * const cat_file_usage[] = {
|
|
N_("git cat-file (-t [--allow-unknown-type] | -s [--allow-unknown-type] | -e | -p | <type> | --textconv | --filters) [--path=<path>] <object>"),
|
|
N_("git cat-file (--batch[=<format>] | --batch-check[=<format>]) [--follow-symlinks] [--textconv | --filters]"),
|
|
NULL
|
|
};
|
|
|
|
static int git_cat_file_config(const char *var, const char *value, void *cb)
|
|
{
|
|
if (userdiff_config(var, value) < 0)
|
|
return -1;
|
|
|
|
return git_default_config(var, value, cb);
|
|
}
|
|
|
|
static int batch_option_callback(const struct option *opt,
|
|
const char *arg,
|
|
int unset)
|
|
{
|
|
struct batch_options *bo = opt->value;
|
|
|
|
BUG_ON_OPT_NEG(unset);
|
|
|
|
if (bo->enabled) {
|
|
return error(_("only one batch option may be specified"));
|
|
}
|
|
|
|
bo->enabled = 1;
|
|
bo->print_contents = !strcmp(opt->long_name, "batch");
|
|
bo->format = arg;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int cmd_cat_file(int argc, const char **argv, const char *prefix)
|
|
{
|
|
int opt = 0;
|
|
const char *exp_type = NULL, *obj_name = NULL;
|
|
struct batch_options batch = {0};
|
|
int unknown_type = 0;
|
|
|
|
const struct option options[] = {
|
|
OPT_GROUP(N_("<type> can be one of: blob, tree, commit, tag")),
|
|
OPT_CMDMODE('t', NULL, &opt, N_("show object type"), 't'),
|
|
OPT_CMDMODE('s', NULL, &opt, N_("show object size"), 's'),
|
|
OPT_CMDMODE('e', NULL, &opt,
|
|
N_("exit with zero when there's no error"), 'e'),
|
|
OPT_CMDMODE('p', NULL, &opt, N_("pretty-print object's content"), 'p'),
|
|
OPT_CMDMODE(0, "textconv", &opt,
|
|
N_("for blob objects, run textconv on object's content"), 'c'),
|
|
OPT_CMDMODE(0, "filters", &opt,
|
|
N_("for blob objects, run filters on object's content"), 'w'),
|
|
OPT_STRING(0, "path", &force_path, N_("blob"),
|
|
N_("use a specific path for --textconv/--filters")),
|
|
OPT_BOOL(0, "allow-unknown-type", &unknown_type,
|
|
N_("allow -s and -t to work with broken/corrupt objects")),
|
|
OPT_BOOL(0, "buffer", &batch.buffer_output, N_("buffer --batch output")),
|
|
OPT_CALLBACK_F(0, "batch", &batch, "format",
|
|
N_("show info and content of objects fed from the standard input"),
|
|
PARSE_OPT_OPTARG | PARSE_OPT_NONEG,
|
|
batch_option_callback),
|
|
OPT_CALLBACK_F(0, "batch-check", &batch, "format",
|
|
N_("show info about objects fed from the standard input"),
|
|
PARSE_OPT_OPTARG | PARSE_OPT_NONEG,
|
|
batch_option_callback),
|
|
OPT_BOOL(0, "follow-symlinks", &batch.follow_symlinks,
|
|
N_("follow in-tree symlinks (used with --batch or --batch-check)")),
|
|
OPT_BOOL(0, "batch-all-objects", &batch.all_objects,
|
|
N_("show all objects with --batch or --batch-check")),
|
|
OPT_BOOL(0, "unordered", &batch.unordered,
|
|
N_("do not order --batch-all-objects output")),
|
|
OPT_END()
|
|
};
|
|
|
|
git_config(git_cat_file_config, NULL);
|
|
|
|
batch.buffer_output = -1;
|
|
argc = parse_options(argc, argv, prefix, options, cat_file_usage, 0);
|
|
|
|
if (opt) {
|
|
if (batch.enabled && (opt == 'c' || opt == 'w'))
|
|
batch.cmdmode = opt;
|
|
else if (argc == 1)
|
|
obj_name = argv[0];
|
|
else
|
|
usage_with_options(cat_file_usage, options);
|
|
}
|
|
if (!opt && !batch.enabled) {
|
|
if (argc == 2) {
|
|
exp_type = argv[0];
|
|
obj_name = argv[1];
|
|
} else
|
|
usage_with_options(cat_file_usage, options);
|
|
}
|
|
if (batch.enabled) {
|
|
if (batch.cmdmode != opt || argc)
|
|
usage_with_options(cat_file_usage, options);
|
|
if (batch.cmdmode && batch.all_objects)
|
|
die("--batch-all-objects cannot be combined with "
|
|
"--textconv nor with --filters");
|
|
}
|
|
|
|
if ((batch.follow_symlinks || batch.all_objects) && !batch.enabled) {
|
|
usage_with_options(cat_file_usage, options);
|
|
}
|
|
|
|
if (force_path && opt != 'c' && opt != 'w') {
|
|
error("--path=<path> needs --textconv or --filters");
|
|
usage_with_options(cat_file_usage, options);
|
|
}
|
|
|
|
if (force_path && batch.enabled) {
|
|
error("--path=<path> incompatible with --batch");
|
|
usage_with_options(cat_file_usage, options);
|
|
}
|
|
|
|
if (batch.buffer_output < 0)
|
|
batch.buffer_output = batch.all_objects;
|
|
|
|
if (batch.enabled)
|
|
return batch_objects(&batch);
|
|
|
|
if (unknown_type && opt != 't' && opt != 's')
|
|
die("git cat-file --allow-unknown-type: use with -s or -t");
|
|
return cat_one_file(opt, exp_type, obj_name, unknown_type);
|
|
}
|