Merge branch 'maint'
[git/platforms.git] / builtin-read-tree.c
blob5a09e17f1ad4fe19cf4e7cc29235d14d35405168
1 /*
2 * GIT - The information manager from hell
4 * Copyright (C) Linus Torvalds, 2005
5 */
7 #include "cache.h"
8 #include "object.h"
9 #include "tree.h"
10 #include "tree-walk.h"
11 #include "cache-tree.h"
12 #include "unpack-trees.h"
13 #include "dir.h"
14 #include "builtin.h"
16 static int nr_trees;
17 static struct tree *trees[MAX_UNPACK_TREES];
19 static int list_tree(unsigned char *sha1)
21 struct tree *tree;
23 if (nr_trees >= MAX_UNPACK_TREES)
24 die("I cannot read more than %d trees", MAX_UNPACK_TREES);
25 tree = parse_tree_indirect(sha1);
26 if (!tree)
27 return -1;
28 trees[nr_trees++] = tree;
29 return 0;
32 static int read_cache_unmerged(void)
34 int i;
35 struct cache_entry **dst;
36 struct cache_entry *last = NULL;
38 read_cache();
39 dst = active_cache;
40 for (i = 0; i < active_nr; i++) {
41 struct cache_entry *ce = active_cache[i];
42 if (ce_stage(ce)) {
43 remove_name_hash(ce);
44 if (last && !strcmp(ce->name, last->name))
45 continue;
46 cache_tree_invalidate_path(active_cache_tree, ce->name);
47 last = ce;
48 continue;
50 *dst++ = ce;
52 active_nr = dst - active_cache;
53 return !!last;
56 static void prime_cache_tree_rec(struct cache_tree *it, struct tree *tree)
58 struct tree_desc desc;
59 struct name_entry entry;
60 int cnt;
62 hashcpy(it->sha1, tree->object.sha1);
63 init_tree_desc(&desc, tree->buffer, tree->size);
64 cnt = 0;
65 while (tree_entry(&desc, &entry)) {
66 if (!S_ISDIR(entry.mode))
67 cnt++;
68 else {
69 struct cache_tree_sub *sub;
70 struct tree *subtree = lookup_tree(entry.sha1);
71 if (!subtree->object.parsed)
72 parse_tree(subtree);
73 sub = cache_tree_sub(it, entry.path);
74 sub->cache_tree = cache_tree();
75 prime_cache_tree_rec(sub->cache_tree, subtree);
76 cnt += sub->cache_tree->entry_count;
79 it->entry_count = cnt;
82 static void prime_cache_tree(void)
84 if (!nr_trees)
85 return;
86 active_cache_tree = cache_tree();
87 prime_cache_tree_rec(active_cache_tree, trees[0]);
91 static const char read_tree_usage[] = "git-read-tree (<sha> | [[-m [--trivial] [--aggressive] | --reset | --prefix=<prefix>] [-u | -i]] [--exclude-per-directory=<gitignore>] [--index-output=<file>] <sha1> [<sha2> [<sha3>]])";
93 static struct lock_file lock_file;
95 int cmd_read_tree(int argc, const char **argv, const char *unused_prefix)
97 int i, newfd, stage = 0;
98 unsigned char sha1[20];
99 struct tree_desc t[MAX_UNPACK_TREES];
100 struct unpack_trees_options opts;
102 memset(&opts, 0, sizeof(opts));
103 opts.head_idx = -1;
104 opts.src_index = &the_index;
105 opts.dst_index = &the_index;
107 git_config(git_default_config, NULL);
109 newfd = hold_locked_index(&lock_file, 1);
111 for (i = 1; i < argc; i++) {
112 const char *arg = argv[i];
114 /* "-u" means "update", meaning that a merge will update
115 * the working tree.
117 if (!strcmp(arg, "-u")) {
118 opts.update = 1;
119 continue;
122 if (!strcmp(arg, "-v")) {
123 opts.verbose_update = 1;
124 continue;
127 /* "-i" means "index only", meaning that a merge will
128 * not even look at the working tree.
130 if (!strcmp(arg, "-i")) {
131 opts.index_only = 1;
132 continue;
135 if (!prefixcmp(arg, "--index-output=")) {
136 set_alternate_index_output(arg + 15);
137 continue;
140 /* "--prefix=<subdirectory>/" means keep the current index
141 * entries and put the entries from the tree under the
142 * given subdirectory.
144 if (!prefixcmp(arg, "--prefix=")) {
145 if (stage || opts.merge || opts.prefix)
146 usage(read_tree_usage);
147 opts.prefix = arg + 9;
148 opts.merge = 1;
149 stage = 1;
150 if (read_cache_unmerged())
151 die("you need to resolve your current index first");
152 continue;
155 /* This differs from "-m" in that we'll silently ignore
156 * unmerged entries and overwrite working tree files that
157 * correspond to them.
159 if (!strcmp(arg, "--reset")) {
160 if (stage || opts.merge || opts.prefix)
161 usage(read_tree_usage);
162 opts.reset = 1;
163 opts.merge = 1;
164 stage = 1;
165 read_cache_unmerged();
166 continue;
169 if (!strcmp(arg, "--trivial")) {
170 opts.trivial_merges_only = 1;
171 continue;
174 if (!strcmp(arg, "--aggressive")) {
175 opts.aggressive = 1;
176 continue;
179 /* "-m" stands for "merge", meaning we start in stage 1 */
180 if (!strcmp(arg, "-m")) {
181 if (stage || opts.merge || opts.prefix)
182 usage(read_tree_usage);
183 if (read_cache_unmerged())
184 die("you need to resolve your current index first");
185 stage = 1;
186 opts.merge = 1;
187 continue;
190 if (!prefixcmp(arg, "--exclude-per-directory=")) {
191 struct dir_struct *dir;
193 if (opts.dir)
194 die("more than one --exclude-per-directory are given.");
196 dir = xcalloc(1, sizeof(*opts.dir));
197 dir->show_ignored = 1;
198 dir->exclude_per_dir = arg + 24;
199 opts.dir = dir;
200 /* We do not need to nor want to do read-directory
201 * here; we are merely interested in reusing the
202 * per directory ignore stack mechanism.
204 continue;
207 /* using -u and -i at the same time makes no sense */
208 if (1 < opts.index_only + opts.update)
209 usage(read_tree_usage);
211 if (get_sha1(arg, sha1))
212 die("Not a valid object name %s", arg);
213 if (list_tree(sha1) < 0)
214 die("failed to unpack tree object %s", arg);
215 stage++;
217 if ((opts.update||opts.index_only) && !opts.merge)
218 usage(read_tree_usage);
219 if ((opts.dir && !opts.update))
220 die("--exclude-per-directory is meaningless unless -u");
222 if (opts.merge) {
223 if (stage < 2)
224 die("just how do you expect me to merge %d trees?", stage-1);
225 switch (stage - 1) {
226 case 1:
227 opts.fn = opts.prefix ? bind_merge : oneway_merge;
228 break;
229 case 2:
230 opts.fn = twoway_merge;
231 break;
232 case 3:
233 default:
234 opts.fn = threeway_merge;
235 cache_tree_free(&active_cache_tree);
236 break;
239 if (stage - 1 >= 3)
240 opts.head_idx = stage - 2;
241 else
242 opts.head_idx = 1;
245 for (i = 0; i < nr_trees; i++) {
246 struct tree *tree = trees[i];
247 parse_tree(tree);
248 init_tree_desc(t+i, tree->buffer, tree->size);
250 if (unpack_trees(nr_trees, t, &opts))
251 return 128;
254 * When reading only one tree (either the most basic form,
255 * "-m ent" or "--reset ent" form), we can obtain a fully
256 * valid cache-tree because the index must match exactly
257 * what came from the tree.
259 if (nr_trees && !opts.prefix && (!opts.merge || (stage == 2))) {
260 cache_tree_free(&active_cache_tree);
261 prime_cache_tree();
264 if (write_cache(newfd, active_cache, active_nr) ||
265 commit_locked_index(&lock_file))
266 die("unable to write new index file");
267 return 0;