1 //===- RewriteRope.cpp - Rope specialized for rewriter --------------------===//
3 // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4 // See https://llvm.org/LICENSE.txt for license information.
5 // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
7 //===----------------------------------------------------------------------===//
9 // This file implements the RewriteRope class, which is a powerful string.
11 //===----------------------------------------------------------------------===//
13 #include "clang/Rewrite/Core/RewriteRope.h"
14 #include "clang/Basic/LLVM.h"
15 #include "llvm/Support/Casting.h"
20 using namespace clang
;
22 /// RewriteRope is a "strong" string class, designed to make insertions and
23 /// deletions in the middle of the string nearly constant time (really, they are
24 /// O(log N), but with a very low constant factor).
26 /// The implementation of this datastructure is a conceptual linear sequence of
27 /// RopePiece elements. Each RopePiece represents a view on a separately
28 /// allocated and reference counted string. This means that splitting a very
29 /// long string can be done in constant time by splitting a RopePiece that
30 /// references the whole string into two rope pieces that reference each half.
31 /// Once split, another string can be inserted in between the two halves by
32 /// inserting a RopePiece in between the two others. All of this is very
33 /// inexpensive: it takes time proportional to the number of RopePieces, not the
34 /// length of the strings they represent.
36 /// While a linear sequences of RopePieces is the conceptual model, the actual
37 /// implementation captures them in an adapted B+ Tree. Using a B+ tree (which
38 /// is a tree that keeps the values in the leaves and has where each node
39 /// contains a reasonable number of pointers to children/values) allows us to
40 /// maintain efficient operation when the RewriteRope contains a *huge* number
41 /// of RopePieces. The basic idea of the B+ Tree is that it allows us to find
42 /// the RopePiece corresponding to some offset very efficiently, and it
43 /// automatically balances itself on insertions of RopePieces (which can happen
44 /// for both insertions and erases of string ranges).
46 /// The one wrinkle on the theory is that we don't attempt to keep the tree
47 /// properly balanced when erases happen. Erases of string data can both insert
48 /// new RopePieces (e.g. when the middle of some other rope piece is deleted,
49 /// which results in two rope pieces, which is just like an insert) or it can
50 /// reduce the number of RopePieces maintained by the B+Tree. In the case when
51 /// the number of RopePieces is reduced, we don't attempt to maintain the
52 /// standard 'invariant' that each node in the tree contains at least
53 /// 'WidthFactor' children/values. For our use cases, this doesn't seem to
56 /// The implementation below is primarily implemented in terms of three classes:
57 /// RopePieceBTreeNode - Common base class for:
59 /// RopePieceBTreeLeaf - Directly manages up to '2*WidthFactor' RopePiece
60 /// nodes. This directly represents a chunk of the string with those
61 /// RopePieces concatenated.
62 /// RopePieceBTreeInterior - An interior node in the B+ Tree, which manages
63 /// up to '2*WidthFactor' other nodes in the tree.
67 //===----------------------------------------------------------------------===//
68 // RopePieceBTreeNode Class
69 //===----------------------------------------------------------------------===//
71 /// RopePieceBTreeNode - Common base class of RopePieceBTreeLeaf and
72 /// RopePieceBTreeInterior. This provides some 'virtual' dispatching methods
73 /// and a flag that determines which subclass the instance is. Also
74 /// important, this node knows the full extend of the node, including any
75 /// children that it has. This allows efficient skipping over entire subtrees
76 /// when looking for an offset in the BTree.
77 class RopePieceBTreeNode
{
79 /// WidthFactor - This controls the number of K/V slots held in the BTree:
80 /// how wide it is. Each level of the BTree is guaranteed to have at least
81 /// 'WidthFactor' elements in it (either ropepieces or children), (except
82 /// the root, which may have less) and may have at most 2*WidthFactor
84 enum { WidthFactor
= 8 };
86 /// Size - This is the number of bytes of file this node (including any
87 /// potential children) covers.
90 /// IsLeaf - True if this is an instance of RopePieceBTreeLeaf, false if it
91 /// is an instance of RopePieceBTreeInterior.
94 RopePieceBTreeNode(bool isLeaf
) : IsLeaf(isLeaf
) {}
95 ~RopePieceBTreeNode() = default;
98 bool isLeaf() const { return IsLeaf
; }
99 unsigned size() const { return Size
; }
103 /// split - Split the range containing the specified offset so that we are
104 /// guaranteed that there is a place to do an insertion at the specified
105 /// offset. The offset is relative, so "0" is the start of the node.
107 /// If there is no space in this subtree for the extra piece, the extra tree
108 /// node is returned and must be inserted into a parent.
109 RopePieceBTreeNode
*split(unsigned Offset
);
111 /// insert - Insert the specified ropepiece into this tree node at the
112 /// specified offset. The offset is relative, so "0" is the start of the
115 /// If there is no space in this subtree for the extra piece, the extra tree
116 /// node is returned and must be inserted into a parent.
117 RopePieceBTreeNode
*insert(unsigned Offset
, const RopePiece
&R
);
119 /// erase - Remove NumBytes from this node at the specified offset. We are
120 /// guaranteed that there is a split at Offset.
121 void erase(unsigned Offset
, unsigned NumBytes
);
124 //===----------------------------------------------------------------------===//
125 // RopePieceBTreeLeaf Class
126 //===----------------------------------------------------------------------===//
128 /// RopePieceBTreeLeaf - Directly manages up to '2*WidthFactor' RopePiece
129 /// nodes. This directly represents a chunk of the string with those
130 /// RopePieces concatenated. Since this is a B+Tree, all values (in this case
131 /// instances of RopePiece) are stored in leaves like this. To make iteration
132 /// over the leaves efficient, they maintain a singly linked list through the
133 /// NextLeaf field. This allows the B+Tree forward iterator to be constant
134 /// time for all increments.
135 class RopePieceBTreeLeaf
: public RopePieceBTreeNode
{
136 /// NumPieces - This holds the number of rope pieces currently active in the
138 unsigned char NumPieces
= 0;
140 /// Pieces - This tracks the file chunks currently in this leaf.
141 RopePiece Pieces
[2*WidthFactor
];
143 /// NextLeaf - This is a pointer to the next leaf in the tree, allowing
144 /// efficient in-order forward iteration of the tree without traversal.
145 RopePieceBTreeLeaf
**PrevLeaf
= nullptr;
146 RopePieceBTreeLeaf
*NextLeaf
= nullptr;
149 RopePieceBTreeLeaf() : RopePieceBTreeNode(true) {}
151 ~RopePieceBTreeLeaf() {
152 if (PrevLeaf
|| NextLeaf
)
153 removeFromLeafInOrder();
157 bool isFull() const { return NumPieces
== 2*WidthFactor
; }
159 /// clear - Remove all rope pieces from this leaf.
162 Pieces
[--NumPieces
] = RopePiece();
166 unsigned getNumPieces() const { return NumPieces
; }
168 const RopePiece
&getPiece(unsigned i
) const {
169 assert(i
< getNumPieces() && "Invalid piece ID");
173 const RopePieceBTreeLeaf
*getNextLeafInOrder() const { return NextLeaf
; }
175 void insertAfterLeafInOrder(RopePieceBTreeLeaf
*Node
) {
176 assert(!PrevLeaf
&& !NextLeaf
&& "Already in ordering");
178 NextLeaf
= Node
->NextLeaf
;
180 NextLeaf
->PrevLeaf
= &NextLeaf
;
181 PrevLeaf
= &Node
->NextLeaf
;
182 Node
->NextLeaf
= this;
185 void removeFromLeafInOrder() {
187 *PrevLeaf
= NextLeaf
;
189 NextLeaf
->PrevLeaf
= PrevLeaf
;
190 } else if (NextLeaf
) {
191 NextLeaf
->PrevLeaf
= nullptr;
195 /// FullRecomputeSizeLocally - This method recomputes the 'Size' field by
196 /// summing the size of all RopePieces.
197 void FullRecomputeSizeLocally() {
199 for (unsigned i
= 0, e
= getNumPieces(); i
!= e
; ++i
)
200 Size
+= getPiece(i
).size();
203 /// split - Split the range containing the specified offset so that we are
204 /// guaranteed that there is a place to do an insertion at the specified
205 /// offset. The offset is relative, so "0" is the start of the node.
207 /// If there is no space in this subtree for the extra piece, the extra tree
208 /// node is returned and must be inserted into a parent.
209 RopePieceBTreeNode
*split(unsigned Offset
);
211 /// insert - Insert the specified ropepiece into this tree node at the
212 /// specified offset. The offset is relative, so "0" is the start of the
215 /// If there is no space in this subtree for the extra piece, the extra tree
216 /// node is returned and must be inserted into a parent.
217 RopePieceBTreeNode
*insert(unsigned Offset
, const RopePiece
&R
);
219 /// erase - Remove NumBytes from this node at the specified offset. We are
220 /// guaranteed that there is a split at Offset.
221 void erase(unsigned Offset
, unsigned NumBytes
);
223 static bool classof(const RopePieceBTreeNode
*N
) {
230 /// split - Split the range containing the specified offset so that we are
231 /// guaranteed that there is a place to do an insertion at the specified
232 /// offset. The offset is relative, so "0" is the start of the node.
234 /// If there is no space in this subtree for the extra piece, the extra tree
235 /// node is returned and must be inserted into a parent.
236 RopePieceBTreeNode
*RopePieceBTreeLeaf::split(unsigned Offset
) {
237 // Find the insertion point. We are guaranteed that there is a split at the
238 // specified offset so find it.
239 if (Offset
== 0 || Offset
== size()) {
240 // Fastpath for a common case. There is already a splitpoint at the end.
244 // Find the piece that this offset lands in.
245 unsigned PieceOffs
= 0;
247 while (Offset
>= PieceOffs
+Pieces
[i
].size()) {
248 PieceOffs
+= Pieces
[i
].size();
252 // If there is already a split point at the specified offset, just return
254 if (PieceOffs
== Offset
)
257 // Otherwise, we need to split piece 'i' at Offset-PieceOffs. Convert Offset
258 // to being Piece relative.
259 unsigned IntraPieceOffset
= Offset
-PieceOffs
;
261 // We do this by shrinking the RopePiece and then doing an insert of the tail.
262 RopePiece
Tail(Pieces
[i
].StrData
, Pieces
[i
].StartOffs
+IntraPieceOffset
,
264 Size
-= Pieces
[i
].size();
265 Pieces
[i
].EndOffs
= Pieces
[i
].StartOffs
+IntraPieceOffset
;
266 Size
+= Pieces
[i
].size();
268 return insert(Offset
, Tail
);
271 /// insert - Insert the specified RopePiece into this tree node at the
272 /// specified offset. The offset is relative, so "0" is the start of the node.
274 /// If there is no space in this subtree for the extra piece, the extra tree
275 /// node is returned and must be inserted into a parent.
276 RopePieceBTreeNode
*RopePieceBTreeLeaf::insert(unsigned Offset
,
277 const RopePiece
&R
) {
278 // If this node is not full, insert the piece.
280 // Find the insertion point. We are guaranteed that there is a split at the
281 // specified offset so find it.
282 unsigned i
= 0, e
= getNumPieces();
283 if (Offset
== size()) {
284 // Fastpath for a common case.
287 unsigned SlotOffs
= 0;
288 for (; Offset
> SlotOffs
; ++i
)
289 SlotOffs
+= getPiece(i
).size();
290 assert(SlotOffs
== Offset
&& "Split didn't occur before insertion!");
293 // For an insertion into a non-full leaf node, just insert the value in
294 // its sorted position. This requires moving later values over.
296 Pieces
[e
] = Pieces
[e
-1];
303 // Otherwise, if this is leaf is full, split it in two halves. Since this
304 // node is full, it contains 2*WidthFactor values. We move the first
305 // 'WidthFactor' values to the LHS child (which we leave in this node) and
306 // move the last 'WidthFactor' values into the RHS child.
308 // Create the new node.
309 RopePieceBTreeLeaf
*NewNode
= new RopePieceBTreeLeaf();
311 // Move over the last 'WidthFactor' values from here to NewNode.
312 std::copy(&Pieces
[WidthFactor
], &Pieces
[2*WidthFactor
],
313 &NewNode
->Pieces
[0]);
314 // Replace old pieces with null RopePieces to drop refcounts.
315 std::fill(&Pieces
[WidthFactor
], &Pieces
[2*WidthFactor
], RopePiece());
317 // Decrease the number of values in the two nodes.
318 NewNode
->NumPieces
= NumPieces
= WidthFactor
;
320 // Recompute the two nodes' size.
321 NewNode
->FullRecomputeSizeLocally();
322 FullRecomputeSizeLocally();
324 // Update the list of leaves.
325 NewNode
->insertAfterLeafInOrder(this);
327 // These insertions can't fail.
328 if (this->size() >= Offset
)
329 this->insert(Offset
, R
);
331 NewNode
->insert(Offset
- this->size(), R
);
335 /// erase - Remove NumBytes from this node at the specified offset. We are
336 /// guaranteed that there is a split at Offset.
337 void RopePieceBTreeLeaf::erase(unsigned Offset
, unsigned NumBytes
) {
338 // Since we are guaranteed that there is a split at Offset, we start by
339 // finding the Piece that starts there.
340 unsigned PieceOffs
= 0;
342 for (; Offset
> PieceOffs
; ++i
)
343 PieceOffs
+= getPiece(i
).size();
344 assert(PieceOffs
== Offset
&& "Split didn't occur before erase!");
346 unsigned StartPiece
= i
;
348 // Figure out how many pieces completely cover 'NumBytes'. We want to remove
350 for (; Offset
+NumBytes
> PieceOffs
+getPiece(i
).size(); ++i
)
351 PieceOffs
+= getPiece(i
).size();
353 // If we exactly include the last one, include it in the region to delete.
354 if (Offset
+NumBytes
== PieceOffs
+getPiece(i
).size()) {
355 PieceOffs
+= getPiece(i
).size();
359 // If we completely cover some RopePieces, erase them now.
360 if (i
!= StartPiece
) {
361 unsigned NumDeleted
= i
-StartPiece
;
362 for (; i
!= getNumPieces(); ++i
)
363 Pieces
[i
-NumDeleted
] = Pieces
[i
];
365 // Drop references to dead rope pieces.
366 std::fill(&Pieces
[getNumPieces()-NumDeleted
], &Pieces
[getNumPieces()],
368 NumPieces
-= NumDeleted
;
370 unsigned CoverBytes
= PieceOffs
-Offset
;
371 NumBytes
-= CoverBytes
;
375 // If we completely removed some stuff, we could be done.
376 if (NumBytes
== 0) return;
378 // Okay, now might be erasing part of some Piece. If this is the case, then
379 // move the start point of the piece.
380 assert(getPiece(StartPiece
).size() > NumBytes
);
381 Pieces
[StartPiece
].StartOffs
+= NumBytes
;
383 // The size of this node just shrunk by NumBytes.
387 //===----------------------------------------------------------------------===//
388 // RopePieceBTreeInterior Class
389 //===----------------------------------------------------------------------===//
393 /// RopePieceBTreeInterior - This represents an interior node in the B+Tree,
394 /// which holds up to 2*WidthFactor pointers to child nodes.
395 class RopePieceBTreeInterior
: public RopePieceBTreeNode
{
396 /// NumChildren - This holds the number of children currently active in the
398 unsigned char NumChildren
= 0;
400 RopePieceBTreeNode
*Children
[2*WidthFactor
];
403 RopePieceBTreeInterior() : RopePieceBTreeNode(false) {}
405 RopePieceBTreeInterior(RopePieceBTreeNode
*LHS
, RopePieceBTreeNode
*RHS
)
406 : RopePieceBTreeNode(false) {
410 Size
= LHS
->size() + RHS
->size();
413 ~RopePieceBTreeInterior() {
414 for (unsigned i
= 0, e
= getNumChildren(); i
!= e
; ++i
)
415 Children
[i
]->Destroy();
418 bool isFull() const { return NumChildren
== 2*WidthFactor
; }
420 unsigned getNumChildren() const { return NumChildren
; }
422 const RopePieceBTreeNode
*getChild(unsigned i
) const {
423 assert(i
< NumChildren
&& "invalid child #");
427 RopePieceBTreeNode
*getChild(unsigned i
) {
428 assert(i
< NumChildren
&& "invalid child #");
432 /// FullRecomputeSizeLocally - Recompute the Size field of this node by
433 /// summing up the sizes of the child nodes.
434 void FullRecomputeSizeLocally() {
436 for (unsigned i
= 0, e
= getNumChildren(); i
!= e
; ++i
)
437 Size
+= getChild(i
)->size();
440 /// split - Split the range containing the specified offset so that we are
441 /// guaranteed that there is a place to do an insertion at the specified
442 /// offset. The offset is relative, so "0" is the start of the node.
444 /// If there is no space in this subtree for the extra piece, the extra tree
445 /// node is returned and must be inserted into a parent.
446 RopePieceBTreeNode
*split(unsigned Offset
);
448 /// insert - Insert the specified ropepiece into this tree node at the
449 /// specified offset. The offset is relative, so "0" is the start of the
452 /// If there is no space in this subtree for the extra piece, the extra tree
453 /// node is returned and must be inserted into a parent.
454 RopePieceBTreeNode
*insert(unsigned Offset
, const RopePiece
&R
);
456 /// HandleChildPiece - A child propagated an insertion result up to us.
457 /// Insert the new child, and/or propagate the result further up the tree.
458 RopePieceBTreeNode
*HandleChildPiece(unsigned i
, RopePieceBTreeNode
*RHS
);
460 /// erase - Remove NumBytes from this node at the specified offset. We are
461 /// guaranteed that there is a split at Offset.
462 void erase(unsigned Offset
, unsigned NumBytes
);
464 static bool classof(const RopePieceBTreeNode
*N
) {
471 /// split - Split the range containing the specified offset so that we are
472 /// guaranteed that there is a place to do an insertion at the specified
473 /// offset. The offset is relative, so "0" is the start of the node.
475 /// If there is no space in this subtree for the extra piece, the extra tree
476 /// node is returned and must be inserted into a parent.
477 RopePieceBTreeNode
*RopePieceBTreeInterior::split(unsigned Offset
) {
478 // Figure out which child to split.
479 if (Offset
== 0 || Offset
== size())
480 return nullptr; // If we have an exact offset, we're already split.
482 unsigned ChildOffset
= 0;
484 for (; Offset
>= ChildOffset
+getChild(i
)->size(); ++i
)
485 ChildOffset
+= getChild(i
)->size();
487 // If already split there, we're done.
488 if (ChildOffset
== Offset
)
491 // Otherwise, recursively split the child.
492 if (RopePieceBTreeNode
*RHS
= getChild(i
)->split(Offset
-ChildOffset
))
493 return HandleChildPiece(i
, RHS
);
494 return nullptr; // Done!
497 /// insert - Insert the specified ropepiece into this tree node at the
498 /// specified offset. The offset is relative, so "0" is the start of the
501 /// If there is no space in this subtree for the extra piece, the extra tree
502 /// node is returned and must be inserted into a parent.
503 RopePieceBTreeNode
*RopePieceBTreeInterior::insert(unsigned Offset
,
504 const RopePiece
&R
) {
505 // Find the insertion point. We are guaranteed that there is a split at the
506 // specified offset so find it.
507 unsigned i
= 0, e
= getNumChildren();
509 unsigned ChildOffs
= 0;
510 if (Offset
== size()) {
511 // Fastpath for a common case. Insert at end of last child.
513 ChildOffs
= size()-getChild(i
)->size();
515 for (; Offset
> ChildOffs
+getChild(i
)->size(); ++i
)
516 ChildOffs
+= getChild(i
)->size();
521 // Insert at the end of this child.
522 if (RopePieceBTreeNode
*RHS
= getChild(i
)->insert(Offset
-ChildOffs
, R
))
523 return HandleChildPiece(i
, RHS
);
528 /// HandleChildPiece - A child propagated an insertion result up to us.
529 /// Insert the new child, and/or propagate the result further up the tree.
531 RopePieceBTreeInterior::HandleChildPiece(unsigned i
, RopePieceBTreeNode
*RHS
) {
532 // Otherwise the child propagated a subtree up to us as a new child. See if
533 // we have space for it here.
535 // Insert RHS after child 'i'.
536 if (i
+ 1 != getNumChildren())
537 memmove(&Children
[i
+2], &Children
[i
+1],
538 (getNumChildren()-i
-1)*sizeof(Children
[0]));
544 // Okay, this node is full. Split it in half, moving WidthFactor children to
545 // a newly allocated interior node.
547 // Create the new node.
548 RopePieceBTreeInterior
*NewNode
= new RopePieceBTreeInterior();
550 // Move over the last 'WidthFactor' values from here to NewNode.
551 memcpy(&NewNode
->Children
[0], &Children
[WidthFactor
],
552 WidthFactor
*sizeof(Children
[0]));
554 // Decrease the number of values in the two nodes.
555 NewNode
->NumChildren
= NumChildren
= WidthFactor
;
557 // Finally, insert the two new children in the side the can (now) hold them.
558 // These insertions can't fail.
560 this->HandleChildPiece(i
, RHS
);
562 NewNode
->HandleChildPiece(i
-WidthFactor
, RHS
);
564 // Recompute the two nodes' size.
565 NewNode
->FullRecomputeSizeLocally();
566 FullRecomputeSizeLocally();
570 /// erase - Remove NumBytes from this node at the specified offset. We are
571 /// guaranteed that there is a split at Offset.
572 void RopePieceBTreeInterior::erase(unsigned Offset
, unsigned NumBytes
) {
573 // This will shrink this node by NumBytes.
576 // Find the first child that overlaps with Offset.
578 for (; Offset
>= getChild(i
)->size(); ++i
)
579 Offset
-= getChild(i
)->size();
581 // Propagate the delete request into overlapping children, or completely
582 // delete the children as appropriate.
584 RopePieceBTreeNode
*CurChild
= getChild(i
);
586 // If we are deleting something contained entirely in the child, pass on the
588 if (Offset
+NumBytes
< CurChild
->size()) {
589 CurChild
->erase(Offset
, NumBytes
);
593 // If this deletion request starts somewhere in the middle of the child, it
594 // must be deleting to the end of the child.
596 unsigned BytesFromChild
= CurChild
->size()-Offset
;
597 CurChild
->erase(Offset
, BytesFromChild
);
598 NumBytes
-= BytesFromChild
;
599 // Start at the beginning of the next child.
605 // If the deletion request completely covers the child, delete it and move
607 NumBytes
-= CurChild
->size();
610 if (i
!= getNumChildren())
611 memmove(&Children
[i
], &Children
[i
+1],
612 (getNumChildren()-i
)*sizeof(Children
[0]));
616 //===----------------------------------------------------------------------===//
617 // RopePieceBTreeNode Implementation
618 //===----------------------------------------------------------------------===//
620 void RopePieceBTreeNode::Destroy() {
621 if (auto *Leaf
= dyn_cast
<RopePieceBTreeLeaf
>(this))
624 delete cast
<RopePieceBTreeInterior
>(this);
627 /// split - Split the range containing the specified offset so that we are
628 /// guaranteed that there is a place to do an insertion at the specified
629 /// offset. The offset is relative, so "0" is the start of the node.
631 /// If there is no space in this subtree for the extra piece, the extra tree
632 /// node is returned and must be inserted into a parent.
633 RopePieceBTreeNode
*RopePieceBTreeNode::split(unsigned Offset
) {
634 assert(Offset
<= size() && "Invalid offset to split!");
635 if (auto *Leaf
= dyn_cast
<RopePieceBTreeLeaf
>(this))
636 return Leaf
->split(Offset
);
637 return cast
<RopePieceBTreeInterior
>(this)->split(Offset
);
640 /// insert - Insert the specified ropepiece into this tree node at the
641 /// specified offset. The offset is relative, so "0" is the start of the
644 /// If there is no space in this subtree for the extra piece, the extra tree
645 /// node is returned and must be inserted into a parent.
646 RopePieceBTreeNode
*RopePieceBTreeNode::insert(unsigned Offset
,
647 const RopePiece
&R
) {
648 assert(Offset
<= size() && "Invalid offset to insert!");
649 if (auto *Leaf
= dyn_cast
<RopePieceBTreeLeaf
>(this))
650 return Leaf
->insert(Offset
, R
);
651 return cast
<RopePieceBTreeInterior
>(this)->insert(Offset
, R
);
654 /// erase - Remove NumBytes from this node at the specified offset. We are
655 /// guaranteed that there is a split at Offset.
656 void RopePieceBTreeNode::erase(unsigned Offset
, unsigned NumBytes
) {
657 assert(Offset
+NumBytes
<= size() && "Invalid offset to erase!");
658 if (auto *Leaf
= dyn_cast
<RopePieceBTreeLeaf
>(this))
659 return Leaf
->erase(Offset
, NumBytes
);
660 return cast
<RopePieceBTreeInterior
>(this)->erase(Offset
, NumBytes
);
663 //===----------------------------------------------------------------------===//
664 // RopePieceBTreeIterator Implementation
665 //===----------------------------------------------------------------------===//
667 static const RopePieceBTreeLeaf
*getCN(const void *P
) {
668 return static_cast<const RopePieceBTreeLeaf
*>(P
);
672 RopePieceBTreeIterator::RopePieceBTreeIterator(const void *n
) {
673 const auto *N
= static_cast<const RopePieceBTreeNode
*>(n
);
675 // Walk down the left side of the tree until we get to a leaf.
676 while (const auto *IN
= dyn_cast
<RopePieceBTreeInterior
>(N
))
679 // We must have at least one leaf.
680 CurNode
= cast
<RopePieceBTreeLeaf
>(N
);
682 // If we found a leaf that happens to be empty, skip over it until we get
683 // to something full.
684 while (CurNode
&& getCN(CurNode
)->getNumPieces() == 0)
685 CurNode
= getCN(CurNode
)->getNextLeafInOrder();
688 CurPiece
= &getCN(CurNode
)->getPiece(0);
689 else // Empty tree, this is an end() iterator.
694 void RopePieceBTreeIterator::MoveToNextPiece() {
695 if (CurPiece
!= &getCN(CurNode
)->getPiece(getCN(CurNode
)->getNumPieces()-1)) {
701 // Find the next non-empty leaf node.
703 CurNode
= getCN(CurNode
)->getNextLeafInOrder();
704 while (CurNode
&& getCN(CurNode
)->getNumPieces() == 0);
707 CurPiece
= &getCN(CurNode
)->getPiece(0);
713 //===----------------------------------------------------------------------===//
714 // RopePieceBTree Implementation
715 //===----------------------------------------------------------------------===//
717 static RopePieceBTreeNode
*getRoot(void *P
) {
718 return static_cast<RopePieceBTreeNode
*>(P
);
721 RopePieceBTree::RopePieceBTree() {
722 Root
= new RopePieceBTreeLeaf();
725 RopePieceBTree::RopePieceBTree(const RopePieceBTree
&RHS
) {
726 assert(RHS
.empty() && "Can't copy non-empty tree yet");
727 Root
= new RopePieceBTreeLeaf();
730 RopePieceBTree::~RopePieceBTree() {
731 getRoot(Root
)->Destroy();
734 unsigned RopePieceBTree::size() const {
735 return getRoot(Root
)->size();
738 void RopePieceBTree::clear() {
739 if (auto *Leaf
= dyn_cast
<RopePieceBTreeLeaf
>(getRoot(Root
)))
742 getRoot(Root
)->Destroy();
743 Root
= new RopePieceBTreeLeaf();
747 void RopePieceBTree::insert(unsigned Offset
, const RopePiece
&R
) {
748 // #1. Split at Offset.
749 if (RopePieceBTreeNode
*RHS
= getRoot(Root
)->split(Offset
))
750 Root
= new RopePieceBTreeInterior(getRoot(Root
), RHS
);
752 // #2. Do the insertion.
753 if (RopePieceBTreeNode
*RHS
= getRoot(Root
)->insert(Offset
, R
))
754 Root
= new RopePieceBTreeInterior(getRoot(Root
), RHS
);
757 void RopePieceBTree::erase(unsigned Offset
, unsigned NumBytes
) {
758 // #1. Split at Offset.
759 if (RopePieceBTreeNode
*RHS
= getRoot(Root
)->split(Offset
))
760 Root
= new RopePieceBTreeInterior(getRoot(Root
), RHS
);
762 // #2. Do the erasing.
763 getRoot(Root
)->erase(Offset
, NumBytes
);
766 //===----------------------------------------------------------------------===//
767 // RewriteRope Implementation
768 //===----------------------------------------------------------------------===//
770 /// MakeRopeString - This copies the specified byte range into some instance of
771 /// RopeRefCountString, and return a RopePiece that represents it. This uses
772 /// the AllocBuffer object to aggregate requests for small strings into one
773 /// allocation instead of doing tons of tiny allocations.
774 RopePiece
RewriteRope::MakeRopeString(const char *Start
, const char *End
) {
775 unsigned Len
= End
-Start
;
776 assert(Len
&& "Zero length RopePiece is invalid!");
778 // If we have space for this string in the current alloc buffer, use it.
779 if (AllocOffs
+Len
<= AllocChunkSize
) {
780 memcpy(AllocBuffer
->Data
+AllocOffs
, Start
, Len
);
782 return RopePiece(AllocBuffer
, AllocOffs
-Len
, AllocOffs
);
785 // If we don't have enough room because this specific allocation is huge,
786 // just allocate a new rope piece for it alone.
787 if (Len
> AllocChunkSize
) {
788 unsigned Size
= End
-Start
+sizeof(RopeRefCountString
)-1;
789 auto *Res
= reinterpret_cast<RopeRefCountString
*>(new char[Size
]);
791 memcpy(Res
->Data
, Start
, End
-Start
);
792 return RopePiece(Res
, 0, End
-Start
);
795 // Otherwise, this was a small request but we just don't have space for it
796 // Make a new chunk and share it with later allocations.
798 unsigned AllocSize
= offsetof(RopeRefCountString
, Data
) + AllocChunkSize
;
799 auto *Res
= reinterpret_cast<RopeRefCountString
*>(new char[AllocSize
]);
801 memcpy(Res
->Data
, Start
, Len
);
805 return RopePiece(AllocBuffer
, 0, Len
);