1/*
2 *
3 * Copyright (c) 1996,1997
4 * Silicon Graphics Computer Systems, Inc.
5 *
6 * Copyright (c) 1997
7 * Moscow Center for SPARC Technology
8 *
9 * Copyright (c) 1999
10 * Boris Fomitchev
11 *
12 * This material is provided "as is", with absolutely no warranty expressed
13 * or implied. Any use is at your own risk.
14 *
15 * Permission to use or copy this software for any purpose is hereby granted
16 * without fee, provided the above notices are retained on all copies.
17 * Permission to modify the code and to distribute modified code is granted,
18 * provided the above notices are retained, and a notice that the code was
19 * modified is included with the above copyright notice.
20 *
21 */
22
23/* NOTE: This is an internal header file, included by other STL headers.
24 *   You should not attempt to use it directly.
25 */
26
27// rope<_CharT,_Alloc> is a sequence of _CharT.
28// Ropes appear to be mutable, but update operations
29// really copy enough of the data structure to leave the original
30// valid.  Thus ropes can be logically copied by just copying
31// a pointer value.
32
33#ifndef _STLP_INTERNAL_ROPE_H
34#define _STLP_INTERNAL_ROPE_H
35
36#ifndef _STLP_INTERNAL_ALGOBASE_H
37#  include <stl/_algobase.h>
38#endif
39
40#if !defined (_STLP_USE_NO_IOSTREAMS) && !defined (_STLP_INTERNAL_IOSFWD)
41#  include <stl/_iosfwd.h>
42#endif
43
44#ifndef _STLP_INTERNAL_ALLOC_H
45#  include <stl/_alloc.h>
46#endif
47
48#ifndef _STLP_INTERNAL_ITERATOR_H
49#  include <stl/_iterator.h>
50#endif
51
52#ifndef _STLP_INTERNAL_ALGO_H
53#  include <stl/_algo.h>
54#endif
55
56#ifndef _STLP_INTERNAL_FUNCTION_BASE_H
57#  include <stl/_function_base.h>
58#endif
59
60#ifndef _STLP_INTERNAL_NUMERIC_H
61#  include <stl/_numeric.h>
62#endif
63
64#ifndef _STLP_INTERNAL_HASH_FUN_H
65#  include <stl/_hash_fun.h>
66#endif
67
68#ifndef _STLP_CHAR_TRAITS_H
69#  include <stl/char_traits.h>
70#endif
71
72#ifndef _STLP_INTERNAL_THREADS_H
73#  include <stl/_threads.h>
74#endif
75
76#ifdef _STLP_SGI_THREADS
77#  include <mutex.h>
78#endif
79
80#ifndef _STLP_DONT_SUPPORT_REBIND_MEMBER_TEMPLATE
81#  define _STLP_CREATE_ALLOCATOR(__atype,__a, _Tp) (_Alloc_traits<_Tp,__atype>::create_allocator(__a))
82#else
83#  define _STLP_CREATE_ALLOCATOR(__atype,__a, _Tp) __stl_alloc_create(__a,(_Tp*)0)
84#endif
85
86_STLP_BEGIN_NAMESPACE
87
88// First a lot of forward declarations.  The standard seems to require
89// much stricter "declaration before use" than many of the implementations
90// that preceded it.
91template<class _CharT, _STLP_DFL_TMPL_PARAM(_Alloc, allocator<_CharT>) > class rope;
92template<class _CharT, class _Alloc> struct _Rope_RopeConcatenation;
93template<class _CharT, class _Alloc> struct _Rope_RopeRep;
94template<class _CharT, class _Alloc> struct _Rope_RopeLeaf;
95template<class _CharT, class _Alloc> struct _Rope_RopeFunction;
96template<class _CharT, class _Alloc> struct _Rope_RopeSubstring;
97template<class _CharT, class _Alloc> class _Rope_iterator;
98template<class _CharT, class _Alloc> class _Rope_const_iterator;
99template<class _CharT, class _Alloc> class _Rope_char_ref_proxy;
100template<class _CharT, class _Alloc> class _Rope_char_ptr_proxy;
101
102_STLP_MOVE_TO_PRIV_NAMESPACE
103
104template <class _CharT>
105struct _BasicCharType { typedef __false_type _Ret; };
106
107_STLP_TEMPLATE_NULL
108struct _BasicCharType<char> { typedef __true_type _Ret; };
109
110#ifdef _STLP_HAS_WCHAR_T
111_STLP_TEMPLATE_NULL
112struct _BasicCharType<wchar_t> { typedef __true_type _Ret; };
113#endif
114
115// Some helpers, so we can use the power algorithm on ropes.
116// See below for why this isn't local to the implementation.
117
118// This uses a nonstandard refcount convention.
119// The result has refcount 0.
120template<class _CharT, class _Alloc>
121struct _Rope_Concat_fn
122  : public binary_function<rope<_CharT,_Alloc>, rope<_CharT,_Alloc>,
123                           rope<_CharT,_Alloc> > {
124  rope<_CharT,_Alloc> operator() (const rope<_CharT,_Alloc>& __x,
125                                  const rope<_CharT,_Alloc>& __y) {
126    return __x + __y;
127  }
128};
129
130template <class _CharT, class _Alloc>
131inline
132rope<_CharT,_Alloc>
133__identity_element(_Rope_Concat_fn<_CharT, _Alloc>)
134{ return rope<_CharT,_Alloc>(); }
135
136_STLP_MOVE_TO_STD_NAMESPACE
137
138// Store an eos
139template <class _CharT>
140inline void _S_construct_null_aux(_CharT *__p, const __true_type&)
141{ *__p = 0; }
142
143template <class _CharT>
144inline void _S_construct_null_aux(_CharT *__p, const __false_type&)
145{ _STLP_STD::_Construct(__p); }
146
147template <class _CharT>
148inline void _S_construct_null(_CharT *__p) {
149  typedef typename _IsIntegral<_CharT>::_Ret _Char_Is_Integral;
150  _S_construct_null_aux(__p, _Char_Is_Integral());
151}
152
153// char_producers are logically functions that generate a section of
154// a string.  These can be converted to ropes.  The resulting rope
155// invokes the char_producer on demand.  This allows, for example,
156// files to be viewed as ropes without reading the entire file.
157template <class _CharT>
158class char_producer {
159public:
160  virtual ~char_producer() {}
161  virtual void operator()(size_t __start_pos, size_t __len,
162                          _CharT* __buffer) = 0;
163  // Buffer should really be an arbitrary output iterator.
164  // That way we could flatten directly into an ostream, etc.
165  // This is thoroughly impossible, since iterator types don't
166  // have runtime descriptions.
167};
168
169// Sequence buffers:
170//
171// Sequence must provide an append operation that appends an
172// array to the sequence.  Sequence buffers are useful only if
173// appending an entire array is cheaper than appending element by element.
174// This is true for many string representations.
175// This should  perhaps inherit from ostream<sequence::value_type>
176// and be implemented correspondingly, so that they can be used
177// for formatted.  For the sake of portability, we don't do this yet.
178//
179// For now, sequence buffers behave as output iterators.  But they also
180// behave a little like basic_ostringstream<sequence::value_type> and a
181// little like containers.
182
183template<class _Sequence
184# if !(defined (_STLP_NON_TYPE_TMPL_PARAM_BUG) || \
185       defined ( _STLP_NO_DEFAULT_NON_TYPE_PARAM ))
186         , size_t _Buf_sz = 100
187#   if defined(__sgi) && !defined(__GNUC__)
188#   define __TYPEDEF_WORKAROUND
189         ,class _V = typename _Sequence::value_type
190#   endif /* __sgi */
191# endif /* _STLP_NON_TYPE_TMPL_PARAM_BUG */
192         >
193// The 3rd parameter works around a common compiler bug.
194class sequence_buffer : public iterator <output_iterator_tag, void, void, void, void> {
195public:
196# ifndef __TYPEDEF_WORKAROUND
197  typedef typename _Sequence::value_type value_type;
198  typedef sequence_buffer<_Sequence
199# if !(defined (_STLP_NON_TYPE_TMPL_PARAM_BUG) || \
200       defined ( _STLP_NO_DEFAULT_NON_TYPE_PARAM ))
201  , _Buf_sz
202  > _Self;
203# else /* _STLP_NON_TYPE_TMPL_PARAM_BUG */
204  > _Self;
205  enum { _Buf_sz = 100};
206# endif /* _STLP_NON_TYPE_TMPL_PARAM_BUG */
207  // # endif
208# else /* __TYPEDEF_WORKAROUND */
209  typedef _V value_type;
210  typedef sequence_buffer<_Sequence, _Buf_sz, _V> _Self;
211# endif /* __TYPEDEF_WORKAROUND */
212protected:
213  _Sequence* _M_prefix;
214  value_type _M_buffer[_Buf_sz];
215  size_t     _M_buf_count;
216public:
217  void flush() {
218    _M_prefix->append(_M_buffer, _M_buffer + _M_buf_count);
219    _M_buf_count = 0;
220  }
221  ~sequence_buffer() { flush(); }
222  sequence_buffer() : _M_prefix(0), _M_buf_count(0) {}
223  sequence_buffer(const _Self& __x) {
224    _M_prefix = __x._M_prefix;
225    _M_buf_count = __x._M_buf_count;
226    _STLP_STD::copy(__x._M_buffer, __x._M_buffer + __x._M_buf_count, _M_buffer);
227  }
228  sequence_buffer(_Self& __x) {
229    __x.flush();
230    _M_prefix = __x._M_prefix;
231    _M_buf_count = 0;
232  }
233  sequence_buffer(_Sequence& __s) : _M_prefix(&__s), _M_buf_count(0) {}
234  _Self& operator= (_Self& __x) {
235    __x.flush();
236    _M_prefix = __x._M_prefix;
237    _M_buf_count = 0;
238    return *this;
239  }
240  _Self& operator= (const _Self& __x) {
241    _M_prefix = __x._M_prefix;
242    _M_buf_count = __x._M_buf_count;
243    _STLP_STD::copy(__x._M_buffer, __x._M_buffer + __x._M_buf_count, _M_buffer);
244    return *this;
245  }
246  void push_back(value_type __x) {
247    if (_M_buf_count < _Buf_sz) {
248      _M_buffer[_M_buf_count] = __x;
249      ++_M_buf_count;
250    } else {
251      flush();
252      _M_buffer[0] = __x;
253      _M_buf_count = 1;
254    }
255  }
256  void append(const value_type *__s, size_t __len) {
257    if (__len + _M_buf_count <= _Buf_sz) {
258      size_t __i = _M_buf_count;
259      size_t __j = 0;
260      for (; __j < __len; __i++, __j++) {
261        _M_buffer[__i] = __s[__j];
262      }
263      _M_buf_count += __len;
264    } else if (0 == _M_buf_count) {
265      _M_prefix->append(__s, __s + __len);
266    } else {
267      flush();
268      append(__s, __len);
269    }
270  }
271  _Self& write(const value_type *__s, size_t __len) {
272    append(__s, __len);
273    return *this;
274  }
275  _Self& put(value_type __x) {
276    push_back(__x);
277    return *this;
278  }
279  _Self& operator=(const value_type& __rhs) {
280    push_back(__rhs);
281    return *this;
282  }
283  _Self& operator*() { return *this; }
284  _Self& operator++() { return *this; }
285  _Self& operator++(int) { return *this; }
286};
287
288// The following should be treated as private, at least for now.
289template<class _CharT>
290class _Rope_char_consumer {
291#if !defined (_STLP_MEMBER_TEMPLATES)
292public:
293  //Without member templates we have to use run-time parameterization.
294  // The symmetry with char_producer is accidental and temporary.
295  virtual ~_Rope_char_consumer() {}
296  virtual bool operator()(const _CharT* __buffer, size_t __len) = 0;
297#endif
298};
299
300//
301// What follows should really be local to rope.  Unfortunately,
302// that doesn't work, since it makes it impossible to define generic
303// equality on rope iterators.  According to the draft standard, the
304// template parameters for such an equality operator cannot be inferred
305// from the occurence of a member class as a parameter.
306// (SGI compilers in fact allow this, but the __result wouldn't be
307// portable.)
308// Similarly, some of the static member functions are member functions
309// only to avoid polluting the global namespace, and to circumvent
310// restrictions on type inference for template functions.
311//
312
313//
314// The internal data structure for representing a rope.  This is
315// private to the implementation.  A rope is really just a pointer
316// to one of these.
317//
318// A few basic functions for manipulating this data structure
319// are members of _RopeRep.  Most of the more complex algorithms
320// are implemented as rope members.
321//
322// Some of the static member functions of _RopeRep have identically
323// named functions in rope that simply invoke the _RopeRep versions.
324//
325
326template<class _CharT, class _Alloc>
327struct _Rope_RopeRep
328  : public _Refcount_Base
329{
330  typedef _Rope_RopeRep<_CharT, _Alloc> _Self;
331public:
332  //
333  // GAB: 11/09/05
334  //
335  // "__ROPE_DEPTH_SIZE" is set to one more then the "__ROPE_MAX_DEPTH".
336  // This was originally just an addition of "__ROPE_MAX_DEPTH + 1"
337  // but this addition causes the sunpro compiler to complain about
338  // multiple declarations during the initialization of "_S_min_len".
339  // Changed to be a fixed value and the sunpro compiler appears to
340  // be happy???
341  //
342#  define __ROPE_MAX_DEPTH  45
343#  define __ROPE_DEPTH_SIZE 46 // __ROPE_MAX_DEPTH + 1
344  enum { _S_max_rope_depth = __ROPE_MAX_DEPTH };
345  enum _Tag {_S_leaf, _S_concat, _S_substringfn, _S_function};
346  // Apparently needed by VC++
347  // The data fields of leaves are allocated with some
348  // extra space, to accomodate future growth and for basic
349  // character types, to hold a trailing eos character.
350  enum { _S_alloc_granularity = 8 };
351
352  _Tag _M_tag:8;
353  bool _M_is_balanced:8;
354
355  _STLP_FORCE_ALLOCATORS(_CharT, _Alloc)
356  typedef _Alloc allocator_type;
357
358  allocator_type get_allocator() const { return allocator_type(_M_size);  }
359
360  unsigned char _M_depth;
361  _CharT* _STLP_VOLATILE _M_c_string;
362  _STLP_PRIV _STLP_alloc_proxy<size_t, _CharT, allocator_type> _M_size;
363
364#ifdef _STLP_NO_ARROW_OPERATOR
365  _Rope_RopeRep() : _Refcount_Base(1), _M_size(allocator_type(), 0) {
366#  if defined (_STLP_CHECK_RUNTIME_COMPATIBILITY)
367    _STLP_CHECK_RUNTIME_COMPATIBILITY();
368#  endif
369  }
370#endif
371
372  /* Flattened version of string, if needed.  */
373  /* typically 0.                             */
374  /* If it's not 0, then the memory is owned  */
375  /* by this node.                            */
376  /* In the case of a leaf, this may point to */
377  /* the same memory as the data field.       */
378  _Rope_RopeRep(_Tag __t, unsigned char __d, bool __b, size_t _p_size,
379                allocator_type __a) :
380    _Refcount_Base(1),
381    _M_tag(__t), _M_is_balanced(__b), _M_depth(__d), _M_c_string(0), _M_size(__a, _p_size) {
382#if defined (_STLP_CHECK_RUNTIME_COMPATIBILITY)
383    _STLP_CHECK_RUNTIME_COMPATIBILITY();
384#endif
385    }
386
387  typedef _STLP_TYPENAME _STLP_PRIV _BasicCharType<_CharT>::_Ret _IsBasicCharType;
388
389#if 0
390  /* Please tell why this code is necessary if you uncomment it.
391   * Problem with it is that rope implementation expect that _S_rounded_up_size(n)
392   * returns a size > n in order to store the terminating null charater. When
393   * instanciation type is not a char or wchar_t this is not guaranty resulting in
394   * memory overrun.
395   */
396  static size_t _S_rounded_up_size_aux(size_t __n, __true_type const& /*_IsBasicCharType*/) {
397    // Allow slop for in-place expansion.
398    return (__n + _S_alloc_granularity) & ~(_S_alloc_granularity - 1);
399  }
400
401  static size_t _S_rounded_up_size_aux(size_t __n, __false_type const& /*_IsBasicCharType*/) {
402    // Allow slop for in-place expansion.
403    return (__n + _S_alloc_granularity - 1) & ~(_S_alloc_granularity - 1);
404  }
405#endif
406  // fbp : moved from RopeLeaf
407  static size_t _S_rounded_up_size(size_t __n)
408  //{ return _S_rounded_up_size_aux(__n, _IsBasicCharType()); }
409  { return (__n + _S_alloc_granularity) & ~(_S_alloc_granularity - 1); }
410
411  static void _S_free_string( _CharT* __s, size_t __len,
412                             allocator_type __a) {
413    _STLP_STD::_Destroy_Range(__s, __s + __len);
414    //  This has to be a static member, so this gets a bit messy
415#   ifndef _STLP_DONT_SUPPORT_REBIND_MEMBER_TEMPLATE
416    __a.deallocate(__s, _S_rounded_up_size(__len));    //*ty 03/24/2001 - restored not to use __stl_alloc_rebind() since it is not defined under _STLP_MEMBER_TEMPLATE_CLASSES
417#   else
418    __stl_alloc_rebind (__a, (_CharT*)0).deallocate(__s, _S_rounded_up_size(__len));
419#   endif
420  }
421
422  // Deallocate data section of a leaf.
423  // This shouldn't be a member function.
424  // But its hard to do anything else at the
425  // moment, because it's templatized w.r.t.
426  // an allocator.
427  // Does nothing if __GC is defined.
428  void _M_free_c_string();
429  void _M_free_tree();
430  // Deallocate t. Assumes t is not 0.
431  void _M_unref_nonnil() {
432    if (_M_decr() == 0) _M_free_tree();
433  }
434  void _M_ref_nonnil() {
435    _M_incr();
436  }
437  static void _S_unref(_Self* __t) {
438    if (0 != __t) {
439      __t->_M_unref_nonnil();
440    }
441  }
442  static void _S_ref(_Self* __t) {
443    if (0 != __t) __t->_M_incr();
444  }
445  //static void _S_free_if_unref(_Self* __t) {
446  //  if (0 != __t && 0 == __t->_M_ref_count) __t->_M_free_tree();
447  //}
448};
449
450template<class _CharT, class _Alloc>
451struct _Rope_RopeLeaf : public _Rope_RopeRep<_CharT,_Alloc> {
452public:
453  _CharT* _M_data; /* Not necessarily 0 terminated. */
454                                /* The allocated size is         */
455                                /* _S_rounded_up_size(size), except */
456                                /* in the GC case, in which it   */
457                                /* doesn't matter.               */
458private:
459  typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep;
460  typedef typename _RopeRep::_IsBasicCharType _IsBasicCharType;
461  void _M_init(__true_type const& /*_IsBasicCharType*/) {
462    this->_M_c_string = _M_data;
463  }
464  void _M_init(__false_type const& /*_IsBasicCharType*/) {}
465
466public:
467  _STLP_FORCE_ALLOCATORS(_CharT, _Alloc)
468  typedef typename _RopeRep::allocator_type allocator_type;
469
470  _Rope_RopeLeaf( _CharT* __d, size_t _p_size, allocator_type __a)
471    : _Rope_RopeRep<_CharT,_Alloc>(_RopeRep::_S_leaf, 0, true, _p_size, __a),
472      _M_data(__d) {
473    _STLP_ASSERT(_p_size > 0)
474    _M_init(_IsBasicCharType());
475  }
476
477# ifdef _STLP_NO_ARROW_OPERATOR
478  _Rope_RopeLeaf() {}
479  _Rope_RopeLeaf(const _Rope_RopeLeaf<_CharT, _Alloc>& ) {}
480# endif
481
482// The constructor assumes that d has been allocated with
483  // the proper allocator and the properly padded size.
484  // In contrast, the destructor deallocates the data:
485  ~_Rope_RopeLeaf() {
486    if (_M_data != this->_M_c_string) {
487      this->_M_free_c_string();
488    }
489    _RopeRep::_S_free_string(_M_data, this->_M_size._M_data, this->get_allocator());
490  }
491};
492
493template<class _CharT, class _Alloc>
494struct _Rope_RopeConcatenation : public _Rope_RopeRep<_CharT, _Alloc> {
495private:
496  typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep;
497
498public:
499  _RopeRep* _M_left;
500  _RopeRep* _M_right;
501  _STLP_FORCE_ALLOCATORS(_CharT, _Alloc)
502  typedef typename _RopeRep::allocator_type allocator_type;
503  _Rope_RopeConcatenation(_RopeRep* __l, _RopeRep* __r, allocator_type __a)
504    : _Rope_RopeRep<_CharT,_Alloc>(_RopeRep::_S_concat,
505                                   (max)(__l->_M_depth, __r->_M_depth) + 1, false,
506                                   __l->_M_size._M_data + __r->_M_size._M_data, __a), _M_left(__l), _M_right(__r)
507  {}
508# ifdef _STLP_NO_ARROW_OPERATOR
509  _Rope_RopeConcatenation() {}
510  _Rope_RopeConcatenation(const _Rope_RopeConcatenation<_CharT, _Alloc>&) {}
511# endif
512
513  ~_Rope_RopeConcatenation() {
514    this->_M_free_c_string();
515    _M_left->_M_unref_nonnil();
516    _M_right->_M_unref_nonnil();
517  }
518};
519
520template <class _CharT, class _Alloc>
521struct _Rope_RopeFunction : public _Rope_RopeRep<_CharT, _Alloc> {
522private:
523  typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep;
524public:
525  char_producer<_CharT>* _M_fn;
526  /*
527   * Char_producer is owned by the
528   * rope and should be explicitly
529   * deleted when the rope becomes
530   * inaccessible.
531   */
532  bool _M_delete_when_done;
533  _STLP_FORCE_ALLOCATORS(_CharT, _Alloc)
534  typedef typename _Rope_RopeRep<_CharT,_Alloc>::allocator_type allocator_type;
535# ifdef _STLP_NO_ARROW_OPERATOR
536  _Rope_RopeFunction() {}
537  _Rope_RopeFunction(const _Rope_RopeFunction<_CharT, _Alloc>& ) {}
538# endif
539
540  _Rope_RopeFunction(char_producer<_CharT>* __f, size_t _p_size,
541                     bool __d, allocator_type __a)
542    : _Rope_RopeRep<_CharT,_Alloc>(_RopeRep::_S_function, 0, true, _p_size, __a), _M_fn(__f)
543    , _M_delete_when_done(__d)
544  { _STLP_ASSERT(_p_size > 0) }
545
546  ~_Rope_RopeFunction() {
547    this->_M_free_c_string();
548    if (_M_delete_when_done) {
549      delete _M_fn;
550    }
551  }
552};
553
554/*
555 * Substring results are usually represented using just
556 * concatenation nodes.  But in the case of very long flat ropes
557 * or ropes with a functional representation that isn't practical.
558 * In that case, we represent the __result as a special case of
559 * RopeFunction, whose char_producer points back to the rope itself.
560 * In all cases except repeated substring operations and
561 * deallocation, we treat the __result as a RopeFunction.
562 */
563template<class _CharT, class _Alloc>
564struct _Rope_RopeSubstring : public char_producer<_CharT>, public _Rope_RopeFunction<_CharT,_Alloc> {
565public:
566  // XXX this whole class should be rewritten.
567  typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep;
568  _RopeRep *_M_base;      // not 0
569  size_t _M_start;
570  /* virtual */ void operator()(size_t __start_pos, size_t __req_len,
571                                _CharT* __buffer) {
572    typedef _Rope_RopeFunction<_CharT,_Alloc> _RopeFunction;
573    typedef _Rope_RopeLeaf<_CharT,_Alloc> _RopeLeaf;
574    switch (_M_base->_M_tag) {
575    case _RopeRep::_S_function:
576    case _RopeRep::_S_substringfn:
577      {
578        char_producer<_CharT>* __fn =
579          __STATIC_CAST(_RopeFunction*, _M_base)->_M_fn;
580        _STLP_ASSERT(__start_pos + __req_len <= this->_M_size._M_data)
581        _STLP_ASSERT(_M_start + this->_M_size._M_data <= _M_base->_M_size._M_data)
582        (*__fn)(__start_pos + _M_start, __req_len, __buffer);
583      }
584      break;
585    case _RopeRep::_S_leaf:
586      {
587        _CharT* __s =
588          __STATIC_CAST(_RopeLeaf*, _M_base)->_M_data;
589        _STLP_PRIV __ucopy_n(__s + __start_pos + _M_start, __req_len, __buffer);
590      }
591      break;
592    default:
593      _STLP_ASSERT(false)
594        ;
595    }
596  }
597
598  _STLP_FORCE_ALLOCATORS(_CharT, _Alloc)
599  typedef typename _RopeRep::allocator_type allocator_type;
600
601  _Rope_RopeSubstring(_RopeRep* __b, size_t __s, size_t __l, allocator_type __a)
602    : _Rope_RopeFunction<_CharT,_Alloc>(this, __l, false, __a),
603      _M_base(__b), _M_start(__s) {
604    _STLP_ASSERT(__l > 0)
605    _STLP_ASSERT(__s + __l <= __b->_M_size._M_data)
606    _M_base->_M_ref_nonnil();
607    this->_M_tag = _RopeRep::_S_substringfn;
608  }
609  virtual ~_Rope_RopeSubstring()
610  { _M_base->_M_unref_nonnil(); }
611};
612
613/*
614 * Self-destructing pointers to Rope_rep.
615 * These are not conventional smart pointers.  Their
616 * only purpose in life is to ensure that unref is called
617 * on the pointer either at normal exit or if an exception
618 * is raised.  It is the caller's responsibility to
619 * adjust reference counts when these pointers are initialized
620 * or assigned to.  (This convention significantly reduces
621 * the number of potentially expensive reference count
622 * updates.)
623 */
624template<class _CharT, class _Alloc>
625struct _Rope_self_destruct_ptr {
626  _Rope_RopeRep<_CharT,_Alloc>* _M_ptr;
627  ~_Rope_self_destruct_ptr()
628  { _Rope_RopeRep<_CharT,_Alloc>::_S_unref(_M_ptr); }
629#   ifdef _STLP_USE_EXCEPTIONS
630  _Rope_self_destruct_ptr() : _M_ptr(0) {}
631#   else
632  _Rope_self_destruct_ptr() {}
633#   endif
634  _Rope_self_destruct_ptr(_Rope_RopeRep<_CharT,_Alloc>* __p) : _M_ptr(__p) {}
635  _Rope_RopeRep<_CharT,_Alloc>& operator*() { return *_M_ptr; }
636  _Rope_RopeRep<_CharT,_Alloc>* operator->() { return _M_ptr; }
637  operator _Rope_RopeRep<_CharT,_Alloc>*() { return _M_ptr; }
638  _Rope_self_destruct_ptr<_CharT, _Alloc>&
639  operator= (_Rope_RopeRep<_CharT,_Alloc>* __x)
640  { _M_ptr = __x; return *this; }
641};
642
643/*
644 * Dereferencing a nonconst iterator has to return something
645 * that behaves almost like a reference.  It's not possible to
646 * return an actual reference since assignment requires extra
647 * work.  And we would get into the same problems as with the
648 * CD2 version of basic_string.
649 */
650template<class _CharT, class _Alloc>
651class _Rope_char_ref_proxy {
652  typedef _Rope_char_ref_proxy<_CharT, _Alloc> _Self;
653  friend class rope<_CharT,_Alloc>;
654  friend class _Rope_iterator<_CharT,_Alloc>;
655  friend class _Rope_char_ptr_proxy<_CharT,_Alloc>;
656  typedef _Rope_self_destruct_ptr<_CharT,_Alloc> _Self_destruct_ptr;
657  typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep;
658  typedef rope<_CharT,_Alloc> _My_rope;
659  size_t _M_pos;
660  _CharT _M_current;
661  bool _M_current_valid;
662  _My_rope* _M_root;     // The whole rope.
663public:
664  _Rope_char_ref_proxy(_My_rope* __r, size_t __p) :
665    _M_pos(__p), _M_current_valid(false), _M_root(__r) {}
666  _Rope_char_ref_proxy(const _Self& __x) :
667    _M_pos(__x._M_pos), _M_current_valid(false), _M_root(__x._M_root) {}
668  // Don't preserve cache if the reference can outlive the
669  // expression.  We claim that's not possible without calling
670  // a copy constructor or generating reference to a proxy
671  // reference.  We declare the latter to have undefined semantics.
672  _Rope_char_ref_proxy(_My_rope* __r, size_t __p, _CharT __c)
673    : _M_pos(__p), _M_current(__c), _M_current_valid(true), _M_root(__r) {}
674  inline operator _CharT () const;
675  _Self& operator= (_CharT __c);
676  _Rope_char_ptr_proxy<_CharT, _Alloc> operator& () const;
677  _Self& operator= (const _Self& __c) {
678    return operator=((_CharT)__c);
679  }
680};
681
682#ifdef _STLP_FUNCTION_TMPL_PARTIAL_ORDER
683template<class _CharT, class __Alloc>
684inline void swap(_Rope_char_ref_proxy <_CharT, __Alloc > __a,
685                 _Rope_char_ref_proxy <_CharT, __Alloc > __b) {
686  _CharT __tmp = __a;
687  __a = __b;
688  __b = __tmp;
689}
690#else
691// There is no really acceptable way to handle this.  The default
692// definition of swap doesn't work for proxy references.
693// It can't really be made to work, even with ugly hacks, since
694// the only unusual operation it uses is the copy constructor, which
695// is needed for other purposes.  We provide a macro for
696// full specializations, and instantiate the most common case.
697# define _ROPE_SWAP_SPECIALIZATION(_CharT, __Alloc) \
698    inline void swap(_Rope_char_ref_proxy <_CharT, __Alloc > __a, \
699                     _Rope_char_ref_proxy <_CharT, __Alloc > __b) { \
700        _CharT __tmp = __a; \
701        __a = __b; \
702        __b = __tmp; \
703    }
704
705_ROPE_SWAP_SPECIALIZATION(char, allocator<char>)
706
707# ifndef _STLP_NO_WCHAR_T
708_ROPE_SWAP_SPECIALIZATION(wchar_t, allocator<wchar_t>)
709# endif
710
711#endif /* !_STLP_FUNCTION_TMPL_PARTIAL_ORDER */
712
713template<class _CharT, class _Alloc>
714class _Rope_char_ptr_proxy {
715  // XXX this class should be rewritten.
716public:
717  typedef _Rope_char_ptr_proxy<_CharT, _Alloc> _Self;
718  friend class _Rope_char_ref_proxy<_CharT,_Alloc>;
719  size_t _M_pos;
720  rope<_CharT,_Alloc>* _M_root;     // The whole rope.
721
722  _Rope_char_ptr_proxy(const _Rope_char_ref_proxy<_CharT,_Alloc>& __x)
723    : _M_pos(__x._M_pos), _M_root(__x._M_root) {}
724  _Rope_char_ptr_proxy(const _Self& __x)
725    : _M_pos(__x._M_pos), _M_root(__x._M_root) {}
726  _Rope_char_ptr_proxy() {}
727  _Rope_char_ptr_proxy(_CharT* __x) : _M_pos(0), _M_root(0) {
728    _STLP_ASSERT(0 == __x)
729  }
730  _Self& operator= (const _Self& __x) {
731    _M_pos = __x._M_pos;
732    _M_root = __x._M_root;
733    return *this;
734  }
735
736  _Rope_char_ref_proxy<_CharT,_Alloc> operator*() const {
737    return _Rope_char_ref_proxy<_CharT,_Alloc>(_M_root, _M_pos);
738  }
739};
740
741
742/*
743 * Rope iterators:
744 * Unlike in the C version, we cache only part of the stack
745 * for rope iterators, since they must be efficiently copyable.
746 * When we run out of cache, we have to reconstruct the iterator
747 * value.
748 * Pointers from iterators are not included in reference counts.
749 * Iterators are assumed to be thread private.  Ropes can
750 * be shared.
751 */
752template<class _CharT, class _Alloc>
753class _Rope_iterator_base
754/*   : public random_access_iterator<_CharT, ptrdiff_t>  */
755{
756  friend class rope<_CharT,_Alloc>;
757  typedef _Rope_iterator_base<_CharT, _Alloc> _Self;
758  typedef _Rope_RopeConcatenation<_CharT,_Alloc> _RopeConcat;
759public:
760  typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep;
761
762  enum { _S_path_cache_len = 4 }; // Must be <= 9 because of _M_path_direction.
763  enum { _S_iterator_buf_len = 15 };
764  size_t _M_current_pos;
765  // The whole rope.
766  _RopeRep* _M_root;
767  // Starting position for current leaf
768  size_t _M_leaf_pos;
769  // Buffer possibly containing current char.
770  _CharT* _M_buf_start;
771  // Pointer to current char in buffer, != 0 ==> buffer valid.
772  _CharT* _M_buf_ptr;
773  // One past __last valid char in buffer.
774  _CharT* _M_buf_end;
775
776  // What follows is the path cache.  We go out of our
777  // way to make this compact.
778  // Path_end contains the bottom section of the path from
779  // the root to the current leaf.
780  struct {
781#  if defined (__BORLANDC__) && (__BORLANDC__ < 0x560)
782    _RopeRep const*_M_data[4];
783#  else
784    _RopeRep const*_M_data[_S_path_cache_len];
785#  endif
786  } _M_path_end;
787  // Last valid __pos in path_end;
788  // _M_path_end[0] ... _M_path_end[_M_leaf_index-1]
789  // point to concatenation nodes.
790  int _M_leaf_index;
791  // (_M_path_directions >> __i) & 1 is 1
792  // if we got from _M_path_end[leaf_index - __i - 1]
793  // to _M_path_end[leaf_index - __i] by going to the
794  // __right. Assumes path_cache_len <= 9.
795  unsigned char _M_path_directions;
796  // Short buffer for surrounding chars.
797  // This is useful primarily for
798  // RopeFunctions.  We put the buffer
799  // here to avoid locking in the
800  // multithreaded case.
801  // The cached path is generally assumed to be valid
802  // only if the buffer is valid.
803  struct {
804#  if defined (__BORLANDC__) && (__BORLANDC__ < 0x560)
805    _CharT _M_data[15];
806#  else
807    _CharT _M_data[_S_iterator_buf_len];
808#  endif
809  } _M_tmp_buf;
810
811  // Set buffer contents given path cache.
812  static void _S_setbuf(_Rope_iterator_base<_CharT, _Alloc>& __x);
813  // Set buffer contents and path cache.
814  static void _S_setcache(_Rope_iterator_base<_CharT, _Alloc>& __x);
815  // As above, but assumes path cache is valid for previous posn.
816  static void _S_setcache_for_incr(_Rope_iterator_base<_CharT, _Alloc>& __x);
817  _Rope_iterator_base() {}
818  _Rope_iterator_base(_RopeRep* __root, size_t __pos)
819    : _M_current_pos(__pos),_M_root(__root),  _M_buf_ptr(0) {}
820  void _M_incr(size_t __n);
821  void _M_decr(size_t __n);
822public:
823  size_t index() const { return _M_current_pos; }
824private:
825  void _M_copy_buf(const _Self& __x) {
826    _M_tmp_buf = __x._M_tmp_buf;
827    if (__x._M_buf_start == __x._M_tmp_buf._M_data) {
828      _M_buf_start = _M_tmp_buf._M_data;
829      _M_buf_end = _M_buf_start + (__x._M_buf_end - __x._M_buf_start);
830      _M_buf_ptr = _M_buf_start + (__x._M_buf_ptr - __x._M_buf_start);
831    } else {
832      _M_buf_end = __x._M_buf_end;
833    }
834  }
835
836public:
837  _Rope_iterator_base(const _Self& __x) :
838      _M_current_pos(__x._M_current_pos),
839      _M_root(__x._M_root),
840      _M_leaf_pos( __x._M_leaf_pos ),
841      _M_buf_start(__x._M_buf_start),
842      _M_buf_ptr(__x._M_buf_ptr),
843      _M_path_end(__x._M_path_end),
844      _M_leaf_index(__x._M_leaf_index),
845      _M_path_directions(__x._M_path_directions)
846      {
847        if (0 != __x._M_buf_ptr) {
848          _M_copy_buf(__x);
849        }
850      }
851  _Self& operator = (const _Self& __x)
852      {
853        _M_current_pos = __x._M_current_pos;
854        _M_root = __x._M_root;
855        _M_buf_start = __x._M_buf_start;
856        _M_buf_ptr = __x._M_buf_ptr;
857        _M_path_end = __x._M_path_end;
858        _M_leaf_index = __x._M_leaf_index;
859        _M_path_directions = __x._M_path_directions;
860        _M_leaf_pos = __x._M_leaf_pos;
861        if (0 != __x._M_buf_ptr) {
862          _M_copy_buf(__x);
863        }
864        return *this;
865      }
866};
867
868template<class _CharT, class _Alloc> class _Rope_iterator;
869
870template<class _CharT, class _Alloc>
871class _Rope_const_iterator : public _Rope_iterator_base<_CharT,_Alloc> {
872  friend class rope<_CharT,_Alloc>;
873  typedef  _Rope_const_iterator<_CharT, _Alloc> _Self;
874  typedef _Rope_iterator_base<_CharT,_Alloc> _Base;
875  //  protected:
876public:
877#   ifndef _STLP_HAS_NO_NAMESPACES
878  typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep;
879  // The one from the base class may not be directly visible.
880#   endif
881  _Rope_const_iterator(const _RopeRep* __root, size_t __pos):
882    _Rope_iterator_base<_CharT,_Alloc>(__CONST_CAST(_RopeRep*,__root), __pos)
883    // Only nonconst iterators modify root ref count
884  {}
885public:
886  typedef _CharT reference;   // Really a value.  Returning a reference
887                              // Would be a mess, since it would have
888                              // to be included in refcount.
889  typedef const _CharT* pointer;
890  typedef _CharT value_type;
891  typedef ptrdiff_t difference_type;
892  typedef random_access_iterator_tag iterator_category;
893
894public:
895  _Rope_const_iterator() {}
896  _Rope_const_iterator(const _Self& __x) :
897    _Rope_iterator_base<_CharT,_Alloc>(__x) { }
898  _Rope_const_iterator(const _Rope_iterator<_CharT,_Alloc>& __x):
899    _Rope_iterator_base<_CharT,_Alloc>(__x) {}
900  _Rope_const_iterator(const rope<_CharT,_Alloc>& __r, size_t __pos) :
901    _Rope_iterator_base<_CharT,_Alloc>(__r._M_tree_ptr._M_data, __pos) {}
902  _Self& operator= (const _Self& __x) {
903    _Base::operator=(__x);
904    return *this;
905  }
906  reference operator*() {
907    if (0 == this->_M_buf_ptr)
908#if defined(__clang__) || (defined(__GNUC__) && (__GNUC__ > 4 || __GNUC_MINOR__ >= 7))
909      this->_S_setcache(*this);
910#elif !defined (__DMC__)
911      _S_setcache(*this);
912#else
913    { _Rope_iterator_base<_CharT, _Alloc>* __x = this; _S_setcache(*__x); }
914#endif
915    return *(this->_M_buf_ptr);
916  }
917  _Self& operator++()
918      {
919        if ( this->_M_buf_ptr != 0 ) {
920          _CharT *__next = this->_M_buf_ptr + 1;
921          if ( __next < this->_M_buf_end ) {
922            this->_M_buf_ptr = __next;
923            ++this->_M_current_pos;
924            return *this;
925          }
926        }
927        this->_M_incr(1);
928        return *this;
929      }
930  _Self& operator+=(ptrdiff_t __n) {
931    if (__n >= 0) {
932      this->_M_incr(__n);
933    } else {
934      this->_M_decr(-__n);
935    }
936    return *this;
937  }
938  _Self& operator--() {
939    this->_M_decr(1);
940    return *this;
941  }
942  _Self& operator-=(ptrdiff_t __n) {
943    if (__n >= 0) {
944      this->_M_decr(__n);
945    } else {
946      this->_M_incr(-__n);
947    }
948    return *this;
949  }
950  _Self operator++(int) {
951    size_t __old_pos = this->_M_current_pos;
952    this->_M_incr(1);
953    return _Rope_const_iterator<_CharT,_Alloc>(this->_M_root, __old_pos);
954    // This makes a subsequent dereference expensive.
955    // Perhaps we should instead copy the iterator
956    // if it has a valid cache?
957  }
958  _Self operator--(int) {
959    size_t __old_pos = this->_M_current_pos;
960    this->_M_decr(1);
961    return _Rope_const_iterator<_CharT,_Alloc>(this->_M_root, __old_pos);
962  }
963  inline reference operator[](size_t __n);
964};
965
966template<class _CharT, class _Alloc>
967class _Rope_iterator : public _Rope_iterator_base<_CharT,_Alloc> {
968  friend class rope<_CharT,_Alloc>;
969  typedef _Rope_iterator<_CharT, _Alloc> _Self;
970  typedef _Rope_iterator_base<_CharT,_Alloc> _Base;
971  typedef _Rope_RopeRep<_CharT,_Alloc> _RopeRep;
972
973public:
974  rope<_CharT,_Alloc>* _M_root_rope;
975  // root is treated as a cached version of this,
976  // and is used to detect changes to the underlying
977  // rope.
978  // Root is included in the reference count.
979  // This is necessary so that we can detect changes reliably.
980  // Unfortunately, it requires careful bookkeeping for the
981  // nonGC case.
982  _Rope_iterator(rope<_CharT,_Alloc>* __r, size_t __pos);
983
984  void _M_check();
985public:
986  typedef _Rope_char_ref_proxy<_CharT,_Alloc>  reference;
987  typedef _Rope_char_ref_proxy<_CharT,_Alloc>* pointer;
988  typedef _CharT value_type;
989  typedef ptrdiff_t difference_type;
990  typedef random_access_iterator_tag iterator_category;
991public:
992  ~_Rope_iterator() {  //*TY 5/6/00 - added dtor to balance reference count
993    _RopeRep::_S_unref(this->_M_root);
994  }
995
996  rope<_CharT,_Alloc>& container() { return *_M_root_rope; }
997  _Rope_iterator() {
998    this->_M_root = 0;  // Needed for reference counting.
999  }
1000  _Rope_iterator(const  _Self& __x) :
1001    _Rope_iterator_base<_CharT,_Alloc>(__x) {
1002    _M_root_rope = __x._M_root_rope;
1003    _RopeRep::_S_ref(this->_M_root);
1004  }
1005  _Rope_iterator(rope<_CharT,_Alloc>& __r, size_t __pos);
1006  _Self& operator= (const  _Self& __x) {
1007    _RopeRep* __old = this->_M_root;
1008    _RopeRep::_S_ref(__x._M_root);
1009    _Base::operator=(__x);
1010    _M_root_rope = __x._M_root_rope;
1011    _RopeRep::_S_unref(__old);
1012    return *this;
1013  }
1014  reference operator*() {
1015    _M_check();
1016    if (0 == this->_M_buf_ptr) {
1017      return reference(_M_root_rope, this->_M_current_pos);
1018    } else {
1019      return reference(_M_root_rope, this->_M_current_pos, *(this->_M_buf_ptr));
1020    }
1021  }
1022  _Self& operator++() {
1023    this->_M_incr(1);
1024    return *this;
1025  }
1026  _Self& operator+=(ptrdiff_t __n) {
1027    if (__n >= 0) {
1028      this->_M_incr(__n);
1029    } else {
1030      this->_M_decr(-__n);
1031    }
1032    return *this;
1033  }
1034  _Self& operator--() {
1035    this->_M_decr(1);
1036    return *this;
1037  }
1038  _Self& operator-=(ptrdiff_t __n) {
1039    if (__n >= 0) {
1040      this->_M_decr(__n);
1041    } else {
1042      this->_M_incr(-__n);
1043    }
1044    return *this;
1045  }
1046  _Self operator++(int) {
1047    size_t __old_pos = this->_M_current_pos;
1048    this->_M_incr(1);
1049    return _Self(_M_root_rope, __old_pos);
1050  }
1051  _Self operator--(int) {
1052    size_t __old_pos = this->_M_current_pos;
1053    this->_M_decr(1);
1054    return _Self(_M_root_rope, __old_pos);
1055  }
1056  reference operator[](ptrdiff_t __n) {
1057    return reference(_M_root_rope, this->_M_current_pos + __n);
1058  }
1059};
1060
1061# ifdef _STLP_USE_OLD_HP_ITERATOR_QUERIES
1062template <class _CharT, class _Alloc>
1063inline random_access_iterator_tag
1064iterator_category(const _Rope_iterator<_CharT,_Alloc>&) {  return random_access_iterator_tag();}
1065template <class _CharT, class _Alloc>
1066inline _CharT* value_type(const _Rope_iterator<_CharT,_Alloc>&) { return 0; }
1067template <class _CharT, class _Alloc>
1068inline ptrdiff_t* distance_type(const _Rope_iterator<_CharT,_Alloc>&) { return 0; }
1069template <class _CharT, class _Alloc>
1070inline random_access_iterator_tag
1071iterator_category(const _Rope_const_iterator<_CharT,_Alloc>&) { return random_access_iterator_tag(); }
1072template <class _CharT, class _Alloc>
1073inline _CharT* value_type(const _Rope_const_iterator<_CharT,_Alloc>&) { return 0; }
1074template <class _CharT, class _Alloc>
1075inline ptrdiff_t* distance_type(const _Rope_const_iterator<_CharT,_Alloc>&) { return 0; }
1076#endif /* _STLP_USE_OLD_HP_ITERATOR_QUERIES */
1077
1078template <class _CharT, class _Alloc, class _CharConsumer>
1079bool _S_apply_to_pieces(_CharConsumer& __c,
1080                        _Rope_RopeRep<_CharT, _Alloc> *__r,
1081                        size_t __begin, size_t __end);
1082                        // begin and end are assumed to be in range.
1083
1084template <class _CharT, class _Alloc>
1085class rope
1086#if defined (_STLP_USE_PARTIAL_SPEC_WORKAROUND)
1087           : public __stlport_class<rope<_CharT, _Alloc> >
1088#endif
1089{
1090  typedef rope<_CharT,_Alloc> _Self;
1091public:
1092  typedef _CharT value_type;
1093  typedef ptrdiff_t difference_type;
1094  typedef size_t size_type;
1095  typedef _CharT const_reference;
1096  typedef const _CharT* const_pointer;
1097  typedef _Rope_iterator<_CharT,_Alloc> iterator;
1098  typedef _Rope_const_iterator<_CharT,_Alloc> const_iterator;
1099  typedef _Rope_char_ref_proxy<_CharT,_Alloc> reference;
1100  typedef _Rope_char_ptr_proxy<_CharT,_Alloc> pointer;
1101
1102  friend class _Rope_iterator<_CharT,_Alloc>;
1103  friend class _Rope_const_iterator<_CharT,_Alloc>;
1104  friend struct _Rope_RopeRep<_CharT,_Alloc>;
1105  friend class _Rope_iterator_base<_CharT,_Alloc>;
1106  friend class _Rope_char_ptr_proxy<_CharT,_Alloc>;
1107  friend class _Rope_char_ref_proxy<_CharT,_Alloc>;
1108  friend struct _Rope_RopeSubstring<_CharT,_Alloc>;
1109
1110  _STLP_DECLARE_RANDOM_ACCESS_REVERSE_ITERATORS;
1111
1112protected:
1113  typedef _CharT* _Cstrptr;
1114
1115  static _CharT _S_empty_c_str[1];
1116
1117  enum { _S_copy_max = 23 };
1118  // For strings shorter than _S_copy_max, we copy to
1119  // concatenate.
1120
1121  typedef _Rope_RopeRep<_CharT, _Alloc> _RopeRep;
1122  typedef typename _RopeRep::_IsBasicCharType _IsBasicCharType;
1123
1124public:
1125  _STLP_FORCE_ALLOCATORS(_CharT, _Alloc)
1126  typedef _Alloc allocator_type;
1127
1128public:
1129  // The only data member of a rope:
1130  _STLP_PRIV _STLP_alloc_proxy<_RopeRep*, _CharT, allocator_type> _M_tree_ptr;
1131
1132public:
1133  allocator_type get_allocator() const { return allocator_type(_M_tree_ptr); }
1134
1135public:
1136  typedef _Rope_RopeConcatenation<_CharT,_Alloc> _RopeConcatenation;
1137  typedef _Rope_RopeLeaf<_CharT,_Alloc> _RopeLeaf;
1138  typedef _Rope_RopeFunction<_CharT,_Alloc> _RopeFunction;
1139  typedef _Rope_RopeSubstring<_CharT,_Alloc> _RopeSubstring;
1140
1141  // Retrieve a character at the indicated position.
1142  static _CharT _S_fetch(_RopeRep* __r, size_type __pos);
1143
1144  // Obtain a pointer to the character at the indicated position.
1145  // The pointer can be used to change the character.
1146  // If such a pointer cannot be produced, as is frequently the
1147  // case, 0 is returned instead.
1148  // (Returns nonzero only if all nodes in the path have a refcount
1149  // of 1.)
1150  static _CharT* _S_fetch_ptr(_RopeRep* __r, size_type __pos);
1151
1152  static void _S_unref(_RopeRep* __t) {
1153    _RopeRep::_S_unref(__t);
1154  }
1155  static void _S_ref(_RopeRep* __t) {
1156    _RopeRep::_S_ref(__t);
1157  }
1158
1159  typedef _Rope_self_destruct_ptr<_CharT,_Alloc> _Self_destruct_ptr;
1160
1161  // _Result is counted in refcount.
1162  static _RopeRep* _S_substring(_RopeRep* __base,
1163                                size_t __start, size_t __endp1);
1164
1165  static _RopeRep* _S_concat_char_iter(_RopeRep* __r,
1166                                       const _CharT* __iter, size_t __slen);
1167  // Concatenate rope and char ptr, copying __s.
1168  // Should really take an arbitrary iterator.
1169  // Result is counted in refcount.
1170  static _RopeRep* _S_destr_concat_char_iter(_RopeRep* __r,
1171                                             const _CharT* __iter, size_t __slen);
1172    // As above, but one reference to __r is about to be
1173    // destroyed.  Thus the pieces may be recycled if all
1174    // relevent reference counts are 1.
1175
1176  // General concatenation on _RopeRep.  _Result
1177  // has refcount of 1.  Adjusts argument refcounts.
1178  static _RopeRep* _S_concat_rep(_RopeRep* __left, _RopeRep* __right);
1179
1180public:
1181#if defined (_STLP_MEMBER_TEMPLATES)
1182  template <class _CharConsumer>
1183#else
1184  typedef _Rope_char_consumer<_CharT> _CharConsumer;
1185#endif
1186  void apply_to_pieces(size_t __begin, size_t __end,
1187                       _CharConsumer& __c) const
1188  { _S_apply_to_pieces(__c, _M_tree_ptr._M_data, __begin, __end); }
1189
1190protected:
1191
1192  static size_t _S_rounded_up_size(size_t __n)
1193  { return _RopeRep::_S_rounded_up_size(__n); }
1194
1195  // Allocate and construct a RopeLeaf using the supplied allocator
1196  // Takes ownership of s instead of copying.
1197  static _RopeLeaf* _S_new_RopeLeaf(_CharT *__s,
1198                                    size_t _p_size, allocator_type __a) {
1199    _RopeLeaf* __space = _STLP_CREATE_ALLOCATOR(allocator_type, __a,
1200                                                _RopeLeaf).allocate(1);
1201    _STLP_TRY {
1202      new(__space) _RopeLeaf(__s, _p_size, __a);
1203    }
1204   _STLP_UNWIND(_STLP_CREATE_ALLOCATOR(allocator_type,__a,
1205                                       _RopeLeaf).deallocate(__space, 1))
1206    return __space;
1207  }
1208
1209  static _RopeConcatenation* _S_new_RopeConcatenation(_RopeRep* __left, _RopeRep* __right,
1210                                                      allocator_type __a) {
1211   _RopeConcatenation* __space = _STLP_CREATE_ALLOCATOR(allocator_type, __a,
1212                                                        _RopeConcatenation).allocate(1);
1213    return new(__space) _RopeConcatenation(__left, __right, __a);
1214  }
1215
1216  static _RopeFunction* _S_new_RopeFunction(char_producer<_CharT>* __f,
1217                                            size_t _p_size, bool __d, allocator_type __a) {
1218   _RopeFunction* __space = _STLP_CREATE_ALLOCATOR(allocator_type, __a,
1219                                                   _RopeFunction).allocate(1);
1220    return new(__space) _RopeFunction(__f, _p_size, __d, __a);
1221  }
1222
1223  static _RopeSubstring* _S_new_RopeSubstring(_Rope_RopeRep<_CharT,_Alloc>* __b, size_t __s,
1224                                              size_t __l, allocator_type __a) {
1225   _RopeSubstring* __space = _STLP_CREATE_ALLOCATOR(allocator_type, __a,
1226                                                    _RopeSubstring).allocate(1);
1227    return new(__space) _RopeSubstring(__b, __s, __l, __a);
1228  }
1229
1230  static
1231  _RopeLeaf* _S_RopeLeaf_from_unowned_char_ptr(const _CharT *__s,
1232                                               size_t _p_size, allocator_type __a) {
1233    if (0 == _p_size) return 0;
1234
1235   _CharT* __buf = _STLP_CREATE_ALLOCATOR(allocator_type,__a, _CharT).allocate(_S_rounded_up_size(_p_size));
1236
1237    _STLP_PRIV __ucopy_n(__s, _p_size, __buf);
1238    _S_construct_null(__buf + _p_size);
1239
1240    _STLP_TRY {
1241      return _S_new_RopeLeaf(__buf, _p_size, __a);
1242    }
1243    _STLP_UNWIND(_RopeRep::_S_free_string(__buf, _p_size, __a))
1244    _STLP_RET_AFTER_THROW(0)
1245  }
1246
1247
1248  // Concatenation of nonempty strings.
1249  // Always builds a concatenation node.
1250  // Rebalances if the result is too deep.
1251  // Result has refcount 1.
1252  // Does not increment left and right ref counts even though
1253  // they are referenced.
1254  static _RopeRep*
1255  _S_tree_concat(_RopeRep* __left, _RopeRep* __right);
1256
1257  // Concatenation helper functions
1258  static _RopeLeaf*
1259  _S_leaf_concat_char_iter(_RopeLeaf* __r,
1260                           const _CharT* __iter, size_t __slen);
1261  // Concatenate by copying leaf.
1262  // should take an arbitrary iterator
1263  // result has refcount 1.
1264  static _RopeLeaf* _S_destr_leaf_concat_char_iter
1265  (_RopeLeaf* __r, const _CharT* __iter, size_t __slen);
1266  // A version that potentially clobbers __r if __r->_M_ref_count == 1.
1267
1268
1269  // A helper function for exponentiating strings.
1270  // This uses a nonstandard refcount convention.
1271  // The result has refcount 0.
1272  typedef _STLP_PRIV _Rope_Concat_fn<_CharT,_Alloc> _Concat_fn;
1273#if !defined (__GNUC__) || (__GNUC__ < 3)
1274  friend _Concat_fn;
1275#else
1276  friend struct _STLP_PRIV _Rope_Concat_fn<_CharT,_Alloc>;
1277#endif
1278
1279public:
1280  static size_t _S_char_ptr_len(const _CharT* __s) {
1281    return char_traits<_CharT>::length(__s);
1282  }
1283
1284public: /* for operators */
1285  rope(_RopeRep* __t, const allocator_type& __a = allocator_type())
1286    : _M_tree_ptr(__a, __t) { }
1287private:
1288  // Copy __r to the _CharT buffer.
1289  // Returns __buffer + __r->_M_size._M_data.
1290  // Assumes that buffer is uninitialized.
1291  static _CharT* _S_flatten(_RopeRep* __r, _CharT* __buffer);
1292
1293  // Again, with explicit starting position and length.
1294  // Assumes that buffer is uninitialized.
1295  static _CharT* _S_flatten(_RopeRep* __r,
1296                            size_t __start, size_t __len,
1297                            _CharT* __buffer);
1298
1299  // fbp : HP aCC prohibits access to protected min_len from within static methods ( ?? )
1300public:
1301  static const unsigned long _S_min_len[__ROPE_DEPTH_SIZE];
1302protected:
1303  static bool _S_is_balanced(_RopeRep* __r)
1304  { return (__r->_M_size._M_data >= _S_min_len[__r->_M_depth]); }
1305
1306  static bool _S_is_almost_balanced(_RopeRep* __r) {
1307    return (__r->_M_depth == 0 ||
1308            __r->_M_size._M_data >= _S_min_len[__r->_M_depth - 1]);
1309  }
1310
1311  static bool _S_is_roughly_balanced(_RopeRep* __r) {
1312    return (__r->_M_depth <= 1 ||
1313            __r->_M_size._M_data >= _S_min_len[__r->_M_depth - 2]);
1314  }
1315
1316  // Assumes the result is not empty.
1317  static _RopeRep* _S_concat_and_set_balanced(_RopeRep* __left,
1318                                              _RopeRep* __right) {
1319    _RopeRep* __result = _S_concat_rep(__left, __right);
1320    if (_S_is_balanced(__result)) __result->_M_is_balanced = true;
1321    return __result;
1322  }
1323
1324  // The basic rebalancing operation.  Logically copies the
1325  // rope.  The result has refcount of 1.  The client will
1326  // usually decrement the reference count of __r.
1327  // The result is within height 2 of balanced by the above
1328  // definition.
1329  static _RopeRep* _S_balance(_RopeRep* __r);
1330
1331  // Add all unbalanced subtrees to the forest of balanceed trees.
1332  // Used only by balance.
1333  static void _S_add_to_forest(_RopeRep*__r, _RopeRep** __forest);
1334
1335  // Add __r to forest, assuming __r is already balanced.
1336  static void _S_add_leaf_to_forest(_RopeRep* __r, _RopeRep** __forest);
1337
1338#ifdef _STLP_DEBUG
1339  // Print to stdout, exposing structure
1340  static void _S_dump(_RopeRep* __r, int __indent = 0);
1341#endif
1342
1343  // Return -1, 0, or 1 if __x < __y, __x == __y, or __x > __y resp.
1344  static int _S_compare(const _RopeRep* __x, const _RopeRep* __y);
1345
1346  void _STLP_FUNCTION_THROWS _M_throw_out_of_range() const;
1347
1348  void _M_reset(_RopeRep* __r) {
1349    //if (__r != _M_tree_ptr._M_data) {
1350      _S_unref(_M_tree_ptr._M_data);
1351      _M_tree_ptr._M_data = __r;
1352    //}
1353  }
1354
1355public:
1356  bool empty() const { return 0 == _M_tree_ptr._M_data; }
1357
1358  // Comparison member function.  This is public only for those
1359  // clients that need a ternary comparison.  Others
1360  // should use the comparison operators below.
1361  int compare(const _Self& __y) const {
1362    return _S_compare(_M_tree_ptr._M_data, __y._M_tree_ptr._M_data);
1363  }
1364
1365  rope(const _CharT* __s, const allocator_type& __a = allocator_type())
1366    : _M_tree_ptr(__a, _S_RopeLeaf_from_unowned_char_ptr(__s, _S_char_ptr_len(__s),__a))
1367  {}
1368
1369  rope(const _CharT* __s, size_t __len,
1370       const allocator_type& __a = allocator_type())
1371    : _M_tree_ptr(__a, (_S_RopeLeaf_from_unowned_char_ptr(__s, __len, __a)))
1372  {}
1373
1374  // Should perhaps be templatized with respect to the iterator type
1375  // and use Sequence_buffer.  (It should perhaps use sequence_buffer
1376  // even now.)
1377  rope(const _CharT *__s, const _CharT *__e,
1378       const allocator_type& __a = allocator_type())
1379    : _M_tree_ptr(__a, _S_RopeLeaf_from_unowned_char_ptr(__s, __e - __s, __a))
1380  {}
1381
1382  rope(const const_iterator& __s, const const_iterator& __e,
1383       const allocator_type& __a = allocator_type())
1384    : _M_tree_ptr(__a, _S_substring(__s._M_root, __s._M_current_pos,
1385                                    __e._M_current_pos))
1386  {}
1387
1388  rope(const iterator& __s, const iterator& __e,
1389       const allocator_type& __a = allocator_type())
1390    : _M_tree_ptr(__a, _S_substring(__s._M_root, __s._M_current_pos,
1391                                    __e._M_current_pos))
1392  {}
1393
1394  rope(_CharT __c, const allocator_type& __a = allocator_type())
1395    : _M_tree_ptr(__a, (_RopeRep*)0) {
1396    _CharT* __buf = _M_tree_ptr.allocate(_S_rounded_up_size(1));
1397
1398    _Copy_Construct(__buf, __c);
1399    _S_construct_null(__buf + 1);
1400
1401    _STLP_TRY {
1402      _M_tree_ptr._M_data = _S_new_RopeLeaf(__buf, 1, __a);
1403    }
1404    _STLP_UNWIND(_RopeRep::_S_free_string(__buf, 1, __a))
1405  }
1406
1407  rope(size_t __n, _CharT __c,
1408       const allocator_type& __a = allocator_type()):
1409    _M_tree_ptr(__a, (_RopeRep*)0) {
1410    if (0 == __n)
1411      return;
1412
1413    rope<_CharT,_Alloc> __result;
1414# define  __exponentiate_threshold size_t(32)
1415    _RopeRep* __remainder;
1416    rope<_CharT,_Alloc> __remainder_rope;
1417
1418    // gcc-2.7.2 bugs
1419    typedef _STLP_PRIV _Rope_Concat_fn<_CharT,_Alloc> _Concat_fn;
1420
1421    size_t __exponent = __n / __exponentiate_threshold;
1422    size_t __rest = __n % __exponentiate_threshold;
1423    if (0 == __rest) {
1424      __remainder = 0;
1425    } else {
1426      _CharT* __rest_buffer = _M_tree_ptr.allocate(_S_rounded_up_size(__rest));
1427      uninitialized_fill_n(__rest_buffer, __rest, __c);
1428      _S_construct_null(__rest_buffer + __rest);
1429      _STLP_TRY {
1430        __remainder = _S_new_RopeLeaf(__rest_buffer, __rest, __a);
1431      }
1432      _STLP_UNWIND(_RopeRep::_S_free_string(__rest_buffer, __rest, __a))
1433    }
1434    __remainder_rope._M_tree_ptr._M_data = __remainder;
1435    if (__exponent != 0) {
1436      _CharT* __base_buffer = _M_tree_ptr.allocate(_S_rounded_up_size(__exponentiate_threshold));
1437      _RopeLeaf* __base_leaf;
1438      rope<_CharT,_Alloc> __base_rope;
1439      uninitialized_fill_n(__base_buffer, __exponentiate_threshold, __c);
1440      _S_construct_null(__base_buffer + __exponentiate_threshold);
1441      _STLP_TRY {
1442        __base_leaf = _S_new_RopeLeaf(__base_buffer,
1443                                      __exponentiate_threshold, __a);
1444      }
1445      _STLP_UNWIND(_RopeRep::_S_free_string(__base_buffer,
1446                                            __exponentiate_threshold, __a))
1447      __base_rope._M_tree_ptr._M_data = __base_leaf;
1448      if (1 == __exponent) {
1449        __result = __base_rope;
1450        // One each for base_rope and __result
1451        //_STLP_ASSERT(2 == __result._M_tree_ptr._M_data->_M_ref_count)
1452      } else {
1453        __result = _STLP_PRIV __power(__base_rope, __exponent, _Concat_fn());
1454      }
1455      if (0 != __remainder) {
1456        __result += __remainder_rope;
1457      }
1458    } else {
1459      __result = __remainder_rope;
1460    }
1461    _M_tree_ptr._M_data = __result._M_tree_ptr._M_data;
1462    _M_tree_ptr._M_data->_M_ref_nonnil();
1463# undef __exponentiate_threshold
1464  }
1465
1466  rope(const allocator_type& __a = allocator_type())
1467    : _M_tree_ptr(__a, (_RopeRep*)0) {}
1468
1469  // Construct a rope from a function that can compute its members
1470  rope(char_producer<_CharT> *__fn, size_t __len, bool __delete_fn,
1471       const allocator_type& __a = allocator_type())
1472    : _M_tree_ptr(__a, (_RopeRep*)0) {
1473    _M_tree_ptr._M_data = (0 == __len) ?
1474      0 : _S_new_RopeFunction(__fn, __len, __delete_fn, __a);
1475  }
1476
1477  rope(const _Self& __x)
1478    : _M_tree_ptr(__x._M_tree_ptr, __x._M_tree_ptr._M_data) {
1479    _S_ref(_M_tree_ptr._M_data);
1480  }
1481
1482#if !defined (_STLP_NO_MOVE_SEMANTIC)
1483  rope(__move_source<_Self> __src)
1484    : _M_tree_ptr(__src.get()._M_tree_ptr, __src.get()._M_tree_ptr._M_data) {
1485    __src.get()._M_tree_ptr._M_data = 0;
1486  }
1487#endif
1488
1489  ~rope() {
1490    _S_unref(_M_tree_ptr._M_data);
1491  }
1492
1493  _Self& operator=(const _Self& __x) {
1494    _STLP_ASSERT(get_allocator() == __x.get_allocator())
1495    _S_ref(__x._M_tree_ptr._M_data);
1496    _M_reset(__x._M_tree_ptr._M_data);
1497    return *this;
1498  }
1499
1500  void clear() {
1501    _S_unref(_M_tree_ptr._M_data);
1502    _M_tree_ptr._M_data = 0;
1503  }
1504  void push_back(_CharT __x) {
1505    _M_reset(_S_destr_concat_char_iter(_M_tree_ptr._M_data, &__x, 1));
1506  }
1507
1508  void pop_back() {
1509    _RopeRep* __old = _M_tree_ptr._M_data;
1510    _M_tree_ptr._M_data =
1511      _S_substring(_M_tree_ptr._M_data, 0, _M_tree_ptr._M_data->_M_size._M_data - 1);
1512    _S_unref(__old);
1513  }
1514
1515  _CharT back() const {
1516    return _S_fetch(_M_tree_ptr._M_data, _M_tree_ptr._M_data->_M_size._M_data - 1);
1517  }
1518
1519  void push_front(_CharT __x) {
1520    _RopeRep* __old = _M_tree_ptr._M_data;
1521    _RopeRep* __left =
1522      _S_RopeLeaf_from_unowned_char_ptr(&__x, 1, _M_tree_ptr);
1523    _STLP_TRY {
1524      _M_tree_ptr._M_data = _S_concat_rep(__left, _M_tree_ptr._M_data);
1525      _S_unref(__old);
1526      _S_unref(__left);
1527    }
1528    _STLP_UNWIND(_S_unref(__left))
1529  }
1530
1531  void pop_front() {
1532    _RopeRep* __old = _M_tree_ptr._M_data;
1533    _M_tree_ptr._M_data = _S_substring(_M_tree_ptr._M_data, 1, _M_tree_ptr._M_data->_M_size._M_data);
1534    _S_unref(__old);
1535  }
1536
1537  _CharT front() const {
1538    return _S_fetch(_M_tree_ptr._M_data, 0);
1539  }
1540
1541  void balance() {
1542    _RopeRep* __old = _M_tree_ptr._M_data;
1543    _M_tree_ptr._M_data = _S_balance(_M_tree_ptr._M_data);
1544    _S_unref(__old);
1545  }
1546
1547  void copy(_CharT* __buffer) const {
1548    _STLP_STD::_Destroy_Range(__buffer, __buffer + size());
1549    _S_flatten(_M_tree_ptr._M_data, __buffer);
1550  }
1551
1552  /*
1553   * This is the copy function from the standard, but
1554   * with the arguments reordered to make it consistent with the
1555   * rest of the interface.
1556   * Note that this guaranteed not to compile if the draft standard
1557   * order is assumed.
1558   */
1559  size_type copy(size_type __pos, size_type __n, _CharT* __buffer) const {
1560    size_t _p_size = size();
1561    size_t __len = (__pos + __n > _p_size? _p_size - __pos : __n);
1562
1563    _STLP_STD::_Destroy_Range(__buffer, __buffer + __len);
1564    _S_flatten(_M_tree_ptr._M_data, __pos, __len, __buffer);
1565    return __len;
1566  }
1567
1568# ifdef _STLP_DEBUG
1569  // Print to stdout, exposing structure.  May be useful for
1570  // performance debugging.
1571  void dump() {
1572    _S_dump(_M_tree_ptr._M_data);
1573  }
1574# endif
1575
1576  // Convert to 0 terminated string in new allocated memory.
1577  // Embedded 0s in the input do not terminate the copy.
1578  const _CharT* c_str() const;
1579
1580  // As above, but also use the flattened representation as the
1581  // the new rope representation.
1582  const _CharT* replace_with_c_str();
1583
1584  // Reclaim memory for the c_str generated flattened string.
1585  // Intentionally undocumented, since it's hard to say when this
1586  // is safe for multiple threads.
1587  void delete_c_str () {
1588    if (0 == _M_tree_ptr._M_data) return;
1589    if (_RopeRep::_S_leaf == _M_tree_ptr._M_data->_M_tag &&
1590        ((_RopeLeaf*)_M_tree_ptr._M_data)->_M_data ==
1591        _M_tree_ptr._M_data->_M_c_string) {
1592      // Representation shared
1593      return;
1594    }
1595    _M_tree_ptr._M_data->_M_free_c_string();
1596    _M_tree_ptr._M_data->_M_c_string = 0;
1597  }
1598
1599  _CharT operator[] (size_type __pos) const {
1600    return _S_fetch(_M_tree_ptr._M_data, __pos);
1601  }
1602
1603  _CharT at(size_type __pos) const {
1604    if (__pos >= size()) _M_throw_out_of_range();
1605    return (*this)[__pos];
1606  }
1607
1608  const_iterator begin() const {
1609    return(const_iterator(_M_tree_ptr._M_data, 0));
1610  }
1611
1612  // An easy way to get a const iterator from a non-const container.
1613  const_iterator const_begin() const {
1614    return(const_iterator(_M_tree_ptr._M_data, 0));
1615  }
1616
1617  const_iterator end() const {
1618    return(const_iterator(_M_tree_ptr._M_data, size()));
1619  }
1620
1621  const_iterator const_end() const {
1622    return(const_iterator(_M_tree_ptr._M_data, size()));
1623  }
1624
1625  size_type size() const {
1626    return(0 == _M_tree_ptr._M_data? 0 : _M_tree_ptr._M_data->_M_size._M_data);
1627  }
1628
1629  size_type length() const {
1630    return size();
1631  }
1632
1633  size_type max_size() const {
1634    return _S_min_len[__ROPE_MAX_DEPTH-1] - 1;
1635    //  Guarantees that the result can be sufficiently
1636    //  balanced.  Longer ropes will probably still work,
1637    //  but it's harder to make guarantees.
1638  }
1639
1640  const_reverse_iterator rbegin() const {
1641    return const_reverse_iterator(end());
1642  }
1643
1644  const_reverse_iterator const_rbegin() const {
1645    return const_reverse_iterator(end());
1646  }
1647
1648  const_reverse_iterator rend() const {
1649    return const_reverse_iterator(begin());
1650  }
1651
1652  const_reverse_iterator const_rend() const {
1653    return const_reverse_iterator(begin());
1654  }
1655  // The symmetric cases are intentionally omitted, since they're presumed
1656  // to be less common, and we don't handle them as well.
1657
1658  // The following should really be templatized.
1659  // The first argument should be an input iterator or
1660  // forward iterator with value_type _CharT.
1661  _Self& append(const _CharT* __iter, size_t __n) {
1662    _M_reset(_S_destr_concat_char_iter(_M_tree_ptr._M_data, __iter, __n));
1663    return *this;
1664  }
1665
1666  _Self& append(const _CharT* __c_string) {
1667    size_t __len = _S_char_ptr_len(__c_string);
1668    append(__c_string, __len);
1669    return *this;
1670  }
1671
1672  _Self& append(const _CharT* __s, const _CharT* __e) {
1673    _M_reset(_S_destr_concat_char_iter(_M_tree_ptr._M_data, __s, __e - __s));
1674    return *this;
1675  }
1676
1677  _Self& append(const_iterator __s, const_iterator __e) {
1678    _STLP_ASSERT(__s._M_root == __e._M_root)
1679    _STLP_ASSERT(get_allocator() == __s._M_root->get_allocator())
1680    _Self_destruct_ptr __appendee(_S_substring(__s._M_root, __s._M_current_pos, __e._M_current_pos));
1681    _M_reset(_S_concat_rep(_M_tree_ptr._M_data, (_RopeRep*)__appendee));
1682    return *this;
1683  }
1684
1685  _Self& append(_CharT __c) {
1686    _M_reset(_S_destr_concat_char_iter(_M_tree_ptr._M_data, &__c, 1));
1687    return *this;
1688  }
1689
1690  _Self& append() { return append(_CharT()); }  // XXX why?
1691
1692  _Self& append(const _Self& __y) {
1693    _STLP_ASSERT(__y.get_allocator() == get_allocator())
1694    _M_reset(_S_concat_rep(_M_tree_ptr._M_data, __y._M_tree_ptr._M_data));
1695    return *this;
1696  }
1697
1698  _Self& append(size_t __n, _CharT __c) {
1699    rope<_CharT,_Alloc> __last(__n, __c);
1700    return append(__last);
1701  }
1702
1703  void swap(_Self& __b) {
1704    _M_tree_ptr.swap(__b._M_tree_ptr);
1705  }
1706#if defined (_STLP_USE_PARTIAL_SPEC_WORKAROUND) && !defined (_STLP_FUNCTION_TMPL_PARTIAL_ORDER)
1707  void _M_swap_workaround(_Self& __x) { swap(__x); }
1708#endif
1709
1710protected:
1711  // Result is included in refcount.
1712  static _RopeRep* replace(_RopeRep* __old, size_t __pos1,
1713                           size_t __pos2, _RopeRep* __r) {
1714    if (0 == __old) { _S_ref(__r); return __r; }
1715    _Self_destruct_ptr __left(_S_substring(__old, 0, __pos1));
1716    _Self_destruct_ptr __right(_S_substring(__old, __pos2, __old->_M_size._M_data));
1717    _STLP_MPWFIX_TRY  //*TY 06/01/2000 -
1718    _RopeRep* __result;
1719
1720    if (0 == __r) {
1721      __result = _S_concat_rep(__left, __right);
1722    } else {
1723      _STLP_ASSERT(__old->get_allocator() == __r->get_allocator())
1724      _Self_destruct_ptr __left_result(_S_concat_rep(__left, __r));
1725      __result = _S_concat_rep(__left_result, __right);
1726    }
1727    return __result;
1728    _STLP_MPWFIX_CATCH  //*TY 06/01/2000 -
1729  }
1730
1731public:
1732  void insert(size_t __p, const _Self& __r) {
1733    if (__p > size()) _M_throw_out_of_range();
1734    _STLP_ASSERT(get_allocator() == __r.get_allocator())
1735    _M_reset(replace(_M_tree_ptr._M_data, __p, __p, __r._M_tree_ptr._M_data));
1736  }
1737
1738  void insert(size_t __p, size_t __n, _CharT __c) {
1739    rope<_CharT,_Alloc> __r(__n,__c);
1740    insert(__p, __r);
1741  }
1742
1743  void insert(size_t __p, const _CharT* __i, size_t __n) {
1744    if (__p > size()) _M_throw_out_of_range();
1745    _Self_destruct_ptr __left(_S_substring(_M_tree_ptr._M_data, 0, __p));
1746    _Self_destruct_ptr __right(_S_substring(_M_tree_ptr._M_data, __p, size()));
1747    _Self_destruct_ptr __left_result(
1748                                     _S_concat_char_iter(__left, __i, __n));
1749    // _S_ destr_concat_char_iter should be safe here.
1750    // But as it stands it's probably not a win, since __left
1751    // is likely to have additional references.
1752    _M_reset(_S_concat_rep(__left_result, __right));
1753  }
1754
1755  void insert(size_t __p, const _CharT* __c_string) {
1756    insert(__p, __c_string, _S_char_ptr_len(__c_string));
1757  }
1758
1759  void insert(size_t __p, _CharT __c) {
1760    insert(__p, &__c, 1);
1761  }
1762
1763  void insert(size_t __p) {
1764    _CharT __c = _CharT();
1765    insert(__p, &__c, 1);
1766  }
1767
1768  void insert(size_t __p, const _CharT* __i, const _CharT* __j) {
1769    _Self __r(__i, __j);
1770    insert(__p, __r);
1771  }
1772
1773  void insert(size_t __p, const const_iterator& __i,
1774                          const const_iterator& __j) {
1775    _Self __r(__i, __j);
1776    insert(__p, __r);
1777  }
1778
1779  void insert(size_t __p, const iterator& __i,
1780                          const iterator& __j) {
1781    _Self __r(__i, __j);
1782    insert(__p, __r);
1783  }
1784
1785  // (position, length) versions of replace operations:
1786  void replace(size_t __p, size_t __n, const _Self& __r) {
1787    if (__p > size()) _M_throw_out_of_range();
1788    _M_reset(replace(_M_tree_ptr._M_data, __p, __p + __n, __r._M_tree_ptr._M_data));
1789  }
1790
1791  void replace(size_t __p, size_t __n,
1792               const _CharT* __i, size_t __i_len) {
1793    _Self __r(__i, __i_len);
1794    replace(__p, __n, __r);
1795  }
1796
1797  void replace(size_t __p, size_t __n, _CharT __c) {
1798    _Self __r(__c);
1799    replace(__p, __n, __r);
1800  }
1801
1802  void replace(size_t __p, size_t __n, const _CharT* __c_string) {
1803    _Self __r(__c_string);
1804    replace(__p, __n, __r);
1805  }
1806
1807  void replace(size_t __p, size_t __n,
1808               const _CharT* __i, const _CharT* __j) {
1809    _Self __r(__i, __j);
1810    replace(__p, __n, __r);
1811  }
1812
1813  void replace(size_t __p, size_t __n,
1814               const const_iterator& __i, const const_iterator& __j) {
1815    _Self __r(__i, __j);
1816    replace(__p, __n, __r);
1817  }
1818
1819  void replace(size_t __p, size_t __n,
1820               const iterator& __i, const iterator& __j) {
1821    _Self __r(__i, __j);
1822    replace(__p, __n, __r);
1823  }
1824
1825  // Single character variants:
1826  void replace(size_t __p, _CharT __c) {
1827    if (__p > size()) _M_throw_out_of_range();
1828    iterator __i(this, __p);
1829    *__i = __c;
1830  }
1831
1832  void replace(size_t __p, const _Self& __r) {
1833    replace(__p, 1, __r);
1834  }
1835
1836  void replace(size_t __p, const _CharT* __i, size_t __i_len) {
1837    replace(__p, 1, __i, __i_len);
1838  }
1839
1840  void replace(size_t __p, const _CharT* __c_string) {
1841    replace(__p, 1, __c_string);
1842  }
1843
1844  void replace(size_t __p, const _CharT* __i, const _CharT* __j) {
1845    replace(__p, 1, __i, __j);
1846  }
1847
1848  void replace(size_t __p, const const_iterator& __i,
1849                           const const_iterator& __j) {
1850    replace(__p, 1, __i, __j);
1851  }
1852
1853  void replace(size_t __p, const iterator& __i,
1854                           const iterator& __j) {
1855    replace(__p, 1, __i, __j);
1856  }
1857
1858  // Erase, (position, size) variant.
1859  void erase(size_t __p, size_t __n) {
1860    if (__p > size()) _M_throw_out_of_range();
1861    _M_reset(replace(_M_tree_ptr._M_data, __p, __p + __n, 0));
1862  }
1863
1864  // Erase, single character
1865  void erase(size_t __p) {
1866    erase(__p, __p + 1);
1867  }
1868
1869  // Insert, iterator variants.
1870  iterator insert(const iterator& __p, const _Self& __r)
1871  { insert(__p.index(), __r); return __p; }
1872  iterator insert(const iterator& __p, size_t __n, _CharT __c)
1873  { insert(__p.index(), __n, __c); return __p; }
1874  iterator insert(const iterator& __p, _CharT __c)
1875  { insert(__p.index(), __c); return __p; }
1876  iterator insert(const iterator& __p )
1877  { insert(__p.index()); return __p; }
1878  iterator insert(const iterator& __p, const _CharT* c_string)
1879  { insert(__p.index(), c_string); return __p; }
1880  iterator insert(const iterator& __p, const _CharT* __i, size_t __n)
1881  { insert(__p.index(), __i, __n); return __p; }
1882  iterator insert(const iterator& __p, const _CharT* __i,
1883                  const _CharT* __j)
1884  { insert(__p.index(), __i, __j);  return __p; }
1885  iterator insert(const iterator& __p,
1886                  const const_iterator& __i, const const_iterator& __j)
1887  { insert(__p.index(), __i, __j); return __p; }
1888  iterator insert(const iterator& __p,
1889                  const iterator& __i, const iterator& __j)
1890  { insert(__p.index(), __i, __j); return __p; }
1891
1892  // Replace, range variants.
1893  void replace(const iterator& __p, const iterator& __q,
1894               const _Self& __r)
1895  { replace(__p.index(), __q.index() - __p.index(), __r); }
1896  void replace(const iterator& __p, const iterator& __q, _CharT __c)
1897  { replace(__p.index(), __q.index() - __p.index(), __c); }
1898  void replace(const iterator& __p, const iterator& __q,
1899               const _CharT* __c_string)
1900  { replace(__p.index(), __q.index() - __p.index(), __c_string); }
1901  void replace(const iterator& __p, const iterator& __q,
1902               const _CharT* __i, size_t __n)
1903  { replace(__p.index(), __q.index() - __p.index(), __i, __n); }
1904  void replace(const iterator& __p, const iterator& __q,
1905               const _CharT* __i, const _CharT* __j)
1906  { replace(__p.index(), __q.index() - __p.index(), __i, __j); }
1907  void replace(const iterator& __p, const iterator& __q,
1908               const const_iterator& __i, const const_iterator& __j)
1909  { replace(__p.index(), __q.index() - __p.index(), __i, __j); }
1910  void replace(const iterator& __p, const iterator& __q,
1911               const iterator& __i, const iterator& __j)
1912  { replace(__p.index(), __q.index() - __p.index(), __i, __j); }
1913
1914  // Replace, iterator variants.
1915  void replace(const iterator& __p, const _Self& __r)
1916  { replace(__p.index(), __r); }
1917  void replace(const iterator& __p, _CharT __c)
1918  { replace(__p.index(), __c); }
1919  void replace(const iterator& __p, const _CharT* __c_string)
1920  { replace(__p.index(), __c_string); }
1921  void replace(const iterator& __p, const _CharT* __i, size_t __n)
1922  { replace(__p.index(), __i, __n); }
1923  void replace(const iterator& __p, const _CharT* __i, const _CharT* __j)
1924  { replace(__p.index(), __i, __j); }
1925  void replace(const iterator& __p, const_iterator __i,
1926               const_iterator __j)
1927  { replace(__p.index(), __i, __j); }
1928  void replace(const iterator& __p, iterator __i, iterator __j)
1929  { replace(__p.index(), __i, __j); }
1930
1931  // Iterator and range variants of erase
1932  iterator erase(const iterator& __p, const iterator& __q) {
1933    size_t __p_index = __p.index();
1934    erase(__p_index, __q.index() - __p_index);
1935    return iterator(this, __p_index);
1936  }
1937  iterator erase(const iterator& __p) {
1938    size_t __p_index = __p.index();
1939    erase(__p_index, 1);
1940    return iterator(this, __p_index);
1941  }
1942
1943  _Self substr(size_t __start, size_t __len = 1) const {
1944    if (__start > size()) _M_throw_out_of_range();
1945    return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __start, __start + __len));
1946  }
1947
1948  _Self substr(iterator __start, iterator __end) const {
1949    return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __start.index(), __end.index()));
1950  }
1951
1952  _Self substr(iterator __start) const {
1953    size_t __pos = __start.index();
1954    return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __pos, __pos + 1));
1955  }
1956
1957  _Self substr(const_iterator __start, const_iterator __end) const {
1958    // This might eventually take advantage of the cache in the
1959    // iterator.
1960    return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __start.index(), __end.index()));
1961  }
1962
1963  rope<_CharT,_Alloc> substr(const_iterator __start) {
1964    size_t __pos = __start.index();
1965    return rope<_CharT,_Alloc>(_S_substring(_M_tree_ptr._M_data, __pos, __pos + 1));
1966  }
1967
1968#include <stl/_string_npos.h>
1969
1970  size_type find(const _Self& __s, size_type __pos = 0) const {
1971    if (__pos >= size())
1972# ifndef _STLP_OLD_ROPE_SEMANTICS
1973      return npos;
1974# else
1975      return size();
1976# endif
1977
1978    size_type __result_pos;
1979    const_iterator __result = _STLP_STD::search(const_begin() + (ptrdiff_t)__pos, const_end(), __s.begin(), __s.end() );
1980    __result_pos = __result.index();
1981# ifndef _STLP_OLD_ROPE_SEMANTICS
1982    if (__result_pos == size()) __result_pos = npos;
1983# endif
1984    return __result_pos;
1985  }
1986  size_type find(_CharT __c, size_type __pos = 0) const;
1987  size_type find(const _CharT* __s, size_type __pos = 0) const {
1988    size_type __result_pos;
1989    const_iterator __result = _STLP_STD::search(const_begin() + (ptrdiff_t)__pos, const_end(),
1990                                                __s, __s + _S_char_ptr_len(__s));
1991    __result_pos = __result.index();
1992# ifndef _STLP_OLD_ROPE_SEMANTICS
1993    if (__result_pos == size()) __result_pos = npos;
1994# endif
1995    return __result_pos;
1996  }
1997
1998  iterator mutable_begin() {
1999    return(iterator(this, 0));
2000  }
2001
2002  iterator mutable_end() {
2003    return(iterator(this, size()));
2004  }
2005
2006  reverse_iterator mutable_rbegin() {
2007    return reverse_iterator(mutable_end());
2008  }
2009
2010  reverse_iterator mutable_rend() {
2011    return reverse_iterator(mutable_begin());
2012  }
2013
2014  reference mutable_reference_at(size_type __pos) {
2015    return reference(this, __pos);
2016  }
2017
2018# ifdef __STD_STUFF
2019  reference operator[] (size_type __pos) {
2020    return reference(this, __pos);
2021  }
2022
2023  reference at(size_type __pos) {
2024    if (__pos >= size()) _M_throw_out_of_range();
2025    return (*this)[__pos];
2026  }
2027
2028  void resize(size_type, _CharT) {}
2029  void resize(size_type) {}
2030  void reserve(size_type = 0) {}
2031  size_type capacity() const {
2032    return max_size();
2033  }
2034
2035  // Stuff below this line is dangerous because it's error prone.
2036  // I would really like to get rid of it.
2037  // copy function with funny arg ordering.
2038  size_type copy(_CharT* __buffer, size_type __n,
2039                 size_type __pos = 0) const {
2040    return copy(__pos, __n, __buffer);
2041  }
2042
2043  iterator end() { return mutable_end(); }
2044
2045  iterator begin() { return mutable_begin(); }
2046
2047  reverse_iterator rend() { return mutable_rend(); }
2048
2049  reverse_iterator rbegin() { return mutable_rbegin(); }
2050
2051# else
2052
2053  const_iterator end() { return const_end(); }
2054
2055  const_iterator begin() { return const_begin(); }
2056
2057  const_reverse_iterator rend() { return const_rend(); }
2058
2059  const_reverse_iterator rbegin() { return const_rbegin(); }
2060
2061# endif
2062}; //class rope
2063
2064#if defined (__GNUC__) && (__GNUC__ == 2) && (__GNUC_MINOR__ == 96)
2065template <class _CharT, class _Alloc>
2066const size_t rope<_CharT, _Alloc>::npos = ~(size_t) 0;
2067#endif
2068
2069template <class _CharT, class _Alloc>
2070inline _CharT
2071_Rope_const_iterator< _CharT, _Alloc>::operator[](size_t __n)
2072{ return rope<_CharT,_Alloc>::_S_fetch(this->_M_root, this->_M_current_pos + __n); }
2073
2074template <class _CharT, class _Alloc>
2075inline bool operator== (const _Rope_const_iterator<_CharT,_Alloc>& __x,
2076                        const _Rope_const_iterator<_CharT,_Alloc>& __y) {
2077  return (__x._M_current_pos == __y._M_current_pos &&
2078          __x._M_root == __y._M_root);
2079}
2080
2081template <class _CharT, class _Alloc>
2082inline bool operator< (const _Rope_const_iterator<_CharT,_Alloc>& __x,
2083                       const _Rope_const_iterator<_CharT,_Alloc>& __y)
2084{ return (__x._M_current_pos < __y._M_current_pos); }
2085
2086#ifdef _STLP_USE_SEPARATE_RELOPS_NAMESPACE
2087
2088template <class _CharT, class _Alloc>
2089inline bool operator!= (const _Rope_const_iterator<_CharT,_Alloc>& __x,
2090                        const _Rope_const_iterator<_CharT,_Alloc>& __y)
2091{ return !(__x == __y); }
2092
2093template <class _CharT, class _Alloc>
2094inline bool operator> (const _Rope_const_iterator<_CharT,_Alloc>& __x,
2095                       const _Rope_const_iterator<_CharT,_Alloc>& __y)
2096{ return __y < __x; }
2097
2098template <class _CharT, class _Alloc>
2099inline bool operator<= (const _Rope_const_iterator<_CharT,_Alloc>& __x,
2100                        const _Rope_const_iterator<_CharT,_Alloc>& __y)
2101{ return !(__y < __x); }
2102
2103template <class _CharT, class _Alloc>
2104inline bool operator>= (const _Rope_const_iterator<_CharT,_Alloc>& __x,
2105                        const _Rope_const_iterator<_CharT,_Alloc>& __y)
2106{ return !(__x < __y); }
2107
2108#endif /* _STLP_USE_SEPARATE_RELOPS_NAMESPACE */
2109
2110template <class _CharT, class _Alloc>
2111inline ptrdiff_t operator-(const _Rope_const_iterator<_CharT,_Alloc>& __x,
2112                           const _Rope_const_iterator<_CharT,_Alloc>& __y)
2113{ return (ptrdiff_t)__x._M_current_pos - (ptrdiff_t)__y._M_current_pos; }
2114
2115#if !defined( __MWERKS__ ) || __MWERKS__ >= 0x2000  // dwa 8/21/97  - "ambiguous access to overloaded function" bug.
2116template <class _CharT, class _Alloc>
2117inline _Rope_const_iterator<_CharT,_Alloc>
2118operator-(const _Rope_const_iterator<_CharT,_Alloc>& __x, ptrdiff_t __n)
2119{ return _Rope_const_iterator<_CharT,_Alloc>(__x._M_root, __x._M_current_pos - __n); }
2120# endif
2121
2122template <class _CharT, class _Alloc>
2123inline _Rope_const_iterator<_CharT,_Alloc>
2124operator+(const _Rope_const_iterator<_CharT,_Alloc>& __x, ptrdiff_t __n)
2125{ return _Rope_const_iterator<_CharT,_Alloc>(__x._M_root, __x._M_current_pos + __n); }
2126
2127template <class _CharT, class _Alloc>
2128inline _Rope_const_iterator<_CharT,_Alloc>
2129operator+(ptrdiff_t __n, const _Rope_const_iterator<_CharT,_Alloc>& __x)
2130{ return _Rope_const_iterator<_CharT,_Alloc>(__x._M_root, __x._M_current_pos + __n); }
2131
2132template <class _CharT, class _Alloc>
2133inline bool operator== (const _Rope_iterator<_CharT,_Alloc>& __x,
2134                        const _Rope_iterator<_CharT,_Alloc>& __y) {
2135  return (__x._M_current_pos == __y._M_current_pos &&
2136          __x._M_root_rope == __y._M_root_rope);
2137}
2138
2139template <class _CharT, class _Alloc>
2140inline bool operator< (const _Rope_iterator<_CharT,_Alloc>& __x,
2141                       const _Rope_iterator<_CharT,_Alloc>& __y)
2142{ return (__x._M_current_pos < __y._M_current_pos); }
2143
2144#if defined (_STLP_USE_SEPARATE_RELOPS_NAMESPACE)
2145template <class _CharT, class _Alloc>
2146inline bool operator!= (const _Rope_iterator<_CharT,_Alloc>& __x,
2147                        const _Rope_iterator<_CharT,_Alloc>& __y)
2148{ return !(__x == __y); }
2149
2150template <class _CharT, class _Alloc>
2151inline bool operator> (const _Rope_iterator<_CharT,_Alloc>& __x,
2152                       const _Rope_iterator<_CharT,_Alloc>& __y)
2153{ return __y < __x; }
2154
2155template <class _CharT, class _Alloc>
2156inline bool operator<= (const _Rope_iterator<_CharT,_Alloc>& __x,
2157                        const _Rope_iterator<_CharT,_Alloc>& __y)
2158{ return !(__y < __x); }
2159
2160template <class _CharT, class _Alloc>
2161inline bool operator>= (const _Rope_iterator<_CharT,_Alloc>& __x,
2162                        const _Rope_iterator<_CharT,_Alloc>& __y)
2163{ return !(__x < __y); }
2164#endif /* _STLP_USE_SEPARATE_RELOPS_NAMESPACE */
2165
2166template <class _CharT, class _Alloc>
2167inline ptrdiff_t operator-(const _Rope_iterator<_CharT,_Alloc>& __x,
2168                           const _Rope_iterator<_CharT,_Alloc>& __y)
2169{ return (ptrdiff_t)__x._M_current_pos - (ptrdiff_t)__y._M_current_pos; }
2170
2171#if !defined( __MWERKS__ ) || __MWERKS__ >= 0x2000  // dwa 8/21/97  - "ambiguous access to overloaded function" bug.
2172template <class _CharT, class _Alloc>
2173inline _Rope_iterator<_CharT,_Alloc>
2174operator-(const _Rope_iterator<_CharT,_Alloc>& __x,
2175          ptrdiff_t __n) {
2176  return _Rope_iterator<_CharT,_Alloc>(__x._M_root_rope, __x._M_current_pos - __n);
2177}
2178# endif
2179
2180template <class _CharT, class _Alloc>
2181inline _Rope_iterator<_CharT,_Alloc>
2182operator+(const _Rope_iterator<_CharT,_Alloc>& __x,
2183          ptrdiff_t __n) {
2184  return _Rope_iterator<_CharT,_Alloc>(__x._M_root_rope, __x._M_current_pos + __n);
2185}
2186
2187template <class _CharT, class _Alloc>
2188inline _Rope_iterator<_CharT,_Alloc>
2189operator+(ptrdiff_t __n, const _Rope_iterator<_CharT,_Alloc>& __x) {
2190  return _Rope_iterator<_CharT,_Alloc>(__x._M_root_rope, __x._M_current_pos + __n);
2191}
2192
2193template <class _CharT, class _Alloc>
2194inline rope<_CharT,_Alloc>
2195operator+ (const rope<_CharT,_Alloc>& __left,
2196           const rope<_CharT,_Alloc>& __right) {
2197  _STLP_ASSERT(__left.get_allocator() == __right.get_allocator())
2198  return rope<_CharT,_Alloc>(rope<_CharT,_Alloc>::_S_concat_rep(__left._M_tree_ptr._M_data, __right._M_tree_ptr._M_data));
2199  // Inlining this should make it possible to keep __left and __right in registers.
2200}
2201
2202template <class _CharT, class _Alloc>
2203inline rope<_CharT,_Alloc>&
2204operator+= (rope<_CharT,_Alloc>& __left,
2205            const rope<_CharT,_Alloc>& __right) {
2206  __left.append(__right);
2207  return __left;
2208}
2209
2210template <class _CharT, class _Alloc>
2211inline rope<_CharT,_Alloc>
2212operator+ (const rope<_CharT,_Alloc>& __left,
2213           const _CharT* __right) {
2214  size_t __rlen = rope<_CharT,_Alloc>::_S_char_ptr_len(__right);
2215  return rope<_CharT,_Alloc>(rope<_CharT,_Alloc>::_S_concat_char_iter(__left._M_tree_ptr._M_data, __right, __rlen));
2216}
2217
2218template <class _CharT, class _Alloc>
2219inline rope<_CharT,_Alloc>&
2220operator+= (rope<_CharT,_Alloc>& __left,
2221            const _CharT* __right) {
2222  __left.append(__right);
2223  return __left;
2224}
2225
2226template <class _CharT, class _Alloc>
2227inline rope<_CharT,_Alloc>
2228operator+ (const rope<_CharT,_Alloc>& __left, _CharT __right) {
2229  return rope<_CharT,_Alloc>(rope<_CharT,_Alloc>::_S_concat_char_iter(__left._M_tree_ptr._M_data, &__right, 1));
2230}
2231
2232template <class _CharT, class _Alloc>
2233inline rope<_CharT,_Alloc>&
2234operator+= (rope<_CharT,_Alloc>& __left, _CharT __right) {
2235  __left.append(__right);
2236  return __left;
2237}
2238
2239template <class _CharT, class _Alloc>
2240inline bool
2241operator< (const rope<_CharT,_Alloc>& __left,
2242           const rope<_CharT,_Alloc>& __right) {
2243  return __left.compare(__right) < 0;
2244}
2245
2246template <class _CharT, class _Alloc>
2247inline bool
2248operator== (const rope<_CharT,_Alloc>& __left,
2249            const rope<_CharT,_Alloc>& __right) {
2250  return __left.compare(__right) == 0;
2251}
2252
2253#ifdef _STLP_USE_SEPARATE_RELOPS_NAMESPACE
2254
2255template <class _CharT, class _Alloc>
2256inline bool
2257operator!= (const rope<_CharT,_Alloc>& __x, const rope<_CharT,_Alloc>& __y) {
2258  return !(__x == __y);
2259}
2260
2261template <class _CharT, class _Alloc>
2262inline bool
2263operator> (const rope<_CharT,_Alloc>& __x, const rope<_CharT,_Alloc>& __y) {
2264  return __y < __x;
2265}
2266
2267template <class _CharT, class _Alloc>
2268inline bool
2269operator<= (const rope<_CharT,_Alloc>& __x, const rope<_CharT,_Alloc>& __y) {
2270  return !(__y < __x);
2271}
2272
2273template <class _CharT, class _Alloc>
2274inline bool
2275operator>= (const rope<_CharT,_Alloc>& __x, const rope<_CharT,_Alloc>& __y) {
2276  return !(__x < __y);
2277}
2278
2279template <class _CharT, class _Alloc>
2280inline bool operator!= (const _Rope_char_ptr_proxy<_CharT,_Alloc>& __x,
2281                        const _Rope_char_ptr_proxy<_CharT,_Alloc>& __y) {
2282  return !(__x == __y);
2283}
2284
2285#endif /* _STLP_USE_SEPARATE_RELOPS_NAMESPACE */
2286
2287template <class _CharT, class _Alloc>
2288inline bool operator== (const _Rope_char_ptr_proxy<_CharT,_Alloc>& __x,
2289                        const _Rope_char_ptr_proxy<_CharT,_Alloc>& __y) {
2290  return (__x._M_pos == __y._M_pos && __x._M_root == __y._M_root);
2291}
2292
2293#if !defined (_STLP_USE_NO_IOSTREAMS)
2294template<class _CharT, class _Traits, class _Alloc>
2295basic_ostream<_CharT, _Traits>& operator<< (basic_ostream<_CharT, _Traits>& __o,
2296                                            const rope<_CharT, _Alloc>& __r);
2297#endif
2298
2299typedef rope<char, allocator<char> > crope;
2300#if defined (_STLP_HAS_WCHAR_T)
2301typedef rope<wchar_t, allocator<wchar_t> > wrope;
2302#endif
2303
2304inline crope::reference __mutable_reference_at(crope& __c, size_t __i)
2305{ return __c.mutable_reference_at(__i); }
2306
2307#if defined (_STLP_HAS_WCHAR_T)
2308inline wrope::reference __mutable_reference_at(wrope& __c, size_t __i)
2309{ return __c.mutable_reference_at(__i); }
2310#endif
2311
2312#if defined (_STLP_FUNCTION_TMPL_PARTIAL_ORDER)
2313template <class _CharT, class _Alloc>
2314inline void swap(rope<_CharT,_Alloc>& __x, rope<_CharT,_Alloc>& __y)
2315{ __x.swap(__y); }
2316#else
2317
2318inline void swap(crope& __x, crope& __y) { __x.swap(__y); }
2319# ifdef _STLP_HAS_WCHAR_T  // dwa 8/21/97
2320inline void swap(wrope& __x, wrope& __y) { __x.swap(__y); }
2321# endif
2322
2323#endif /* _STLP_FUNCTION_TMPL_PARTIAL_ORDER */
2324
2325
2326// Hash functions should probably be revisited later:
2327_STLP_TEMPLATE_NULL struct hash<crope> {
2328  size_t operator()(const crope& __str) const {
2329    size_t _p_size = __str.size();
2330
2331    if (0 == _p_size) return 0;
2332    return 13*__str[0] + 5*__str[_p_size - 1] + _p_size;
2333  }
2334};
2335
2336#if defined (_STLP_HAS_WCHAR_T)  // dwa 8/21/97
2337_STLP_TEMPLATE_NULL struct hash<wrope> {
2338  size_t operator()(const wrope& __str) const {
2339    size_t _p_size = __str.size();
2340
2341    if (0 == _p_size) return 0;
2342    return 13*__str[0] + 5*__str[_p_size - 1] + _p_size;
2343  }
2344};
2345#endif
2346
2347#if (!defined (_STLP_MSVC) || (_STLP_MSVC >= 1310))
2348// I couldn't get this to work with VC++
2349template<class _CharT,class _Alloc>
2350#  if defined (__DMC__)
2351extern
2352#  endif
2353void _Rope_rotate(_Rope_iterator<_CharT, _Alloc> __first,
2354                  _Rope_iterator<_CharT, _Alloc> __middle,
2355                  _Rope_iterator<_CharT, _Alloc> __last);
2356
2357inline void rotate(_Rope_iterator<char, allocator<char> > __first,
2358                   _Rope_iterator<char, allocator<char> > __middle,
2359                   _Rope_iterator<char, allocator<char> > __last)
2360{ _Rope_rotate(__first, __middle, __last); }
2361#endif
2362
2363template <class _CharT, class _Alloc>
2364inline _Rope_char_ref_proxy<_CharT, _Alloc>::operator _CharT () const {
2365  if (_M_current_valid) {
2366    return _M_current;
2367  } else {
2368    return _My_rope::_S_fetch(_M_root->_M_tree_ptr._M_data, _M_pos);
2369  }
2370}
2371
2372#if defined (_STLP_CLASS_PARTIAL_SPECIALIZATION) && !defined (_STLP_NO_MOVE_SEMANTIC)
2373template <class _CharT, class _Alloc>
2374struct __move_traits<rope<_CharT, _Alloc> > {
2375  typedef __true_type implemented;
2376  //Completness depends on the allocator:
2377  typedef typename __move_traits<_Alloc>::complete complete;
2378};
2379#endif
2380
2381_STLP_END_NAMESPACE
2382
2383#if !defined (_STLP_LINK_TIME_INSTANTIATION)
2384#  include <stl/_rope.c>
2385#endif
2386
2387#endif /* _STLP_INTERNAL_ROPE_H */
2388
2389// Local Variables:
2390// mode:C++
2391// End:
2392