Initial release: December 2005
Kris
- class Iterator(T) : InputFilter ¶
-
The base class for a set of stream iterators. These operate
upon a buffered input stream, and are designed to deal with
partial content. That is, stream iterators go to work the
moment any data becomes available in the buffer. Contrast
this behaviour with the tango.text.Util iterators, which
operate upon the extent of an array.
There are two types of iterators supported; exclusive and
inclusive. The former are the more common kind, where a token
is delimited by elements that are considered foreign. Examples
include space, comma, and end-of-line delineation. Inclusive
tokens are just the opposite: they look for patterns in the
text that should be part of the token itself - everything else
is considered foreign. Currently tango.io.stream includes the
exclusive variety only.
Each pattern is exposed to the client as a slice of the original
content, where the slice is transient. If you need to retain the
exposed content, then you should .dup it appropriately.
The content provided to these iterators is intended to be fully
read-only. All current tokenizers abide by this rule, but it is
possible a user could mutate the content through a token slice.
To enforce the desired read-only aspect, the code would have to
introduce redundant copying or the compiler would have to support
read-only arrays (now in D2).
See Delimiters, Lines, Patterns, Quotes.
- size_t scan(const(void)[] data) [protected, abstract] ¶
-
The pattern scanner, implemented via subclasses.
- this(InputStream stream = null) ¶
-
Instantiate with a buffer.
- Iterator set(InputStream stream) ¶
-
Set the provided stream as the scanning source.
- const(T)[] get() [final] ¶
-
Return the current token as a slice of the content.
- int opApply(scope int delegate(ref const(T)[]) dg) ¶
-
Iterate over the set of tokens. This should really
provide read-only access to the tokens, but D does
not support that at this time.
- int opApply(scope int delegate(ref int, ref const(T)[]) dg) ¶
-
Iterate over a set of tokens, exposing a token count
starting at zero.
- int opApply(scope int delegate(ref int, ref const(T)[], ref const(T)[]) dg) ¶
-
Iterate over a set of tokens and delimiters, exposing a
token count starting at zero.
- const(T)[] next() [@property, final] ¶
-
Locate the next token. Returns the token if found, null
otherwise. Null indicates an end of stream condition. To
sweep a conduit for lines using method next():
1
2
3
|
auto lines = new Lines!(char) (new File("myfile"));
while (lines.next)
Cout (lines.get).newline;
|
Alternatively, we can extract one line from a conduit:
1
|
auto line = (new Lines!(char) (new File("myfile"))).next;
|
The difference between next() and foreach() is that the
latter processes all tokens in one go, whereas the former
processes in a piecemeal fashion. To wit:
1
2
|
foreach (line; new Lines!(char) (new File("myfile")))
Cout(line).newline;
|
- size_t set(const(T)* content, size_t start, size_t end) [protected, final] ¶
-
Set the content of the current slice to the provided
start and end points.
- size_t set(const(T)* content, size_t start, size_t end, size_t next) [protected, final] ¶
-
Set the content of the current slice to the provided
start and end points, and delimiter to the segment
between end & next (inclusive.)
- size_t notFound() [protected, final] ¶
-
Called when a scanner fails to find a matching pattern.
This may cause more content to be loaded, and a rescan
initiated.
- size_t found(size_t i) [protected, final] ¶
-
Invoked when a scanner matches a pattern. The provided
value should be the index of the last element of the
matching pattern, which is converted back to a void[]
index.
- bool has(const(T)[] set, T match) [protected, final] ¶
-
See if set of characters holds a particular instance.