aboutsummaryrefslogtreecommitdiffstats
path: root/lib/rdoc/token_stream.rb
diff options
context:
space:
mode:
authordrbrain <drbrain@b2dd03c8-39d4-4d8f-98ff-823fe69b080e>2010-12-20 03:22:49 +0000
committerdrbrain <drbrain@b2dd03c8-39d4-4d8f-98ff-823fe69b080e>2010-12-20 03:22:49 +0000
commit2ef9c50c6e405717d06362787c4549ca4f1c6485 (patch)
treeee99486567461dd5796f3d6edcc9e204187f2666 /lib/rdoc/token_stream.rb
parentd7effd506f5b91a636f2e6452ef1946b923007c7 (diff)
downloadruby-2ef9c50c6e405717d06362787c4549ca4f1c6485.tar.gz
Import RDoc 3
git-svn-id: svn+ssh://ci.ruby-lang.org/ruby/trunk@30249 b2dd03c8-39d4-4d8f-98ff-823fe69b080e
Diffstat (limited to 'lib/rdoc/token_stream.rb')
-rw-r--r--lib/rdoc/token_stream.rb50
1 files changed, 50 insertions, 0 deletions
diff --git a/lib/rdoc/token_stream.rb b/lib/rdoc/token_stream.rb
new file mode 100644
index 0000000000..fb887f2fa4
--- /dev/null
+++ b/lib/rdoc/token_stream.rb
@@ -0,0 +1,50 @@
+##
+# A TokenStream is a list of tokens, gathered during the parse of some entity
+# (say a method). Entities populate these streams by being registered with the
+# lexer. Any class can collect tokens by including TokenStream. From the
+# outside, you use such an object by calling the start_collecting_tokens
+# method, followed by calls to add_token and pop_token.
+
+module RDoc::TokenStream
+
+ ##
+ # Adds +tokens+ to the collected tokens
+
+ def add_tokens(*tokens)
+ tokens.flatten.each { |token| @token_stream << token }
+ end
+
+ alias add_token add_tokens
+
+ ##
+ # Starts collecting tokens
+
+ def collect_tokens
+ @token_stream = []
+ end
+
+ alias start_collecting_tokens collect_tokens
+
+ ##
+ # Remove the last token from the collected tokens
+
+ def pop_token
+ @token_stream.pop
+ end
+
+ ##
+ # Current token stream
+
+ def token_stream
+ @token_stream
+ end
+
+ ##
+ # Returns a string representation of the token stream
+
+ def tokens_to_s
+ token_stream.map { |token| token.text }.join ''
+ end
+
+end
+