Server IP : 103.119.228.120 / Your IP : 3.145.91.152 Web Server : Apache System : Linux v8.techscape8.com 3.10.0-1160.119.1.el7.tuxcare.els2.x86_64 #1 SMP Mon Jul 15 12:09:18 UTC 2024 x86_64 User : nobody ( 99) PHP Version : 5.6.40 Disable Function : shell_exec,symlink,system,exec,proc_get_status,proc_nice,proc_terminate,define_syslog_variables,syslog,openlog,closelog,escapeshellcmd,passthru,ocinum cols,ini_alter,leak,listen,chgrp,apache_note,apache_setenv,debugger_on,debugger_off,ftp_exec,dl,dll,myshellexec,proc_open,socket_bind,proc_close,escapeshellarg,parse_ini_filepopen,fpassthru,exec,passthru,escapeshellarg,escapeshellcmd,proc_close,proc_open,ini_alter,popen,show_source,proc_nice,proc_terminate,proc_get_status,proc_close,pfsockopen,leak,apache_child_terminate,posix_kill,posix_mkfifo,posix_setpgid,posix_setsid,posix_setuid,dl,symlink,shell_exec,system,dl,passthru,escapeshellarg,escapeshellcmd,myshellexec,c99_buff_prepare,c99_sess_put,fpassthru,getdisfunc,fx29exec,fx29exec2,is_windows,disp_freespace,fx29sh_getupdate,fx29_buff_prepare,fx29_sess_put,fx29shexit,fx29fsearch,fx29ftpbrutecheck,fx29sh_tools,fx29sh_about,milw0rm,imagez,sh_name,myshellexec,checkproxyhost,dosyayicek,c99_buff_prepare,c99_sess_put,c99getsource,c99sh_getupdate,c99fsearch,c99shexit,view_perms,posix_getpwuid,posix_getgrgid,posix_kill,parse_perms,parsesort,view_perms_color,set_encoder_input,ls_setcheckboxall,ls_reverse_all,rsg_read,rsg_glob,selfURL,dispsecinfo,unix2DosTime,addFile,system,get_users,view_size,DirFiles,DirFilesWide,DirPrintHTMLHeaders,GetFilesTotal,GetTitles,GetTimeTotal,GetMatchesCount,GetFileMatchesCount,GetResultFiles,fs_copy_dir,fs_copy_obj,fs_move_dir,fs_move_obj,fs_rmdir,SearchText,getmicrotime MySQL : ON | cURL : ON | WGET : ON | Perl : ON | Python : ON | Sudo : ON | Pkexec : ON Directory : /lib/mysqlsh/lib/python3.9/site-packages/antlr4/ |
Upload File : |
# Copyright (c) 2012-2017 The ANTLR Project. All rights reserved. # Use of this file is governed by the BSD 3-clause license that # can be found in the LICENSE.txt file in the project root. # # A token has properties: text, type, line, character position in the line # (so we can ignore tabs), token channel, index, and source from which # we obtained this token. from io import StringIO class Token (object): __slots__ = ('source', 'type', 'channel', 'start', 'stop', 'tokenIndex', 'line', 'column', '_text') INVALID_TYPE = 0 # During lookahead operations, this "token" signifies we hit rule end ATN state # and did not follow it despite needing to. EPSILON = -2 MIN_USER_TOKEN_TYPE = 1 EOF = -1 # All tokens go to the parser (unless skip() is called in that rule) # on a particular "channel". The parser tunes to a particular channel # so that whitespace etc... can go to the parser on a "hidden" channel. DEFAULT_CHANNEL = 0 # Anything on different channel than DEFAULT_CHANNEL is not parsed # by parser. HIDDEN_CHANNEL = 1 def __init__(self): self.source = None self.type = None # token type of the token self.channel = None # The parser ignores everything not on DEFAULT_CHANNEL self.start = None # optional; return -1 if not implemented. self.stop = None # optional; return -1 if not implemented. self.tokenIndex = None # from 0..n-1 of the token object in the input stream self.line = None # line=1..n of the 1st character self.column = None # beginning of the line at which it occurs, 0..n-1 self._text = None # text of the token. @property def text(self): return self._text # Explicitly set the text for this token. If {code text} is not # {@code null}, then {@link #getText} will return this value rather than # extracting the text from the input. # # @param text The explicit text of the token, or {@code null} if the text # should be obtained from the input along with the start and stop indexes # of the token. @text.setter def text(self, text:str): self._text = text def getTokenSource(self): return self.source[0] def getInputStream(self): return self.source[1] class CommonToken(Token): # An empty {@link Pair} which is used as the default value of # {@link #source} for tokens that do not have a source. EMPTY_SOURCE = (None, None) def __init__(self, source:tuple = EMPTY_SOURCE, type:int = None, channel:int=Token.DEFAULT_CHANNEL, start:int=-1, stop:int=-1): super().__init__() self.source = source self.type = type self.channel = channel self.start = start self.stop = stop self.tokenIndex = -1 if source[0] is not None: self.line = source[0].line self.column = source[0].column else: self.column = -1 # Constructs a new {@link CommonToken} as a copy of another {@link Token}. # # <p> # If {@code oldToken} is also a {@link CommonToken} instance, the newly # constructed token will share a reference to the {@link #text} field and # the {@link Pair} stored in {@link #source}. Otherwise, {@link #text} will # be assigned the result of calling {@link #getText}, and {@link #source} # will be constructed from the result of {@link Token#getTokenSource} and # {@link Token#getInputStream}.</p> # # @param oldToken The token to copy. # def clone(self): t = CommonToken(self.source, self.type, self.channel, self.start, self.stop) t.tokenIndex = self.tokenIndex t.line = self.line t.column = self.column t.text = self.text return t @property def text(self): if self._text is not None: return self._text input = self.getInputStream() if input is None: return None n = input.size if self.start < n and self.stop < n: return input.getText(self.start, self.stop) else: return "<EOF>" @text.setter def text(self, text:str): self._text = text def __str__(self): with StringIO() as buf: buf.write("[@") buf.write(str(self.tokenIndex)) buf.write(",") buf.write(str(self.start)) buf.write(":") buf.write(str(self.stop)) buf.write("='") txt = self.text if txt is not None: txt = txt.replace("\n","\\n") txt = txt.replace("\r","\\r") txt = txt.replace("\t","\\t") else: txt = "<no text>" buf.write(txt) buf.write("',<") buf.write(str(self.type)) buf.write(">") if self.channel > 0: buf.write(",channel=") buf.write(str(self.channel)) buf.write(",") buf.write(str(self.line)) buf.write(":") buf.write(str(self.column)) buf.write("]") return buf.getvalue()