/freebsd-12-stable/sbin/devfs/ |
H A D | extern.h | 54 void tokenize(const char *, int *, char ***);
|
H A D | devfs.c | 186 tokenize(const char *line, int *acp, char ***avp) function
|
H A D | rule.c | 300 tokenize(str, &ac, &av);
|
/freebsd-12-stable/contrib/googletest/googlemock/scripts/generator/cpp/ |
H A D | ast.py | 46 from cpp import tokenize namespace 551 if parts[-1].token_type == tokenize.NAME: 581 if (type_name and type_name[-1].token_type == tokenize.NAME and 582 p.token_type == tokenize.NAME): 583 type_name.append(tokenize.Token(tokenize.SYNTAX, ' ', 0, 0)) 739 if token.token_type == tokenize.NAME: 750 if next.token_type == tokenize.SYNTAX and next.name == '(': 755 syntax = tokenize.SYNTAX 764 new_temp = self._GetTokensUpTo(tokenize [all...] |
/freebsd-12-stable/contrib/llvm-project/lld/ELF/ |
H A D | ScriptLexer.h | 26 void tokenize(MemoryBufferRef mb);
|
H A D | ScriptLexer.cpp | 73 ScriptLexer::ScriptLexer(MemoryBufferRef mb) { tokenize(mb); } 88 void ScriptLexer::tokenize(MemoryBufferRef mb) { function in class:lld::elf::ScriptLexer
|
H A D | ScriptParser.cpp | 351 tokenize(*mb);
|
/freebsd-12-stable/tools/tools/vt/keymaps/ |
H A D | convert-keymap.pl | 90 sub tokenize { # split on white space and parentheses (but not within token) subroutine 118 my @KEYTOKEN = tokenize($_);
|
/freebsd-12-stable/contrib/llvm-project/lld/COFF/ |
H A D | Driver.h | 50 llvm::opt::InputArgList parse(StringRef s) { return parse(tokenize(s)); } 62 std::vector<const char *> tokenize(StringRef s);
|
H A D | DriverUtils.cpp | 868 for (StringRef tok : tokenize(s)) { 894 std::vector<const char *> v = tokenize(*s); 898 std::vector<const char *> v = tokenize(*s); 903 std::vector<const char *> ArgParser::tokenize(StringRef s) { function in class:lld::coff::ArgParser
|
/freebsd-12-stable/contrib/ntp/sntp/libopts/ |
H A D | Makefile.am | 63 text_mmap.c time.c tokenize.c \
|
H A D | Makefile.in | 479 tokenize.c usage.c version.c
|
/freebsd-12-stable/tools/tools/notescheck/ |
H A D | notescheck.py | 218 def tokenize(line): function 229 print("Failed to tokenize: %s%s" (line, location), file=sys.stderr) 249 words = tokenize(line)
|
/freebsd-12-stable/crypto/openssl/ |
H A D | Configure | 1861 my @x = tokenize($2); 1870 my @x = tokenize($2); 1879 my @x = tokenize($2); 1888 my @x = tokenize($2); 1894 => sub { push @extra, tokenize($1) 1897 => sub { push @overrides, tokenize($1) 1901 => sub { push @{$ordinals{$1}}, tokenize($2) 1904 => sub { push @{$sources{$1}}, tokenize($2) 1907 => sub { push @{$shared_sources{$1}}, tokenize($2) 1910 => sub { push @{$includes{$1}}, tokenize( [all...] |
/freebsd-12-stable/contrib/llvm-project/clang/include/clang/Tooling/Syntax/ |
H A D | Tokens.h | 172 /// the spelled tokens of a file using the tokenize() helper. 335 std::vector<syntax::Token> tokenize(FileID FID, const SourceManager &SM,
|
/freebsd-12-stable/contrib/googletest/googlemock/ |
H A D | Makefile.am | 157 scripts/generator/cpp/tokenize.py \
|
/freebsd-12-stable/contrib/llvm-project/clang/lib/Tooling/Syntax/ |
H A D | Tree.cpp | 46 auto It = ExtraTokens.try_emplace(FID, tokenize(FID, SourceMgr, LangOpts));
|
H A D | Tokens.cpp | 296 std::vector<syntax::Token> syntax::tokenize(FileID FID, const SourceManager &SM, function in class:syntax 580 File.SpelledTokens = tokenize(FID, SM, LangOpts);
|
/freebsd-12-stable/contrib/ntp/ntpdc/ |
H A D | ntpdc.c | 83 static void tokenize (const char *, char **, int *); 1164 tokenize(cmdline, tokens, &ntok); 1270 * tokenize - turn a command line into tokens 1273 tokenize( function
|
/freebsd-12-stable/contrib/ntp/ntpq/ |
H A D | ntpq.c | 197 static void tokenize (const char *, char **, int *); 1615 tokenize(cmdline, tokens, &ntok); 1705 * tokenize - turn a command line into tokens 1714 tokenize( function
|
/freebsd-12-stable/sys/dev/ocs_fc/ |
H A D | ocs_hw_queues.c | 1193 tokenize(const char *s, tok_t *tok) function 1616 ((s = tokenize(s, &tokarray.tokens[tokarray.inuse_count]))) != NULL; ) {
|