Searched refs:tokenize (Results 1 - 21 of 21) sorted by relevance

/freebsd-12-stable/sbin/devfs/
H A Dextern.h54 void tokenize(const char *, int *, char ***);
H A Ddevfs.c186 tokenize(const char *line, int *acp, char ***avp) function
H A Drule.c300 tokenize(str, &ac, &av);
/freebsd-12-stable/contrib/googletest/googlemock/scripts/generator/cpp/
H A Dast.py46 from cpp import tokenize namespace
551 if parts[-1].token_type == tokenize.NAME:
581 if (type_name and type_name[-1].token_type == tokenize.NAME and
582 p.token_type == tokenize.NAME):
583 type_name.append(tokenize.Token(tokenize.SYNTAX, ' ', 0, 0))
739 if token.token_type == tokenize.NAME:
750 if next.token_type == tokenize.SYNTAX and next.name == '(':
755 syntax = tokenize.SYNTAX
764 new_temp = self._GetTokensUpTo(tokenize
[all...]
/freebsd-12-stable/contrib/llvm-project/lld/ELF/
H A DScriptLexer.h26 void tokenize(MemoryBufferRef mb);
H A DScriptLexer.cpp73 ScriptLexer::ScriptLexer(MemoryBufferRef mb) { tokenize(mb); }
88 void ScriptLexer::tokenize(MemoryBufferRef mb) { function in class:lld::elf::ScriptLexer
H A DScriptParser.cpp351 tokenize(*mb);
/freebsd-12-stable/tools/tools/vt/keymaps/
H A Dconvert-keymap.pl90 sub tokenize { # split on white space and parentheses (but not within token) subroutine
118 my @KEYTOKEN = tokenize($_);
/freebsd-12-stable/contrib/llvm-project/lld/COFF/
H A DDriver.h50 llvm::opt::InputArgList parse(StringRef s) { return parse(tokenize(s)); }
62 std::vector<const char *> tokenize(StringRef s);
H A DDriverUtils.cpp868 for (StringRef tok : tokenize(s)) {
894 std::vector<const char *> v = tokenize(*s);
898 std::vector<const char *> v = tokenize(*s);
903 std::vector<const char *> ArgParser::tokenize(StringRef s) { function in class:lld::coff::ArgParser
/freebsd-12-stable/contrib/ntp/sntp/libopts/
H A DMakefile.am63 text_mmap.c time.c tokenize.c \
H A DMakefile.in479 tokenize.c usage.c version.c
/freebsd-12-stable/tools/tools/notescheck/
H A Dnotescheck.py218 def tokenize(line): function
229 print("Failed to tokenize: %s%s" (line, location), file=sys.stderr)
249 words = tokenize(line)
/freebsd-12-stable/crypto/openssl/
H A DConfigure1861 my @x = tokenize($2);
1870 my @x = tokenize($2);
1879 my @x = tokenize($2);
1888 my @x = tokenize($2);
1894 => sub { push @extra, tokenize($1)
1897 => sub { push @overrides, tokenize($1)
1901 => sub { push @{$ordinals{$1}}, tokenize($2)
1904 => sub { push @{$sources{$1}}, tokenize($2)
1907 => sub { push @{$shared_sources{$1}}, tokenize($2)
1910 => sub { push @{$includes{$1}}, tokenize(
[all...]
/freebsd-12-stable/contrib/llvm-project/clang/include/clang/Tooling/Syntax/
H A DTokens.h172 /// the spelled tokens of a file using the tokenize() helper.
335 std::vector<syntax::Token> tokenize(FileID FID, const SourceManager &SM,
/freebsd-12-stable/contrib/googletest/googlemock/
H A DMakefile.am157 scripts/generator/cpp/tokenize.py \
/freebsd-12-stable/contrib/llvm-project/clang/lib/Tooling/Syntax/
H A DTree.cpp46 auto It = ExtraTokens.try_emplace(FID, tokenize(FID, SourceMgr, LangOpts));
H A DTokens.cpp296 std::vector<syntax::Token> syntax::tokenize(FileID FID, const SourceManager &SM, function in class:syntax
580 File.SpelledTokens = tokenize(FID, SM, LangOpts);
/freebsd-12-stable/contrib/ntp/ntpdc/
H A Dntpdc.c83 static void tokenize (const char *, char **, int *);
1164 tokenize(cmdline, tokens, &ntok);
1270 * tokenize - turn a command line into tokens
1273 tokenize( function
/freebsd-12-stable/contrib/ntp/ntpq/
H A Dntpq.c197 static void tokenize (const char *, char **, int *);
1615 tokenize(cmdline, tokens, &ntok);
1705 * tokenize - turn a command line into tokens
1714 tokenize( function
/freebsd-12-stable/sys/dev/ocs_fc/
H A Docs_hw_queues.c1193 tokenize(const char *s, tok_t *tok) function
1616 ((s = tokenize(s, &tokarray.tokens[tokarray.inuse_count]))) != NULL; ) {

Completed in 214 milliseconds