Tokenization (Q2438971)

From Wikidata
Jump to: navigation, search
breaking a stream of text up into chunks for analysis or further processing
edit
Language Label Description Also known as
English
Tokenization
breaking a stream of text up into chunks for analysis or further processing

    Statements

     
    edit
      edit
        edit
          edit
            edit
              edit
                edit