Python | Tokenizing strings in a list of strings



Method # 1: Using a comprehension list + split()

We can solve this particular problem by using a list comprehension to traverse each line from a list of strings, and the split function does the tokenization task.

# Python3 demo code
# Tokenizing strings in a list of strings
# using a list comprehension + split ()

 
# initializing list

test_list = [ `Geeks for Geeks` , `is` , ` best computer science portal` ]

 

 
# print original list

print ( "The original list:" + str (test_list))

 
# using a list comprehension + split ()
# Tokenizing strings in a string list

res = [sub.split () for sub in test_list]

 
# print result

print ( " The list after split of strings is : " + str (res))

Output:

The original list: [` Geeks for Geeks`, `is`,` best computer science portal `]
The list after split of strings is: [[` Geeks`, `for`, `Geeks`], [` is`], [ `best`, `computer`, `science`, `portal`]]

Method # 2: Using map () + split ()
This is another method by which this particular task can be accomplished. In this method, we are just doing the same task as above, just using the map function to associate the splitting logic with the entire list.

# Python3 demo code
# Tokenizing strings in the string list
# using map () + split ()

 
# initializing list

test_list = [ `Geeks for Geeks` , `is` , ` best computer science portal` ]

 
# print original list

print ( "The original list:" + str (test_list))

  
# using the map () + split ()
# Tokenizing strings in the string list

res = list ( map ( str . split, test_list))

 
# print result

print ( "The list after split of strings is:" + str (res))

Output:

The original list: [`Geeks for Geeks`, `is`, `best computer science portal`]
The list after split of strings is: [[`Geeks`, `for` , `Geeks`], [`is`], [`best`, `computer`, `science`, `portal`]]