NAME
Lingua::EN::Segment - split English-language domain names etc. into words
SYNOPSIS
my $segmenter = Lingua::EN::Segment->new;
for my $domain (<>) {
chomp $domain;
my @words = $segmenter->segment($domain);
print "$domain: ", join(', ', @words), "\n";
}
DESCRIPTION
Sometimes you have a string that to a human eye is clearly made up of many words glommed together without spaces or hyphens. This module uses some mild cunning and a large list of known words from Google to try and work out how the string should be split into words.
new
Out: $segmenter
Returns a Lingua::EN::Segment object.
dist_dir
Out: $dist_dir
Returns the name of the directory where distribution-specific files are installed.
segment
In: $unsegmented_string
Out: @words
Supplied with an unsegmented string - e.g. a domain name - returns a list of words that are most statistically likely to be the words that make up this string.
unigrams
Out: \%unigrams
Returns a hashref of word => likelihood to appear in Google's huge list of words that they got off the Internet. The higher the likelihood, the more likely that this is a genuine regularly-used word, rather than an obscure word or a typo.
bigrams
Out: \%bigrams
As unigrams, but returns a lookup table of "word1 word2" => likelihood for combinations of words.
ACKNOWLEDGEMENTS
This code is based on chapter 14 of Peter Norvig's book Beautiful Data.