Differences Between: [Versions 310 and 400] [Versions 310 and 401] [Versions 310 and 402] [Versions 310 and 403]
(no description)
File Size: | 166 lines (4 kb) |
Included or required: | 0 times |
Referenced: | 0 times |
Includes or requires: | 0 files |
TokenCountVectorizer:: (13 methods):
__construct()
fit()
transform()
getVocabulary()
buildVocabulary()
transformSample()
getTokenIndex()
addTokenToVocabulary()
isStopWord()
updateFrequency()
checkDocumentFrequency()
resetBeyondMinimum()
getBeyondMinimumIndexes()
Class: TokenCountVectorizer - X-Ref
__construct(Tokenizer $tokenizer, ?StopWords $stopWords = null, float $minDF = 0.0) X-Ref |
fit(array $samples, ?array $targets = null) X-Ref |
No description |
transform(array &$samples) X-Ref |
No description |
getVocabulary() X-Ref |
No description |
buildVocabulary(array &$samples) X-Ref |
No description |
transformSample(string &$sample) X-Ref |
No description |
getTokenIndex(string $token) X-Ref |
return: int|bool |
addTokenToVocabulary(string $token) X-Ref |
No description |
isStopWord(string $token) X-Ref |
No description |
updateFrequency(string $token) X-Ref |
No description |
checkDocumentFrequency(array &$samples) X-Ref |
No description |
resetBeyondMinimum(array &$sample, array $beyondMinimum) X-Ref |
No description |
getBeyondMinimumIndexes(int $samplesCount) X-Ref |
No description |