Open
Description
The links to the Wikipedia page on N-grams and the L-p vector normalization are not being interpreted correctly on the TextFeaturizingEstimator page:
This estimator gives the user one-stop solution for doing:
Language Detection
[Tokenization](https://en.wikipedia.org/wiki/Lexical_analysis#Tokenization)
[Text normalization](https://en.wikipedia.org/wiki/Text_normalization)
[Predefined and custom stopwords removal](https://en.wikipedia.org/wiki/Stop_words)
[Word-based or character-based Ngram extraction and SkipGram extraction (through the advanced [options](https://learn.microsoft.com/en-us/dotnet/api/microsoft.ml.transforms.text.textfeaturizingestimator?view=ml-dotnet))](https://en.wikipedia.org/wiki/N-gram)
[TF, IDF or TF-IDF](https://en.wikipedia.org/wiki/Tf%E2%80%93idf)
[L-p vector normalization](xref: Microsoft.ML.Transforms.LpNormNormalizingTransformer)
Document Details
⚠ Do not edit this section. It is required for learn.microsoft.com ➟ GitHub issue linking.
- ID: deeb31e0-6843-e9ae-e6ad-028c4ec9121b
- Version Independent ID: bf1e4b11-7652-7da4-4791-aa2d9488971c
- Content: TextFeaturizingEstimator Class (Microsoft.ML.Transforms.Text)
- Content Source: dotnet/xml/Microsoft.ML.Transforms.Text/TextFeaturizingEstimator.xml
- Product: dotnet-ml-api
- GitHub Login: @natke
- Microsoft Alias: nakersha
Metadata
Metadata
Assignees
Labels
No labels