Tanzu Greenplum text saves configuration files for an index in the ZooKeeper /gptext/configs/<index_name>
znode, for example /gptext/configs/demo.twitter.message
. The configuration files are copied from the $GPTXTHOME/share/gp_index_template/conf
directory and modified with information passed in the gptext.create_index()
function arguments and the VMware Greenplum table definition.
After an index has been created, you can modify the index's configuration files using the gptext-config
command-line utility. You can also edit the template files in the $GPTXTHOME/share/gp_index_template/conf
directory so that any new index you create has your customizations.
If you choose to customize the template files in the $GPTXTHOME/share/gp_index_template/conf
directory, you should first back up the files so that you can restore the default versions if necessary.
You can edit the index configuration files saved in ZooKeeper using the gptext-config
command-line utility with the edit
option. You provide the name of the index and the name of the configuration file you want to modify. To edit the managed-schema
file for the demo.twitter.message
index, for example:
$ gptext-config edit -i demo.twitter.message -f managed-schema
The utility loads the file into an editor, vi
by default. You can specify a different editor with the -e
option. This command uses the nano
editor to edit the stopwords.txt
file.
$ gptext-config edit -i demo.twitter.message -f stopwords.txt -e nano
Warning: When editing XML files such as managed-schema
, be sure that you save a valid XML document. Invalid XML syntax will cause Solr errors and prevent access to your index.
You can use the gptext-config upload
command to upload a local configuration file to ZooKeeper. This example uploads a local configuration file named protwords.custom
to ZooKeeper, overwriting the existing protwords.txt
file.
$ gptext-config upload -i demo.twitter.message -l protwords.custom -f protwords.txt
20171011:11:24:59:030178 gptext-config:gpdb:gpadmin-[INFO]:-Execute GPText config.
20171011:11:25:00:030178 gptext-config:gpdb:gpadmin-[INFO]:-Check zookeeper cluster state ...
20171011:11:25:00:030178 gptext-config:gpdb:gpadmin-[INFO]:-Upload file protwords.custom to zookeeper...
20171011:11:25:01:030178 gptext-config:gpdb:gpadmin-[INFO]:-Reloading configuration...
20171011:11:25:02:030178 gptext-config:gpdb:gpadmin-[INFO]:-Modifications to protwords.txt require that all data be reindexed.
20171011:11:25:02:030178 gptext-config:gpdb:gpadmin-[INFO]:-Done.
Use the gptext-config append
command to append a local text file to an existing configuration file. For example, you could create an additional list of stop words in a local file stopwords.add
and append them to the stopwords.txt
file.
$ gptext-config append -i demo.twitter.message -l stopwords.add -f stopwords.txt
20171010:09:52:59:019764 gptext-config:gpdb:gpadmin-[INFO]:-Execute GPText config.
20171010:09:53:00:019764 gptext-config:gpdb:gpadmin-[INFO]:-Check zookeeper cluster state ...
20171010:09:53:00:019764 gptext-config:gpdb:gpadmin-[INFO]:-Creating temporary copy of stopwords.txt...
20171010:09:53:01:019764 gptext-config:gpdb:gpadmin-[INFO]:-Appending contents of stopwords.add to stopwords.txt
20171010:09:53:01:019764 gptext-config:gpdb:gpadmin-[INFO]:-Backing up stopwords.txt for index demo.twitter.message...
20171010:09:53:03:019764 gptext-config:gpdb:gpadmin-[INFO]:-Reloading configuration...
20171010:09:53:22:019764 gptext-config:gpdb:gpadmin-[INFO]:-Modifications to stopwords.txt require that all data be reindexed.
20171010:09:53:22:019764 gptext-config:gpdb:gpadmin-[INFO]:-Done.
See the gptext-config
command reference for gptext-config
command-line options and for descriptions of the files you can edit with gptext-config
.
The main configuration file for an index is the managed-schema
file. The managed-schema
file is an XML file containing definitions for the fields, field types, and analyzer chains that define the contents and behavior of a Tanzu Greenplum text index.
<field>
XML element) maps a VMware Greenplum table column to a field in the Tanzu Greenplum text index.<fieldType>
XML element) assigns Solr Java classes and analyzer chains that handle a data type to a field.<analyzer>
XML element) is a container element that specifies the Java classes that tokenize and filter the content of a field that is to be indexed. An <analyzer>
element is a child of a <fieldType>
element.In addition to the managed-schema
file, the Solr configuration files for an index include text files that contain lists of words to treat specially when indexing data, localization files, character set collation maps used for sorting, and a Solr server configuration file.
The following sections provide an overview of the contents of the managed-schema
file and the relationships between the XML elements that define fields, field types, and analyzers. By editing the managed-schema
file, you can specify at the field level how Solr indexes and stores VMware Greenplum data.
For detailed documentation of the contents of the managed-schema
file, refer to the comments in the file or to the Apache SolrCloud documentation.
Tanzu Greenplum text adds field
elements to the managed-schema
file for columns included when the index was created with the gptext.create_index()
function. This example is the definition for a text field named description
:
<field name="description" stored="false" type="text_intl" indexed="true"/>
name
attribute is the name of the database column. If the column name is not a valid Solr field name, it is altered to conform.stored
attribute determines if the content of the field will be stored in the index. If the field is stored in the index, Tanzu Greenplum text search results can return the content of the field. If the attribute is not stored, retrieving the field content requires a SQL join.type
attribute maps the VMware Greenplum type to a Solr type, defined in the same file with a <fieldType>
element.indexed
attribute determines whether the field content will be indexed.The <field>
element can specify additional attributes for some types that Tanzu Greenplum text or that you add. For example, columns may include the storeOffsetsWithPositions
, termVectors
, termPositions
, and termOffsets
attributes that control highlighting behaviour.
See the comment after the <fields>
element for a complete list of attributes.
The type
attribute of the <field>
element is mapped to the name
attribute of a <fieldType>
element in the managed-schema
file. The <fieldType>
element determines how Solr parses and stores a field in the index.
The class
attribute maps the field type to a Solr Java class that recognizes and processes the data type. Solr includes many base field types. See Tanzu Greenplum text and Solr Data Type Mappings for a mapping of Solr types to VMware Greenplum types.
You can map a field to a different type by changing the field's type
attribute. For example, to use the Tanzu Greenplum text social media text analyzer chain, you can change the type of a text field from text_intl
to text_sm
. Both of text_intl
and text_sm
use the Solr.TextField
class, but specify different filters in their analyzer chains.
The Tanzu Greenplum text gptext.list_field_types()
function is a convenience function that lets you see the text field types defined in the managed-schema
file for an index without having to edit the file. All of the types listed have the class Solr.TextField
.
SELECT * FROM gptext.list_field_types('demo.wikipedia.articles');
list_field_types
---------------------------
ancestor_path
delimited_payloads_float
delimited_payloads_int
delimited_payloads_string
descendent_path
lowercase
phonetic_en
text
text_ar
text_bg
text_ca
text_cjk
text_cz
text_da
text_de
text_el
text_en
text_en_splitting
text_en_splitting_tight
text_es
text_eu
text_fa
text_fi
text_fr
text_ga
text_general
text_general_rev
text_gl
text_hi
text_hu
text_hy
text_icu
text_id
text_intl
text_intl_prev
text_it
text_ja
text_lv
text_nl
text_no
text_pt
text_ro
text_ru
text_sm
text_sv
text_th
text_tr
text_ws
text_zhsmart
(49 rows)
To add a custom type, you can add a new field type by implementing Solr Java type interfaces, or you can specify an existing base type and customize it with an analyzer chain, as described in the next section.
An analyzer examines the contents of field or search query phrase and returns a stream of tokens used to index the field or search the index. The <analyzer>
element is a child of a <fieldType>
element that specifies how text will be tokenized and processed before it is indexed or applied to a search. An <analyzer>
can be of type index
or query
.
Different indexing and query chains can be defined for indexing and querying operations by adding a type
attribute to the <analyzer>
element. If no type
attribute appears the chain is applied to both field text that is to be indexed and query text that searches the index.
Field analysis begins with a <tokenizer>
that divides the contents of a field into tokens. In Latin-based text documents, the tokens are words or terms. In Chinese, Japanese, and Korean (CJK) documents, the tokens are characters.
The tokenizer can be followed by one or more <filter>
elements which are applied in succession. Filters restrict the query results, for example, by removing unnecessary terms ("a", "an", "the"), converting term formats, or by performing other actions to ensure that only important, relevant terms appear in the result set. Each filter operates on the output of the tokenizer or filter that precedes it. Solr includes many tokenizers and filters that allow analyzer chains to process different character sets, languages, and transformations. See Analyzers, Tokenizers and Filters for more information.
Field types are assigned analyzers in an index's managed-schema
file. The following example shows the Solr text
field type specification:
<fieldType name="text" class="solr.TextField" positionIncrementGap="100" autoGeneratePhraseQueries="true">
<analyzer type="index">
<tokenizer class="solr.WhitespaceTokenizerFactory"/>
<!-- in this example, we will only use synonyms at query time
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="false"/>
-->
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="1" catenateNumbers="1" catenateAll="0" splitOnCaseChange="1"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.KeywordMarkerFilterFactory" protected="protwords.txt"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
<analyzer type="query">
<tokenizer class="solr.WhitespaceTokenizerFactory"/>
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="true"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt"/>
<filter class="solr.WordDelimiterFilterFactory" generateWordParts="1" generateNumberParts="1" catenateWords="0" catenateNumbers="0" catenateAll="0" splitOnCaseChange="1"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.KeywordMarkerFilterFactory" protected="protwords.txt"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldType>
An analyzer has only one tokenizer, solr.WhitespaceTokenizerFactory
in this example. The tokenizer can be followed by one or more filters executed in succession.
Filters restrict the query results. Each filter operates on the output of the tokenizer or filter that precedes it. For example, the solr.StopFilterFactory
filter removes unnecessary terms ("a", "an", "the") from the stream of tokens. The words to filter out of the stream are listed in the stopwords.txt
configuration file. You can edit the stopwords.txt
file with the gptext-config
utility to change the list of words excluded from the index.
There are separate analyzer types for index and query operations. The query analyzer chain in this example includes a solr.SynonymFilterFactory
that looks up each token in a file synonyms.txt
and, if found, returns the synonym in place of the token.
The analyzer chain can include a "stemmer", solr.PorterStemFilterFactory
in this example. The stemmer employs an algorithm to change words to their "stems". For example, "confidential", "confidentiality", and "confidentis" are all stemmed to "confidenti". Using a stemmer can dramatically reduce the size of the index, but users executing searches should be aware that some search expressions will not work as expected because of stemming. For example, searching with a wildcard such as "confidential*"
will return no matches because the words were stemmed to "confidenti" during indexing. Without a wildcard, the word in the search expression is also stemmed and therefore the search succeeds.
The gptext.get_field_type()
convenience function retrieves the field type definition for a field type from the managed-schema file for an index, as a JSON string. This example shows the field type definition for the Solr text
field type.
#= SELECT * FROM gptext.get_field_type('demo.wikipedia.articles', 'text');
field_type
-------------------------------------------------
{
"name": "text",
"class": "solr.TextField",
"indexAnalyzer": {
"tokenizer": {
"class": "solr.WhitespaceTokenizerFactory"
},
"filters": [
{
"class": "solr.StopFilterFactory",
"attributes": [
{
"name": "words",
"value": "stopwords.txt"
},
{
"name": "ignoreCase",
"value": "true"
}
]
},
{
"class": "solr.WordDelimiterFilterFactory",
"attributes": [
{
"name": "catenateNumbers",
"value": "1"
},
{
"name": "generateNumberParts",
"value": "1"
},
{
"name": "splitOnCaseChange",
"value": "1"
},
{
"name": "generateWordParts",
"value": "1"
},
{
"name": "catenateAll",
"value": "0"
},
{
"name": "catenateWords",
"value": "1"
}
]
},
{
"class": "solr.LowerCaseFilterFactory"
},
{
"class": "solr.KeywordMarkerFilterFactory",
"attributes": [
{
"name": "protected",
"value": "protwords.txt"
}
]
},
{
"class": "solr.PorterStemFilterFactory"
}
]
},
"queryAnalyzer": {
"tokenizer": {
"class": "solr.WhitespaceTokenizerFactory"
},
"filters": [
{
"class": "solr.SynonymFilterFactory",
"attributes": [
{
"name": "expand",
"value": "true"
},
{
"name": "ignoreCase",
"value": "true"
},
{
"name": "synonyms",
"value": "synonyms.txt"
}
]
},
{
"class": "solr.StopFilterFactory",
"attributes": [
{
"name": "words",
"value": "stopwords.txt"
},
{
"name": "ignoreCase",
"value": "true"
}
]
},
{
"class": "solr.WordDelimiterFilterFactory",
"attributes": [
{
"name": "catenateNumbers",
"value": "0"
},
{
"name": "generateNumberParts",
"value": "1"
},
{
"name": "splitOnCaseChange",
"value": "1"
},
{
"name": "generateWordParts",
"value": "1"
},
{
"name": "catenateAll",
"value": "0"
},
{
"name": "catenateWords",
"value": "0"
}
]
},
{
"class": "solr.LowerCaseFilterFactory"
},
{
"class": "solr.KeywordMarkerFilterFactory",
"attributes": [
{
"name": "protected",
"value": "protwords.txt"
}
]
},
{
"class": "solr.PorterStemFilterFactory"
}
]
},
"attributes": [
{
"name": "autoGeneratePhraseQueries",
"value": "true"
},
{
"name": "positionIncrementGap",
"value": "100"
}
]
}
(1 row)
The gptext.analyzer()
function lets you test an analyzer chain for a field without altering the index. It shows the output of the tokenizer and each filter in the chain. You supply the text to analyze and specify whether to test the index or the query analyzer chain. It is useful for testing tokenizers and filters and for troubleshooting search queries that do not return the expected results.
=# SELECT * FROM gptext.analyzer('demo.wikipedia.articles', 'index',
'If You Optimize Everything, You will Always be Unhappy.');
class | tokens
------------------------+-----------------------------------------------------------------------------------------------
WhitespaceTokenizer | {{"If"},{"You"},{"Optimize"},{"Everything,"},{"You"},{"will"},{"Always"},{"be"},{"Unhappy."}}
StopFilter | {{},{"You"},{"Optimize"},{"Everything,"},{"You"},{},{"Always"},{},{"Unhappy."}}
WordDelimiterFilter | {{},{"You"},{"Optimize"},{"Everything"},{"You"},{},{"Always"},{},{"Unhappy"}}
LowerCaseFilter | {{},{"you"},{"optimize"},{"everything"},{"you"},{},{"always"},{},{"unhappy"}}
SetKeywordMarkerFilter | {{},{"you"},{"optimize"},{"everything"},{"you"},{},{"always"},{},{"unhappy"}}
PorterStemFilter | {{},{"you"},{"optim"},{"everyth"},{"you"},{},{"alwai"},{},{"unhappi"}}
(6 rows)
In addition to the text analyzer chains Solr provides, Tanzu Greenplum text provides the following text analyzer chains:
text_intl
is the default Tanzu Greenplum text analyzer. It is a multiple language text analyzer for text
fields. It handles Latin-based words and Chinese, Japanese, and Korean (CJK) characters.
text_intl
processes documents as follows.
Note that CJK and non-CJK text are treated as separate tokens. Preserving the original Korean word increases the number of tokens in a document.
Following is the definition from the Solr managed-schema
template.
<fieldType autoGeneratePhraseQueries="true" class="solr.TextField"
name="text_intl" positionIncrementGap="100">
<analyzer type="index">
<tokenizer class="com.emc.solr.analysis.worldlexer.WorldLexerTokenizerFactory"/>
<filter class="solr.CJKWidthFilterFactory"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="com.emc.solr.analysis.worldlexer.WorldLexerBigramFilterFactory" han="true"
hiragana="true" katakana="true" hangul="true" />
<filter class="solr.StopFilterFactory" enablePositionIncrements="true"
ignoreCase="true" words="stopwords.txt"/>
<filter class="solr.KeywordMarkerFilterFactory"
protected="protwords.txt"/>
<filter class="solr.PorterStemFilterFactory"/> </analyzer>
<analyzer type="query">
<tokenizer class="com.emc.solr.analysis.worldlexer.WorldLexerTokenizerFactory"/>
<filter class="solr.CJKWidthFilterFactory"/>
<filter class="com.emc.solr.analysis.worldlexer.WorldLexerBigramFilterFactory" han="true"
hiragana="true" katakana="true" hangul="true" />
<filter class="solr.StopFilterFactory" enablePositionIncrements="true" ignoreCase="true"
words="stopwords.txt"/>
<filter class="solr.KeywordMarkerFilterFactory" protected="protwords.txt"/>
<filter class="solr.PorterStemFilterFactory"/>
</analyzer>
</fieldType>
Following are the analysis steps for text_intl
.
WorldLexerTokenizerFactory
. This tokenizer handles most modern languages. It separates CJK characters from other language text and identifies any currency tokens or symbols.solr.CJKWidthFilterFactory
filter normalizes the CJK characters based on character width.solr.LowerCaseFilterFactory
filter changes all letters to lower case.WorldLexerBigramFilterFactory
filter generates a bigram for any CJK characters, leaves any non-CJK characters intact, and preserves original Korean-language words. Set the han
, hiragana
, katakana
, and hangul
attributes to "true"
to generate bigrams for all supported CJK languages.solr.StopFilterFactory
removes common words, such as "a", "an", and "the", which are listed in the stopwords.txt
configuration file (see To configure an index). If there are no words in the stopwords.txt
file, no words are removed.solr.KeywordMarkerFilterFactory
marks the English words to protect from stemming, using the words listed in the protwords.txt
configuration file (see To configure an index). If protwords.txt
does not contain a list of words, all words in the document are stemmed.solr.PorterStemFilterFactory
, a fast stemmer for the English language.NoteThe
text_intl
analyzer chain for querying is the same as thetext
analyzer chain for indexing.
An analyzer chain, text
, is included in Tanzu Greenplum text's Solr managed-schema
and is based on Solr's default analyzer chain. Because its tokenizer splits on white space, text
cannot process CJK languages: white space is meaningless for CJK languages. Best practice is to use the text_intl
analyzer.
For information about using an analyzer chain other than the default, see Using the text_sm Social Media Analyzer.
The root-level tokenizer, WorldLexerTokenizerFactory
, tokenizes international languages, including CJK languages. WorldLexerTokenizerFactory
tokenizes languages based on their Unicode points and, for Latin-based languages, white space.
NoteUnicode is the encoding for all text in VMware Greenplum.
The following are sample input to, and output from, Tanzu Greenplum text. Each line in the output corresponds to a term.
English and CJK input:
English and CJK output:
Bulgarian input:
Bulgarian output:
Danish input:
Danish output:
The text_intl analyzer uses the following filters:
The CJKWidthFilterFactory
normalizes width differences in CJK characters. This filter normalizes all character widths to fullwidth.
The WorldLexerBigramFilterFactory
filter forms bigrams (pairs) of CJK terms that are generated from WorldLexerTokenizerFactory
. This filter does not modify non-CJK text.
WorldLexerBigramFilterFactory
accepts attributes that guide the creation of bigrams for CJK scripts. For example, if the input contains HANGUL script but the hangul
attribute is set to false,
this filter will not create bigrams for that script. To ensure that WorldLexerBigramFilterFactory
creates bigrams as required, set the CJK attributes han
, hiragana
, katakana
, and hangul
to true
.
The Tanzu Greenplum text text_sm
text analyzer analyzes text from sources such as social media feeds. text_sm
consists of a tokenizer and two filters. To configure the text_sm
text analyzer, use the gptext-config
utility to edit the managed-schema
file. See To use the text_sm Social Media Analyzer for details.
text_sm
normalizes emoticons: it replaces emoticons with text using the emoticons.txt
configuration file. For example, it replaces a happy face emoticon, :-)
, with the text "happy".
The following is the definition from the Solr managed-schema
template.
<fieldType autoGeneratePhraseQueries="true"
class="solr.TextField" name="text_sm"
positionIncrementGap="100" termVectors="true"
termPositions="true" termOffsets="true">
<analyzer type="index">
<tokenizer class =
"com.emc.solr.analysis.text_sm.twitter.TwitterTokenizerFactory"
delimiter="\t"
emoticons="emoticons.txt"/>
<!-- Case insensitive stop word removal.
Add enablePositionIncrements=true in both the index and query
analyzers to leave a 'gap' for more accurate phrase queries. -->
<filter class="solr.StopFilterFactory"
enablePositionIncrements="true" ignoreCase="true"
words="stopwords.txt"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.KeywordMarkerFilterFactory"
protected="protwords.txt"/>
<filter class =
"com.emc.solr.analysis.text_sm.twitter.EmoticonsClassifierFilterFactory"
delimiter="\t" emoticons="emoticons.txt"/>
<filter class =
"com.emc.solr.analysis.text_sm.twitter.TwitterStemFilterFactory"/>
<analyzer type="query">
<tokenizer class =
"com.emc.solr.analysis.text_sm.twitter.TwitterTokenizerFactory"
delimiter="\t"
emoticons="emoticons.txt"
/>
<filter class="solr.StopFilterFactory"
enablePositionIncrements="true" ignoreCase="true"
words="stopwords.txt"/>
<filter class="solr.LowerCaseFilterFactory"/>
<filter class="solr.KeywordMarkerFilterFactory"
protected="protwords.txt"/>
<filter class =
"com.emc.solr.analysis.text_sm.twitter.EmoticonsClassifierFilterFactory"
delimiter="\t"
emoticons="emoticons.txt"/>
<filter class =
"com.emc.solr.analysis.text_sm.twitter.TwitterStemFilterFactory"/>
</analyzer>
</fieldType>
The Twitter tokenizer extends the English language tokenizer, solr.WhitespaceTokenizerFactory,
to recognize the following elements as terms.
com.emc.solr.analysis.socialmedia.twitter.EmoticonsClassifierFilterFactory
classifies emoticons as happy
, sad
, or wink
. It is based on the emoticons.txt
file (one of the files you can edit with gptext-config
, and is intended for future use, such as in sentiment analysis.
com.emc.solr.analysis.socialmedia.twitter.TwitterStemFilterFactory
extends the solr.PorterStemFilterFactory
class to bypass stemming of the social media patterns recognized by the twitter.TwitterTokenizerFactory
.
This file contains lists of emoticons for "happy," "sad," and "wink." They are separated by a tab by default. You can change the separation to any character or string by changing the value of delimiter
in the social media analyzer chain. The following is a sample line from the text_sm
analyzer chain:
<filter class =
"com.emc.solr.analysis.text_sm.twitter.EmoticonsClassifierFilterFactory"
delimiter="\t" emoticons="emoticons.txt"/>
The Solr managed-schema
file created for an index specifies an analyzer to use to index each field. The default analyzer for text fields is text_intl
. To specify the text_sm
social media analyzer, you use the gptext-config
utility to modify the Solr managed-schema
for your index.
The steps are:
Create an index using gptext.create_index()
.
Use the gptext-config
utility to edit the managed-schema
file created for the index:
gptext-config edit -f managed-schema -i <index_name>
The managed-schema
file contains a <field>
element for each text field. For example:
<field name="message_text" stored="false" type="text_intl" indexed="true"/>
The type
attribute specifies the analyzer to use. text_intl
is the default analyzer.
Modify the <field>
element for each text field you want to use the Tanzu Greenplum text social media analyzer and change the type
attribute as follows:
<field name="text_search_col" indexed="true" stored="false" type="text_sm"/>
Save the managed-schema
file.
If you want to index a field using two different analyzer chains simultaneously, you can do this:
Create a new empty index. Then use the gptext-config
utility to add a new field to the index that is a copy of the field you are interested in, but with a different name and analyzer chain.
Let us assume that your index, as initially created, includes a field to index named mytext
. Also assume that this field will be indexed using the default international analyzer (text_intl
).
You want to add a new field to the index's managed-schema
that is a copy of mytext
and that will be indexed with a different analyzer (say the text_sm
analyzer). To do so, follow these steps:
Create an empty index with gptext.create_index()
.
Open the index's managed-schema
file for editing with gptext-config
.
Add a <field>
in the managed-schema
for a new field that will use a different analyzer chain. For example:
<field indexed="true" name="mytext2" stored="false" type="text_sm"/>
By defining the type of this new field to be text_sm
, it will be indexed using the social media analyzer rather than the default text_intl
.
Add a <copyField>
in managed-schema
to copy the original field to the new field. For example:
<copyField dest="mytext2" source="mytext"/>
Index and commit as you normally would.
The database column mytext
is now in the index twice with two different analyzer chains. One column is mytext
, which uses the default international analyzer chain, and the other is the newly created mytext2,
which uses the social media analyzer chain.
You can use different analyzers for individual fields by editing the managed-schema configuration file. For example, if one field contains English text and another contains Chinese language text, you can specify different analyzers for the two fields.
You have a table named email_tbl
with the following definition:
create table email_tbl (
id bigint,
english_content text,
chinese_content text,
timestamp date,
username text,
age int,
... ) # additional columns that are not indexed
id
, english_content
, chinese_content
, timestamp
, username
, and age
.english_content
, you want to use the English language analyzer called "text_en" for the text segmentation.chinese_content
, you want to use the international language analyzer named "text_intl".Here are steps to implement this example:
Create the Tanzu Greenplum text index for the table.
SELECT * FROM gptext.create_index('public', 'email_tbl', 'id', 'english_content');
Modify the analyzer for each column in managed-schema
.
$ gptext-config edit -i db.public.email_tbl -f managed-schema
Find the element for the english_content
field.
<field name="english_content" type="*" indexed="true" stored="true" />
Change the type
attribute to text_en
.
<field name="english_content" type="text_en" indexed="true" stored="true" />
Find the element for the chinese_content
field.
<field name="chinese_content" type="*" indexed="true" stored="true" />
Change the type
attribute to text_intl
.
<field name="chinese_content" type="text_intl" indexed="true" stored="true" />
Index the table.
SELECT * FROM gptext.index(TABLE(SELECT id, english_content, chinese_content, timestamp, username, age FROM email_tbl),
'db.public.email_tbl');
Commit the index.
SELECT * FROM gptext.commit_index('db.public.email_tbl');
The field types text_en
and text_intl
are defined in <fieldType>
entries in the managed-schema file and then referenced in the type
attribute of the <field>
element.
You can define a custom field type by adding a <fieldType>
entry with custom analyzers and then setting the field's type
attribute to the name of the custom field type. For example, the following "text_customize" field type is a copy of the "text_en" field type entry with the synonym filter commented out in the index analyzer. This custom field type will apply the synonym filter to queries, but not to the index.
<fieldType name="text_customize" class="solr.TextField" positionIncrementGap="100">
<analyzer type="index">
<tokenizer class="solr.StandardTokenizerFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt" />
<!-- in this example, we will only use synonyms at query time
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="false"/>
-->
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
<analyzer type="query">
<tokenizer class="solr.StandardTokenizerFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt" />
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="true"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</fieldType>
A field type can also be customized by adding analyzers as child elements of the <field>
element:
<field name="english_content" type="text" indexed="true" stored="false">
<analyzer type="index">
<tokenizer class="solr.StandardTokenizerFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt" />
<!-- in this example, we will only use synonyms at query time
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="false"/>
-->
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
<analyzer type="query">
<tokenizer class="solr.StandardTokenizerFactory"/>
<filter class="solr.StopFilterFactory" ignoreCase="true" words="stopwords.txt" />
<filter class="solr.SynonymFilterFactory" synonyms="synonyms.txt" ignoreCase="true" expand="true"/>
<filter class="solr.LowerCaseFilterFactory"/>
</analyzer>
</field>