elasticsearch - UTF8 encoding is longer than the max length 32766 -


i've upgraded elasticsearch cluster 1.1 1.2 , have errors when indexing big string.

{   "error": "illegalargumentexception[document contains @ least 1 immense term in field=\"response_body\" (whose utf8 encoding longer max length 32766), of skipped.  please correct analyzer not produce such terms.  prefix of first immense term is: '[7b 22 58 48 49 5f 48 6f 74 65 6c 41 76 61 69 6c 52 53 22 3a 7b 22 6d 73 67 56 65 72 73 69]...']",   "status": 500 } 

the mapping of index :

{   "template": "partner_requests-*",   "settings": {     "number_of_shards": 1,     "number_of_replicas": 1   },   "mappings": {     "request": {       "properties": {         "asn_id": { "index": "not_analyzed", "type": "string" },         "search_id": { "index": "not_analyzed", "type": "string" },         "partner": { "index": "not_analyzed", "type": "string" },         "start": { "type": "date" },         "duration": { "type": "float" },         "request_method": { "index": "not_analyzed", "type": "string" },         "request_url": { "index": "not_analyzed", "type": "string" },         "request_body": { "index": "not_analyzed", "type": "string" },         "response_status": { "type": "integer" },         "response_body": { "index": "not_analyzed", "type": "string" }       }     }   } } 

i've searched documentation , didn't find related maximum field size. according core types section don't understand why should "correct analyzer" not_analyzed field.

so running issue maximum size single term. when set field not_analyzed treat 1 single term. maximum size single term in underlying lucene index 32766 bytes, believe hard coded.

your 2 primary options either change type binary or continue use string set index type "no".


Comments

Popular posts from this blog

commonjs - How to write a typescript definition file for a node module that exports a function? -

openid - Okta: Failed to get authorization code through API call -

thorough guide for profiling racket code -