elasticsearch - UTF8 encoding is longer than the max length 32766 -
i've upgraded elasticsearch cluster 1.1 1.2 , have errors when indexing big string.
{ "error": "illegalargumentexception[document contains @ least 1 immense term in field=\"response_body\" (whose utf8 encoding longer max length 32766), of skipped. please correct analyzer not produce such terms. prefix of first immense term is: '[7b 22 58 48 49 5f 48 6f 74 65 6c 41 76 61 69 6c 52 53 22 3a 7b 22 6d 73 67 56 65 72 73 69]...']", "status": 500 }
the mapping of index :
{ "template": "partner_requests-*", "settings": { "number_of_shards": 1, "number_of_replicas": 1 }, "mappings": { "request": { "properties": { "asn_id": { "index": "not_analyzed", "type": "string" }, "search_id": { "index": "not_analyzed", "type": "string" }, "partner": { "index": "not_analyzed", "type": "string" }, "start": { "type": "date" }, "duration": { "type": "float" }, "request_method": { "index": "not_analyzed", "type": "string" }, "request_url": { "index": "not_analyzed", "type": "string" }, "request_body": { "index": "not_analyzed", "type": "string" }, "response_status": { "type": "integer" }, "response_body": { "index": "not_analyzed", "type": "string" } } } } }
i've searched documentation , didn't find related maximum field size. according core types section don't understand why should "correct analyzer" not_analyzed
field.
so running issue maximum size single term. when set field not_analyzed treat 1 single term. maximum size single term in underlying lucene index 32766 bytes, believe hard coded.
your 2 primary options either change type binary or continue use string set index type "no".
Comments
Post a Comment