Elasticsearch connector#
The Elasticsearch Connector allows access to Elasticsearch data from Trino. This document describes how to setup the Elasticsearch Connector to run SQL queries against Elasticsearch.
Note
Elasticsearch (6.6.0 or later) or OpenSearch (1.1.0 or later) is required.
Configuration#
To configure the Elasticsearch connector, create a catalog properties file
etc/catalog/example.properties
with the following contents, replacing the
properties as appropriate for your setup:
connector.name=elasticsearch
elasticsearch.host=localhost
elasticsearch.port=9200
elasticsearch.default-schema-name=default
Configuration properties#
Property name |
Description |
Default |
---|---|---|
|
The comma-separated list of host names for the Elasticsearch node to connect to. This property is required. |
|
|
Port of the Elasticsearch node to connect to. |
|
|
The schema that contains all tables defined without a qualifying schema name. |
|
|
Sets the maximum number of hits that can be returned with each Elasticsearch scroll request. |
|
|
Amount of time Elasticsearch keeps the search context alive for scroll requests. |
|
|
Timeout value for all Elasticsearch requests. |
|
|
Timeout value for all Elasticsearch connection attempts. |
|
|
The minimum duration between backpressure retry attempts for a single request to Elasticsearch. Setting it too low might overwhelm an already struggling ES cluster. |
|
|
The maximum duration between backpressure retry attempts for a single request to Elasticsearch. |
|
|
The maximum duration across all retry attempts for a single request to Elasticsearch. |
|
|
How often the list of available Elasticsearch nodes is refreshed. |
|
|
Disables using the address published by Elasticsearch to connect for queries. |
TLS security#
The Elasticsearch connector provides additional security options to support Elasticsearch clusters that have been configured to use TLS.
If your cluster has globally-trusted certificates, you should only need to enable TLS. If you require custom configuration for certificates, the connector supports key stores and trust stores in PEM or Java Key Store (JKS) format.
The allowed configuration values are:
Property name |
Description |
---|---|
|
Enables TLS security. |
|
|
|
|
|
The key password for the key store specified by
|
|
The key password for the trust store specified by
|
Type mapping#
Because Trino and Elasticsearch each support types that the other does not, this connector maps some types when reading data.
Elasticsearch type to Trino type mapping#
The connector maps Elasticsearch types to the corresponding Trino types according to the following table:
Elasticsearch type |
Trino type |
Notes |
---|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
For more information, see Date types. |
|
|
No other types are supported.
Array types#
Fields in Elasticsearch can contain zero or more values , but there is no dedicated array type. To indicate a field contains an array, it can be annotated in a Trino-specific structure in the _meta section of the index mapping.
For example, you can have an Elasticsearch index that contains documents with the following structure:
{
"array_string_field": ["trino","the","lean","machine-ohs"],
"long_field": 314159265359,
"id_field": "564e6982-88ee-4498-aa98-df9e3f6b6109",
"timestamp_field": "1987-09-17T06:22:48.000Z",
"object_field": {
"array_int_field": [86,75,309],
"int_field": 2
}
}
The array fields of this structure can be defined by using the following command to add the field
property definition to the _meta.trino
property of the target index mapping.
curl --request PUT \
--url localhost:9200/doc/_mapping \
--header 'content-type: application/json' \
--data '
{
"_meta": {
"trino":{
"array_string_field":{
"isArray":true
},
"object_field":{
"array_int_field":{
"isArray":true
}
},
}
}
}'
Note
It is not allowed to use asRawJson
and isArray
flags simultaneously for the same column.
Date types#
Elasticsearch supports a wide array of date formats including
built-in date formats and also custom date formats.
The Elasticsearch connector supports only the default date
type. All other
date formats including built-in date formats and custom date formats are
not supported. Dates with the format property are ignored.
Raw JSON transform#
There are many occurrences where documents in Elasticsearch have more complex
structures that are not represented in the mapping. For example, a single
keyword
field can have widely different content including a single
keyword
value, an array, or a multidimensional keyword
array with any
level of nesting.
curl --request PUT \
--url localhost:9200/doc/_mapping \
--header 'content-type: application/json' \
--data '
{
"properties": {
"array_string_field":{
"type": "keyword"
}
}
}'
Notice for the array_string_field
that all the following documents are legal
for Elasticsearch. See the Elasticsearch array documentation
for more details.
[
{
"array_string_field": "trino"
},
{
"array_string_field": ["trino","is","the","besto"]
},
{
"array_string_field": ["trino",["is","the","besto"]]
},
{
"array_string_field": ["trino",["is",["the","besto"]]]
}
]
Further, Elasticsearch supports types, such as
dense_vector,
that are not supported in Trino. New types are constantly emerging which can
cause parsing exceptions for users that use of these types in Elasticsearch. To
manage all of these scenarios, you can transform fields to raw JSON by
annotating it in a Trino-specific structure in the _meta
section of the index mapping. This indicates to Trino that the field, and all
nested fields beneath, need to be cast to a VARCHAR
field that contains
the raw JSON content. These fields can be defined by using the following command
to add the field property definition to the _meta.presto
property of the
target index mapping.
curl --request PUT \
--url localhost:9200/doc/_mapping \
--header 'content-type: application/json' \
--data '
{
"_meta": {
"presto":{
"array_string_field":{
"asRawJson":true
}
}
}
}'
This preceding configurations causes Trino to return the array_string_field
field as a VARCHAR
containing raw JSON. You can parse these fields with the
built-in JSON functions.
Note
It is not allowed to use asRawJson
and isArray
flags simultaneously for the same column.
Special columns#
The following hidden columns are available:
Column |
Description |
---|---|
_id |
The Elasticsearch document ID |
_score |
The document score returned by the Elasticsearch query |
_source |
The source of the original document |
Full text queries#
Trino SQL queries can be combined with Elasticsearch queries by providing the full text query as part of the table name, separated by a colon. For example:
SELECT * FROM "tweets: +trino SQL^2"
Predicate push down#
The connector supports predicate push down of below data types:
Elasticsearch |
Trino |
Supports |
---|---|---|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
(all others) |
(unsupported) |
(unsupported) |
Password authentication#
To enable password authentication, the elasticsearch.security
option needs to be set to PASSWORD
.
Additionally the following options need to be configured appropriately:
Property name |
Description |
---|---|
|
User name to use to connect to Elasticsearch. |
|
Password to use to connect to Elasticsearch. |
SQL support#
The connector provides globally available and read operation statements to access data and metadata in the Elasticsearch catalog.
Table functions#
The connector provides specific table functions to access Elasticsearch.
raw_query(varchar) -> table
#
The raw_query
function allows you to query the underlying database directly.
This function requires Elastic Query DSL
syntax, because the full query is pushed down and processed in Elasticsearch.
This can be useful for accessing native features which are not available in
Trino or for improving query performance in situations where running a query
natively may be faster.
The native query passed to the underlying data source is required to return a table as a result set. Only the data source performs validation or security checks for these queries using its own configuration. Trino does not perform these tasks. Only use passthrough queries to read data.
The raw_query
function requires three parameters:
schema
: The schema in the catalog that the query is to be executed on.index
: The index in Elasticsearch to be searched.query
: The query to be executed, written in Elastic Query DSL.
Once executed, the query returns a single row containing the resulting JSON payload returned by Elasticsearch.
For example, query the example
catalog and use the raw_query
table
function to search for documents in the orders
index where the country name
is ALGERIA
:
SELECT
*
FROM
TABLE(
example.system.raw_query(
schema => 'sales',
index => 'orders',
query => '{
"query": {
"match": {
"name": "ALGERIA"
}
}
}'
)
);
Note
The query engine does not preserve the order of the results of this
function. If the passed query contains an ORDER BY
clause, the
function result may not be ordered as expected.