id | title |
---|---|
aliyun-oss |
Aliyun OSS |
Alibaba Cloud is the 3rd largest cloud infrastructure provider in the world. It provides its own storage solution known as OSS, Object Storage Service. This document describes how to use OSS as Druid deep storage.
Use the pull-deps tool shipped with Druid to install the aliyun-oss-extensions
extension, as described here on middle manager and historical nodes.
java -classpath "{YOUR_DRUID_DIR}/lib/*" org.apache.druid.cli.Main tools pull-deps -c org.apache.druid.extensions.contrib:aliyun-oss-extensions:{YOUR_DRUID_VERSION}
After installation, add this aliyun-oss-extensions
extension to druid.extensions.loadList
in common.runtime.properties and then restart middle manager and historical nodes.
First add the following OSS configurations to common.runtime.properties
Property | Description | Required |
---|---|---|
druid.oss.accessKey |
The AccessKey ID of the account to be used to access the OSS bucket |
yes |
druid.oss.secretKey |
The AccessKey Secret of the account to be used to access the OSS bucket |
yes |
druid.oss.endpoint |
The endpoint URL of your OSS storage. If your Druid cluster is also hosted in the same region on Alibaba Cloud as the region of your OSS bucket, it's recommended to use the internal network endpoint url, so that any inbound and outbound traffic to the OSS bucket is free of charge. |
yes |
To use OSS as deep storage, add the following configurations:
Property | Description | Required |
---|---|---|
druid.storage.type |
Global deep storage provider. Must be set to oss to make use of this extension. |
yes |
druid.storage.oss.bucket |
Storage bucket name. | yes |
druid.storage.oss.prefix |
Folder where segments will be published to. druid/segments is recommended. |
No |
If OSS is used as deep storage for segment files, it's also recommended saving index logs in the OSS too. To do this, add following configurations:
Property | Description | Required |
---|---|---|
druid.indexer.logs.type |
Global deep storage provider. Must be set to oss to make use of this extension. |
yes |
druid.indexer.logs.oss.bucket |
The bucket used to keep logs. It could be the same as druid.storage.oss.bucket |
yes |
druid.indexer.logs.oss.prefix |
Folder where log files will be published to. druid/logs is recommended. |
no |
Currently, Web Console does not support ingestion from OSS, but it could be done by submitting an ingestion task with OSS's input source configuration.
Below shows the configurations of OSS's input source.
property | description | Required |
---|---|---|
type | This should be oss . |
yes |
uris | JSON array of URIs where OSS objects to be ingested are located. For example, oss://{your_bucket}/{source_file_path} |
uris or prefixes or objects must be set |
prefixes | JSON array of URI prefixes for the locations of OSS objects to be ingested. Empty objects starting with one of the given prefixes will be skipped. | uris or prefixes or objects must be set |
objects | JSON array of OSS Objects to be ingested. | uris or prefixes or objects must be set |
properties | Properties Object for overriding the default OSS configuration. See below for more information. | no (defaults will be used if not given) |
Property | Description | Default | Required |
---|---|---|---|
bucket | Name of the OSS bucket | None | yes |
path | The path where data is located. | None | yes |
Property | Description | Default | Required |
---|---|---|---|
accessKey | The Password Provider or plain text string of this OSS InputSource's access key | None | yes |
secretKey | The Password Provider or plain text string of this OSS InputSource's secret key | None | yes |
endpoint | The endpoint of this OSS InputSource | None | no |
Say that the file rollup-data.json
, which can be found under Druid's quickstart/tutorial
directory, has been uploaded to a folder druid
in your OSS bucket, the bucket for which your Druid is configured.
In this case, the uris
property of the OSS's input source can be used for reading, as shown:
{
"type" : "index_parallel",
"spec" : {
"dataSchema" : {
"dataSource" : "rollup-tutorial-from-oss",
"timestampSpec": {
"column": "timestamp",
"format": "iso"
},
"dimensionsSpec" : {
"dimensions" : [
"srcIP",
"dstIP"
]
},
"metricsSpec" : [
{ "type" : "count", "name" : "count" },
{ "type" : "longSum", "name" : "packets", "fieldName" : "packets" },
{ "type" : "longSum", "name" : "bytes", "fieldName" : "bytes" }
],
"granularitySpec" : {
"type" : "uniform",
"segmentGranularity" : "week",
"queryGranularity" : "minute",
"intervals" : ["2018-01-01/2018-01-03"],
"rollup" : true
}
},
"ioConfig" : {
"type" : "index_parallel",
"inputSource" : {
"type" : "oss",
"uris" : [
"oss://{YOUR_BUCKET_NAME}/druid/rollup-data.json"
]
},
"inputFormat" : {
"type" : "json"
},
"appendToExisting" : false
},
"tuningConfig" : {
"type" : "index_parallel",
"maxRowsPerSegment" : 5000000,
"maxRowsInMemory" : 25000
}
}
}
By posting the above ingestion task spec to http://{YOUR_ROUTER_IP}:8888/druid/indexer/v1/task
, an ingestion task will be created by the indexing service to ingest.
If we want to read files in a same folder, we could use the prefixes
property to specify the folder name where Druid could find input files instead of specifying file URIs one by one.
...
"ioConfig" : {
"type" : "index_parallel",
"inputSource" : {
"type" : "oss",
"prefixes" : [
"oss://{YOUR_BUCKET_NAME}/2020", "oss://{YOUR_BUCKET_NAME}/2021"
]
},
"inputFormat" : {
"type" : "json"
},
"appendToExisting" : false
}
...
The spec above tells the ingestion task to read all files under 2020
and 2021
folders.
If you want to read from files in buckets which are different from the bucket Druid is configured, use objects
property of OSS's InputSource for task submission as below:
...
"ioConfig" : {
"type" : "index_parallel",
"inputSource" : {
"type" : "oss",
"objects" : [
{"bucket": "YOUR_BUCKET_NAME", "path": "druid/rollup-data.json"}
]
},
"inputFormat" : {
"type" : "json"
},
"appendToExisting" : false
}
...
If the default druid.oss.accessKey
is not able to access a bucket, properties
could be used to customize these secret information as below:
...
"ioConfig" : {
"type" : "index_parallel",
"inputSource" : {
"type" : "oss",
"objects" : [
{"bucket": "YOUR_BUCKET_NAME", "path": "druid/rollup-data.json"}
],
"properties": {
"endpoint": "YOUR_ENDPOINT_OF_BUCKET",
"accessKey": "YOUR_ACCESS_KEY",
"secretKey": "YOUR_SECRET_KEY"
}
},
"inputFormat" : {
"type" : "json"
},
"appendToExisting" : false
}
...
This properties
could be applied to any of uris
, objects
, prefixes
property above.
When using OSS as deep storage or reading from OSS, the most problems that users will encounter are related to OSS permission. Please refer to the official OSS permission troubleshooting document to find a solution.