git clone --depth 1 https://github.com/apache/shardingsphere.git
cd shardingsphere
mvn clean install -Dmaven.javadoc.skip=true -Dcheckstyle.skip=true -Drat.skip=true -Djacoco.skip=true -DskipITs -DskipTests -Prelease
The binary packages:
Or get binary package from download page.
Scaling is an experimental feature, if scaling job fail, you could try nightly version, click here to download nightly build.
Unzip the proxy distribution package, modify the configuration file conf/config-sharding.yaml
. Please refer to proxy startup manual for more details.
Modify the configuration file conf/server.yaml
. Please refer to Mode Configuration for more details.
Type of mode
must be Cluster
for now, please start the registry center before running proxy.
Configuration Example:
mode:
type: Cluster
repository:
type: ZooKeeper
props:
namespace: governance_ds
server-lists: localhost:2181
retryIntervalMilliseconds: 500
timeToLiveSeconds: 60
maxRetries: 3
operationTimeoutMilliseconds: 500
overwrite: false
Way 1. Modify scalingName
and scaling
configuration in conf/config-sharding.yaml
.
Configuration Items Explanation:
rules:
- !SHARDING
# ignored configuration
scalingName: # Enabled scaling action config name
scaling:
<scaling-action-config-name> (+):
input: # Data read configuration. If it's not configured, then part of its configuration will take effect.
workerThread: # Worker thread pool size for inventory data ingestion from source. If it's not configured, then use system default value.
batchSize: # Maximum records count of a DML select operation. If it's not configured, then use system default value.
rateLimiter: # Rate limit algorithm. If it's not configured, then system will skip rate limit.
type: # Algorithm type. Options:
props: # Algorithm properties
output: # Data write configuration. If it's not configured, then part of its configuration will take effect.
workerThread: # Worker thread pool size for data importing to target. If it's not configured, then use system default value.
batchSize: # Maximum records count of a DML insert/delete/update operation. If it's not configured, then use system default value.
rateLimiter: # Rate limit algorithm. If it's not configured, then system will skip rate limit.
type: # Algorithm type. Options:
props: # Algorithm properties
streamChannel: # Algorithm of channel that connect producer and consumer, used for input and output. If it's not configured, then system will use MEMORY type
type: # Algorithm type. Options: MEMORY
props: # Algorithm properties
block-queue-size: # Property: data channel block queue size. Available for types: MEMORY
completionDetector: # Completion detect algorithm. If it's not configured, then system won't continue to do next steps automatically.
type: # Algorithm type. Options: IDLE
props: # Algorithm properties
incremental-task-idle-minute-threshold: # If incremental tasks is idle more than so much minutes, then it could be considered as almost completed. Available for types: IDLE
dataConsistencyChecker: # Data consistency check algorithm. If it's not configured, then system will skip this step.
type: # Algorithm type. Options: DATA_MATCH, CRC32_MATCH
props: # Algorithm properties
chunk-size: # Maximum records count of a query operation for check
Configuration Example:
rules:
- !SHARDING
# ignored configuration
scalingName: default_scaling
scaling:
default_scaling:
input:
workerThread: 40
batchSize: 1000
output:
workerThread: 40
batchSize: 1000
streamChannel:
type: MEMORY
props:
block-queue-size: 10000
completionDetector:
type: IDLE
props:
incremental-task-idle-minute-threshold: 30
dataConsistencyChecker:
type: DATA_MATCH
props:
chunk-size: 1000
You could customize completionDetector
, dataConsistencyChecker
algorithm by implementing SPI. Current implementation could be referenced, please refer to Dev Manual#Scaling for more details.
Way 2: Configure scaling by DistSQL
Create scaling configuration example:
CREATE SHARDING SCALING RULE default_scaling (
INPUT(
WORKER_THREAD=40,
BATCH_SIZE=1000
),
OUTPUT(
WORKER_THREAD=40,
BATCH_SIZE=1000
),
STREAM_CHANNEL(TYPE(NAME=MEMORY, PROPERTIES("block-queue-size"=10000))),
COMPLETION_DETECTOR(TYPE(NAME=IDLE, PROPERTIES("incremental-task-idle-minute-threshold"=3))),
DATA_CONSISTENCY_CHECKER(TYPE(NAME=DATA_MATCH, PROPERTIES("chunk-size"=1000)))
);
Please refer to RDL#Sharding for more details.
sh bin/start.sh
logs/stdout.log
:[INFO ] [main] o.a.s.p.frontend.ShardingSphereProxy - ShardingSphere-Proxy start success
It means proxy
start up successfully.
sh bin/stop.sh