日韩av黄I国产麻豆传媒I国产91av视频在线观看I日韩一区二区三区在线看I美女国产在线I麻豆视频国产在线观看I成人黄色短片

歡迎訪問 生活随笔!

生活随笔

當前位置: 首頁 > 编程资源 > 编程问答 >内容正文

编程问答

docker 安装redis第三方集群方案 codis

發布時間:2025/3/20 编程问答 42 豆豆
生活随笔 收集整理的這篇文章主要介紹了 docker 安装redis第三方集群方案 codis 小編覺得挺不錯的,現在分享給大家,幫大家做個參考.
docker 安裝redis第三方集群方案 codis

docker 安裝redis第三方集群方案 codis

?首先,安裝好docker環境,這里不提,需要看的可以在我的博客里找,

事先準備好zookeeper,這里僅僅是起一個單實例做測試,生產環境應該是3個節點以上的集群,比如:

docker run -d -it --name zookeeper -p 2181:2181 zookeeper;

docker倉庫 和 docker服務都跑在一臺服務器上172.16.37.74;

然后

docker?pull centos;

docker run -d -it --name codis ? centos bash;

docker exec -it codis bash;

yum makecache;yum update -y;yum install -y golang;yum install -y git;yum install -y make;

mkdir?/golang

export GOROOT=/usr/lib/golang;

export GOPATH=/golang;

export GOBIN=$GOROOT/bin;

export GOARCH=amd64;

export GOOS=linux;

export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:$GOBIN:$GOROOT/bin;

source /etc/profile;

這時候golang環境裝好了,我這里沒有指定golang1.8.x版本,如果到了將來centos的官方golang版本升級了,可能需要指定一下版本;

mkdir -p $GOPATH/src/github.com/CodisLabs;

cd $GOPATH/src/github.com/CodisLabs;

git clone https://github.com/CodisLabs/codis.git -b release3.2; make;

如果能編譯成功,最好,經常我自己都編譯不成功。。。好在官方有編譯好的bin目錄命令包可以直接下載使用,到https://github.com/CodisLabs/codis/releases找最新版本,我寫這博客時是codis3.2.1-go1.8.3-linux.tar.gz,瀏覽器下載到工作電腦本地,然后在本地起webserver,然后改文件名為codis3.2.1.zip;我的工作電腦和我的測試服務器在同一個內網,所以:

cd /root;

curl -O http://工作電腦ip/codis3.2.1.zip;

tar -xvf?codis3.2.1.zip;

cd?codis3.2.1-go1.8.3-linux;

?mv -f * /golang/src/github.com/CodisLabs/codis/bin/;

cd?/golang/src/github.com/CodisLabs/codis/bin/;

?ls;

顯示:

[root@bbca348beae6 bin]# ls
assets codis-dashboard codis-ha codis-server redis-cli version
codis-admin codis-fe codis-proxy redis-benchmark redis-sentinel

修改一下 /etc/hosts

添加一下幾行

172.16.37.74 zookeeper1
172.16.37.74 zookeeper2
172.16.37.74 zookeeper3
172.16.37.74 codis-dashboard
172.16.37.74 codis-fe

172.16.37.74 codis-proxy1

172.16.37.74 codis-proxy2

172.16.37.74 codis-proxy3

172.16.37.74 codis-server1main

172.16.37.74 codis-server1slave

172.16.37.74 codis-server2main

172.16.37.74 codis-server2slave

172.16.37.74 codis-server3main

172.16.37.74 codis-server3slave

172.16.37.74 codis-server4main

172.16.37.74 codis-server4slave

172.16.37.74 codis-server5main

172.16.37.74 codis-server5slave

172.16.37.74 codis-server6main

172.16.37.74 codis-server6slave

注意?172.16.37.74是我環境的ip ?現實里應該按照實際情況修改

我做了一個簡單的偽實例啟動腳本:

vi /start.sh

#!/bin/bash
export GOROOT=/usr/local/go
export GOPATH=/golang
export GOBIN=$GOPATH/bin
export GOARCH=amd64
export GOOS=linux
export PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:$GOBIN:$GOROOT/bin
##export ZOOKEEPER_HOME=/zookeeper
##export PATH=$ZOOKEEPER_HOME/bin:$PATH
source /etc/profile
cd /golang/src/github.com/CodisLabs/codis/log
##/zookeeper/bin/zkServer.sh start
nohup /golang/src/github.com/CodisLabs/codis/bin/codis-dashboard --ncpu=2 --config=/golang/src/github.com/CodisLabs/codis/config/dashboard.toml --log=/golang/src/github.com/CodisLabs/codis/log/dashboard.log --log-level=WARN >> /golang/src/github.com/CodisLabs/codis/log/codis_dashboard.log &
ps -aux |egrep "(codis-proxy)" | cut -c 9-15 | xargs kill
nohup /golang/src/github.com/CodisLabs/codis/bin/codis-proxy --ncpu=1 --config=/golang/src/github.com/CodisLabs/codis/config/proxy.toml --log=/golang/src/github.com/CodisLabs/codis/log/proxy.log --log-level=WARN >> /golang/src/github.com/CodisLabs/codis/log/codis_proxy.log &
/golang/src/github.com/CodisLabs/codis/bin/codis-admin --dashboard=codis-dashboard:18080 --create-proxy -x codis-proxy1:11080
/golang/src/github.com/CodisLabs/codis/bin/codis-server /golang/src/github.com/CodisLabs/codis/config/redis.conf;
/golang/src/github.com/CodisLabs/codis/bin/codis-server /golang/src/github.com/CodisLabs/codis/config/redis1.conf;
/golang/src/github.com/CodisLabs/codis/bin/codis-server /golang/src/github.com/CodisLabs/codis/config/redis2.conf;
nohup /golang/src/github.com/CodisLabs/codis/bin/codis-fe --ncpu=1 --log=/golang/src/github.com/CodisLabs/codis/log/fe.log --log-level=WARN --zookeeper=zookeeper1:2181 --listen=0.0.0.0:9090 &
bash

wq 保存

chmod +x /start.sh

該腳本依賴的幾個配置文件,首先主板進程的配置文件

vi?/golang/src/github.com/CodisLabs/codis/config/dashboard.toml

##################################################
# #
# Codis-Dashboard #
# #
##################################################

# Set Coordinator, only accept "zookeeper" & "etcd" & "filesystem".
# Quick Start
#coordinator_name = "filesystem"
#coordinator_addr = "/tmp/codis"
coordinator_name = "zookeeper"
coordinator_addr = "zookeeper1:2181,zookeeper2:2181,zookeeper3:2181"

# Set Codis Product Name/Auth.
product_name = "codis-test-machine"
product_auth = ""

# Set bind address for admin(rpc), tcp only.
admin_addr = "0.0.0.0:18080"

# Set arguments for data migration (only accept 'sync' & 'semi-async').
migration_method = "semi-async"
migration_parallel_slots = 100
migration_async_maxbulks = 200
migration_async_maxbytes = "32mb"
migration_async_numkeys = 500
migration_timeout = "30s"

# Set configs for redis sentinel.
sentinel_quorum = 2
sentinel_parallel_syncs = 1
sentinel_down_after = "30s"
sentinel_failover_timeout = "5m"
sentinel_notification_script = ""
sentinel_client_reconfig_script = ""

wq保存

vi /golang/src/github.com/CodisLabs/codis/config/proxy.toml

##################################################
# #
# Codis-Proxy #
# #
##################################################

# Set Codis Product Name/Auth.
product_name = "codis-test-machine"
product_auth = ""

# Set auth for client session
# 1. product_auth is used for auth validation among codis-dashboard,
# codis-proxy and codis-server.
# 2. session_auth is different from product_auth, it requires clients
# to issue AUTH <PASSWORD> before processing any other commands.
session_auth = ""

# Set bind address for admin(rpc), tcp only.
admin_addr = "0.0.0.0:11080"

# Set bind address for proxy, proto_type can be "tcp", "tcp4", "tcp6", "unix" or "unixpacket".
proto_type = "tcp4"
proxy_addr = "0.0.0.0:19000"

# Set jodis address & session timeout
# 1. jodis_name is short for jodis_coordinator_name, only accept "zookeeper" & "etcd".
# 2. jodis_addr is short for jodis_coordinator_addr
# 3. proxy will be registered as node:
# if jodis_compatible = true (not suggested):
# /zk/codis/db_{PRODUCT_NAME}/proxy-{HASHID} (compatible with Codis2.0)
# or else
# /jodis/{PRODUCT_NAME}/proxy-{HASHID}
jodis_name = "zookeeper"
jodis_addr = "zookeeper1:2181,zookeeper2:2181,zookeeper3:2181"
jodis_timeout = "20s"
jodis_compatible = true

# Set datacenter of proxy.
proxy_datacenter = ""

# Set max number of alive sessions.
proxy_max_clients = 1000

# Set max offheap memory size. (0 to disable)
proxy_max_offheap_size = "1024mb"

# Set heap placeholder to reduce GC frequency.
proxy_heap_placeholder = "256mb"

# Proxy will ping backend redis (and clear 'MASTERDOWN' state) in a predefined interval. (0 to disable)
backend_ping_period = "5s"

# Set backend recv buffer size & timeout.
backend_recv_bufsize = "128kb"
backend_recv_timeout = "30s"

# Set backend send buffer & timeout.
backend_send_bufsize = "128kb"
backend_send_timeout = "30s"

# Set backend pipeline buffer size.
backend_max_pipeline = 20480

# Set backend never read replica groups, default is false
backend_primary_only = false

# Set backend parallel connections per server
backend_primary_parallel = 1
backend_replica_parallel = 1

# Set backend tcp keepalive period. (0 to disable)
backend_keepalive_period = "75s"

# Set number of databases of backend.
backend_number_databases = 16

# If there is no request from client for a long time, the connection will be closed. (0 to disable)
# Set session recv buffer size & timeout.
session_recv_bufsize = "128kb"
session_recv_timeout = "30m"

# Set session send buffer size & timeout.
session_send_bufsize = "64kb"
session_send_timeout = "30s"

# Make sure this is higher than the max number of requests for each pipeline request, or your client may be blocked.
# Set session pipeline buffer size.
session_max_pipeline = 10000

# Set session tcp keepalive period. (0 to disable)
session_keepalive_period = "75s"

# Set session to be sensitive to failures. Default is false, instead of closing socket, proxy will send an error response to client.
session_break_on_failure = false

# Set metrics server (such as http://localhost:28000), proxy will report json formatted metrics to specified server in a predefined period.
metrics_report_server = ""
metrics_report_period = "1s"

# Set influxdb server (such as http://localhost:8086), proxy will report metrics to influxdb.
metrics_report_influxdb_server = ""
metrics_report_influxdb_period = "1s"
metrics_report_influxdb_username = ""
metrics_report_influxdb_password = ""
metrics_report_influxdb_database = ""

# Set statsd server (such as localhost:8125), proxy will report metrics to statsd.
metrics_report_statsd_server = ""
metrics_report_statsd_period = "1s"
metrics_report_statsd_prefix = ""

wq保存

分片的配置文件:

vi??/golang/src/github.com/CodisLabs/codis/config/redis.conf

# Redis configuration file example.
#
# Note that in order to read the configuration file, Redis must be
# started with the file path as first argument:
#
# ./redis-server /path/to/redis.conf

# Note on units: when memory size is needed, it is possible to specify
# it in the usual form of 1k 5GB 4M and so forth:
#
# 1k => 1000 bytes
# 1kb => 1024 bytes
# 1m => 1000000 bytes
# 1mb => 1024*1024 bytes
# 1g => 1000000000 bytes
# 1gb => 1024*1024*1024 bytes
#
# units are case insensitive so 1GB 1Gb 1gB are all the same.

################################## INCLUDES ###################################

# Include one or more other config files here. This is useful if you
# have a standard template that goes to all Redis servers but also need
# to customize a few per-server settings. Include files can include
# other files, so use this wisely.
#
# Notice option "include" won't be rewritten by command "CONFIG REWRITE"
# from admin or Redis Sentinel. Since Redis always uses the last processed
# line as value of a configuration directive, you'd better put includes
# at the beginning of this file to avoid overwriting config change at runtime.
#
# If instead you are interested in using includes to override configuration
# options, it is better to use include as the last line.
#
# include /path/to/local.conf
# include /path/to/other.conf

################################## NETWORK #####################################

# By default, if no "bind" configuration directive is specified, Redis listens
# for connections from all the network interfaces available on the server.
# It is possible to listen to just one or multiple selected interfaces using
# the "bind" configuration directive, followed by one or more IP addresses.
#
# Examples:
#
# bind 192.168.1.100 10.0.0.1
# bind 127.0.0.1 ::1
#
# ~~~ WARNING ~~~ If the computer running Redis is directly exposed to the
# internet, binding to all the interfaces is dangerous and will expose the
# instance to everybody on the internet. So by default we uncomment the
# following bind directive, that will force Redis to listen only into
# the IPv4 lookback interface address (this means Redis will be able to
# accept connections only from clients running into the same computer it
# is running).
#
# IF YOU ARE SURE YOU WANT YOUR INSTANCE TO LISTEN TO ALL THE INTERFACES
# JUST COMMENT THE FOLLOWING LINE.
# ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
bind 0.0.0.0

# Protected mode is a layer of security protection, in order to avoid that
# Redis instances left open on the internet are accessed and exploited.
#
# When protected mode is on and if:
#
# 1) The server is not binding explicitly to a set of addresses using the
# "bind" directive.
# 2) No password is configured.
#
# The server only accepts connections from clients connecting from the
# IPv4 and IPv6 loopback addresses 127.0.0.1 and ::1, and from Unix domain
# sockets.
#
# By default protected mode is enabled. You should disable it only if
# you are sure you want clients from other hosts to connect to Redis
# even if no authentication is configured, nor a specific set of interfaces
# are explicitly listed using the "bind" directive.
protected-mode yes

# Accept connections on the specified port, default is 6379 (IANA #815344).
# If port 0 is specified Redis will not listen on a TCP socket.
port 6379

# TCP listen() backlog.
#
# In high requests-per-second environments you need an high backlog in order
# to avoid slow clients connections issues. Note that the Linux kernel
# will silently truncate it to the value of /proc/sys/net/core/somaxconn so
# make sure to raise both the value of somaxconn and tcp_max_syn_backlog
# in order to get the desired effect.
tcp-backlog 511

# Unix socket.
#
# Specify the path for the Unix socket that will be used to listen for
# incoming connections. There is no default, so Redis will not listen
# on a unix socket when not specified.
#
# unixsocket /tmp/redis.sock
# unixsocketperm 700

# Close the connection after a client is idle for N seconds (0 to disable)
timeout 0

# TCP keepalive.
#
# If non-zero, use SO_KEEPALIVE to send TCP ACKs to clients in absence
# of communication. This is useful for two reasons:
#
# 1) Detect dead peers.
# 2) Take the connection alive from the point of view of network
# equipment in the middle.
#
# On Linux, the specified value (in seconds) is the period used to send ACKs.
# Note that to close the connection the double of the time is needed.
# On other kernels the period depends on the kernel configuration.
#
# A reasonable value for this option is 300 seconds, which is the new
# Redis default starting with Redis 3.2.1.
tcp-keepalive 300

################################# GENERAL #####################################

# By default Redis does not run as a daemon. Use 'yes' if you need it.
# Note that Redis will write a pid file in /var/run/redis.pid when daemonized.
daemonize yes

# If you run Redis from upstart or systemd, Redis can interact with your
# supervision tree. Options:
# supervised no - no supervision interaction
# supervised upstart - signal upstart by putting Redis into SIGSTOP mode
# supervised systemd - signal systemd by writing READY=1 to $NOTIFY_SOCKET
# supervised auto - detect upstart or systemd method based on
# UPSTART_JOB or NOTIFY_SOCKET environment variables
# Note: these supervision methods only signal "process is ready."
# They do not enable continuous liveness pings back to your supervisor.
supervised no

# If a pid file is specified, Redis writes it where specified at startup
# and removes it at exit.
#
# When the server runs non daemonized, no pid file is created if none is
# specified in the configuration. When the server is daemonized, the pid file
# is used even if not specified, defaulting to "/var/run/redis.pid".
#
# Creating a pid file is best effort: if Redis is not able to create it
# nothing bad happens, the server will start and run normally.
pidfile /tmp/redis_6379.pid

# Specify the server verbosity level.
# This can be one of:
# debug (a lot of information, useful for development/testing)
# verbose (many rarely useful info, but not a mess like the debug level)
# notice (moderately verbose, what you want in production probably)
# warning (only very important / critical messages are logged)
loglevel notice

# Specify the log file name. Also the empty string can be used to force
# Redis to log on the standard output. Note that if you use standard
# output for logging but daemonize, logs will be sent to /dev/null
logfile "/tmp/redis_6379.log"

# To enable logging to the system logger, just set 'syslog-enabled' to yes,
# and optionally update the other syslog parameters to suit your needs.
# syslog-enabled no

# Specify the syslog identity.
# syslog-ident redis

# Specify the syslog facility. Must be USER or between LOCAL0-LOCAL7.
# syslog-facility local0

# Set the number of databases. The default database is DB 0, you can select
# a different one on a per-connection basis using SELECT <dbid> where
# dbid is a number between 0 and 'databases'-1
databases 16

################################ SNAPSHOTTING ################################
#
# Save the DB on disk:
#
# save <seconds> <changes>
#
# Will save the DB if both the given number of seconds and the given
# number of write operations against the DB occurred.
#
# In the example below the behaviour will be to save:
# after 900 sec (15 min) if at least 1 key changed
# after 300 sec (5 min) if at least 10 keys changed
# after 60 sec if at least 10000 keys changed
#
# Note: you can disable saving completely by commenting out all "save" lines.
#
# It is also possible to remove all the previously configured save
# points by adding a save directive with a single empty string argument
# like in the following example:
#
# save ""

save 900 1
save 300 10
save 60 10000

# By default Redis will stop accepting writes if RDB snapshots are enabled
# (at least one save point) and the latest background save failed.
# This will make the user aware (in a hard way) that data is not persisting
# on disk properly, otherwise chances are that no one will notice and some
# disaster will happen.
#
# If the background saving process will start working again Redis will
# automatically allow writes again.
#
# However if you have setup your proper monitoring of the Redis server
# and persistence, you may want to disable this feature so that Redis will
# continue to work as usual even if there are problems with disk,
# permissions, and so forth.
stop-writes-on-bgsave-error yes

# Compress string objects using LZF when dump .rdb databases?
# For default that's set to 'yes' as it's almost always a win.
# If you want to save some CPU in the saving child set it to 'no' but
# the dataset will likely be bigger if you have compressible values or keys.
rdbcompression yes

# Since version 5 of RDB a CRC64 checksum is placed at the end of the file.
# This makes the format more resistant to corruption but there is a performance
# hit to pay (around 10%) when saving and loading RDB files, so you can disable it
# for maximum performances.
#
# RDB files created with checksum disabled have a checksum of zero that will
# tell the loading code to skip the check.
rdbchecksum yes

# The filename where to dump the DB
dbfilename dump.rdb

# The working directory.
#
# The DB will be written inside this directory, with the filename specified
# above using the 'dbfilename' configuration directive.
#
# The Append Only File will also be created inside this directory.
#
# Note that you must specify a directory here, not a file name.
dir ./

################################# REPLICATION #################################

# Master-Slave replication. Use slaveof to make a Redis instance a copy of
# another Redis server. A few things to understand ASAP about Redis replication.
#
# 1) Redis replication is asynchronous, but you can configure a master to
# stop accepting writes if it appears to be not connected with at least
# a given number of slaves.
# 2) Redis slaves are able to perform a partial resynchronization with the
# master if the replication link is lost for a relatively small amount of
# time. You may want to configure the replication backlog size (see the next
# sections of this file) with a sensible value depending on your needs.
# 3) Replication is automatic and does not need user intervention. After a
# network partition slaves automatically try to reconnect to masters
# and resynchronize with them.
#
# slaveof <masterip> <masterport>

# If the master is password protected (using the "requirepass" configuration
# directive below) it is possible to tell the slave to authenticate before
# starting the replication synchronization process, otherwise the master will
# refuse the slave request.
#
# masterauth <master-password>

# When a slave loses its connection with the master, or when the replication
# is still in progress, the slave can act in two different ways:
#
# 1) if slave-serve-stale-data is set to 'yes' (the default) the slave will
# still reply to client requests, possibly with out of date data, or the
# data set may just be empty if this is the first synchronization.
#
# 2) if slave-serve-stale-data is set to 'no' the slave will reply with
# an error "SYNC with master in progress" to all the kind of commands
# but to INFO and SLAVEOF.
#
slave-serve-stale-data yes

# You can configure a slave instance to accept writes or not. Writing against
# a slave instance may be useful to store some ephemeral data (because data
# written on a slave will be easily deleted after resync with the master) but
# may also cause problems if clients are writing to it because of a
# misconfiguration.
#
# Since Redis 2.6 by default slaves are read-only.
#
# Note: read only slaves are not designed to be exposed to untrusted clients
# on the internet. It's just a protection layer against misuse of the instance.
# Still a read only slave exports by default all the administrative commands
# such as CONFIG, DEBUG, and so forth. To a limited extent you can improve
# security of read only slaves using 'rename-command' to shadow all the
# administrative / dangerous commands.
slave-read-only yes

# Replication SYNC strategy: disk or socket.
#
# -------------------------------------------------------
# WARNING: DISKLESS REPLICATION IS EXPERIMENTAL CURRENTLY
# -------------------------------------------------------
#
# New slaves and reconnecting slaves that are not able to continue the replication
# process just receiving differences, need to do what is called a "full
# synchronization". An RDB file is transmitted from the master to the slaves.
# The transmission can happen in two different ways:
#
# 1) Disk-backed: The Redis master creates a new process that writes the RDB
# file on disk. Later the file is transferred by the parent
# process to the slaves incrementally.
# 2) Diskless: The Redis master creates a new process that directly writes the
# RDB file to slave sockets, without touching the disk at all.
#
# With disk-backed replication, while the RDB file is generated, more slaves
# can be queued and served with the RDB file as soon as the current child producing
# the RDB file finishes its work. With diskless replication instead once
# the transfer starts, new slaves arriving will be queued and a new transfer
# will start when the current one terminates.
#
# When diskless replication is used, the master waits a configurable amount of
# time (in seconds) before starting the transfer in the hope that multiple slaves
# will arrive and the transfer can be parallelized.
#
# With slow disks and fast (large bandwidth) networks, diskless replication
# works better.
repl-diskless-sync no

# When diskless replication is enabled, it is possible to configure the delay
# the server waits in order to spawn the child that transfers the RDB via socket
# to the slaves.
#
# This is important since once the transfer starts, it is not possible to serve
# new slaves arriving, that will be queued for the next RDB transfer, so the server
# waits a delay in order to let more slaves arrive.
#
# The delay is specified in seconds, and by default is 5 seconds. To disable
# it entirely just set it to 0 seconds and the transfer will start ASAP.
repl-diskless-sync-delay 5

# Slaves send PINGs to server in a predefined interval. It's possible to change
# this interval with the repl_ping_slave_period option. The default value is 10
# seconds.
#
# repl-ping-slave-period 10

# The following option sets the replication timeout for:
#
# 1) Bulk transfer I/O during SYNC, from the point of view of slave.
# 2) Master timeout from the point of view of slaves (data, pings).
# 3) Slave timeout from the point of view of masters (REPLCONF ACK pings).
#
# It is important to make sure that this value is greater than the value
# specified for repl-ping-slave-period otherwise a timeout will be detected
# every time there is low traffic between the master and the slave.
#
# repl-timeout 60

# Disable TCP_NODELAY on the slave socket after SYNC?
#
# If you select "yes" Redis will use a smaller number of TCP packets and
# less bandwidth to send data to slaves. But this can add a delay for
# the data to appear on the slave side, up to 40 milliseconds with
# Linux kernels using a default configuration.
#
# If you select "no" the delay for data to appear on the slave side will
# be reduced but more bandwidth will be used for replication.
#
# By default we optimize for low latency, but in very high traffic conditions
# or when the master and slaves are many hops away, turning this to "yes" may
# be a good idea.
repl-disable-tcp-nodelay no

# Set the replication backlog size. The backlog is a buffer that accumulates
# slave data when slaves are disconnected for some time, so that when a slave
# wants to reconnect again, often a full resync is not needed, but a partial
# resync is enough, just passing the portion of data the slave missed while
# disconnected.
#
# The bigger the replication backlog, the longer the time the slave can be
# disconnected and later be able to perform a partial resynchronization.
#
# The backlog is only allocated once there is at least a slave connected.
#
# repl-backlog-size 1mb

# After a master has no longer connected slaves for some time, the backlog
# will be freed. The following option configures the amount of seconds that
# need to elapse, starting from the time the last slave disconnected, for
# the backlog buffer to be freed.
#
# A value of 0 means to never release the backlog.
#
# repl-backlog-ttl 3600

# The slave priority is an integer number published by Redis in the INFO output.
# It is used by Redis Sentinel in order to select a slave to promote into a
# master if the master is no longer working correctly.
#
# A slave with a low priority number is considered better for promotion, so
# for instance if there are three slaves with priority 10, 100, 25 Sentinel will
# pick the one with priority 10, that is the lowest.
#
# However a special priority of 0 marks the slave as not able to perform the
# role of master, so a slave with priority of 0 will never be selected by
# Redis Sentinel for promotion.
#
# By default the priority is 100.
slave-priority 100

# It is possible for a master to stop accepting writes if there are less than
# N slaves connected, having a lag less or equal than M seconds.
#
# The N slaves need to be in "online" state.
#
# The lag in seconds, that must be <= the specified value, is calculated from
# the last ping received from the slave, that is usually sent every second.
#
# This option does not GUARANTEE that N replicas will accept the write, but
# will limit the window of exposure for lost writes in case not enough slaves
# are available, to the specified number of seconds.
#
# For example to require at least 3 slaves with a lag <= 10 seconds use:
#
# min-slaves-to-write 3
# min-slaves-max-lag 10
#
# Setting one or the other to 0 disables the feature.
#
# By default min-slaves-to-write is set to 0 (feature disabled) and
# min-slaves-max-lag is set to 10.

# A Redis master is able to list the address and port of the attached
# slaves in different ways. For example the "INFO replication" section
# offers this information, which is used, among other tools, by
# Redis Sentinel in order to discover slave instances.
# Another place where this info is available is in the output of the
# "ROLE" command of a masteer.
#
# The listed IP and address normally reported by a slave is obtained
# in the following way:
#
# IP: The address is auto detected by checking the peer address
# of the socket used by the slave to connect with the master.
#
# Port: The port is communicated by the slave during the replication
# handshake, and is normally the port that the slave is using to
# list for connections.
#
# However when port forwarding or Network Address Translation (NAT) is
# used, the slave may be actually reachable via different IP and port
# pairs. The following two options can be used by a slave in order to
# report to its master a specific set of IP and port, so that both INFO
# and ROLE will report those values.
#
# There is no need to use both the options if you need to override just
# the port or the IP address.
#
# slave-announce-ip 5.5.5.5
# slave-announce-port 1234

################################## SECURITY ###################################

# Require clients to issue AUTH <PASSWORD> before processing any other
# commands. This might be useful in environments in which you do not trust
# others with access to the host running redis-server.
#
# This should stay commented out for backward compatibility and because most
# people do not need auth (e.g. they run their own servers).
#
# Warning: since Redis is pretty fast an outside user can try up to
# 150k passwords per second against a good box. This means that you should
# use a very strong password otherwise it will be very easy to break.
#
# requirepass foobared

# Command renaming.
#
# It is possible to change the name of dangerous commands in a shared
# environment. For instance the CONFIG command may be renamed into something
# hard to guess so that it will still be available for internal-use tools
# but not available for general clients.
#
# Example:
#
# rename-command CONFIG b840fc02d524045429941cc15f59e41cb7be6c52
#
# It is also possible to completely kill a command by renaming it into
# an empty string:
#
# rename-command CONFIG ""
#
# Please note that changing the name of commands that are logged into the
# AOF file or transmitted to slaves may cause problems.

################################### LIMITS ####################################

# Set the max number of connected clients at the same time. By default
# this limit is set to 10000 clients, however if the Redis server is not
# able to configure the process file limit to allow for the specified limit
# the max number of allowed clients is set to the current file limit
# minus 32 (as Redis reserves a few file descriptors for internal uses).
#
# Once the limit is reached Redis will close all the new connections sending
# an error 'max number of clients reached'.
#
# maxclients 10000

# Don't use more memory than the specified amount of bytes.
# When the memory limit is reached Redis will try to remove keys
# according to the eviction policy selected (see maxmemory-policy).
#
# If Redis can't remove keys according to the policy, or if the policy is
# set to 'noeviction', Redis will start to reply with errors to commands
# that would use more memory, like SET, LPUSH, and so on, and will continue
# to reply to read-only commands like GET.
#
# This option is usually useful when using Redis as an LRU cache, or to set
# a hard memory limit for an instance (using the 'noeviction' policy).
#
# WARNING: If you have slaves attached to an instance with maxmemory on,
# the size of the output buffers needed to feed the slaves are subtracted
# from the used memory count, so that network problems / resyncs will
# not trigger a loop where keys are evicted, and in turn the output
# buffer of slaves is full with DELs of keys evicted triggering the deletion
# of more keys, and so forth until the database is completely emptied.
#
# In short... if you have slaves attached it is suggested that you set a lower
# limit for maxmemory so that there is some free RAM on the system for slave
# output buffers (but this is not needed if the policy is 'noeviction').
#
# maxmemory <bytes>

# MAXMEMORY POLICY: how Redis will select what to remove when maxmemory
# is reached. You can select among five behaviors:
#
# volatile-lru -> remove the key with an expire set using an LRU algorithm
# allkeys-lru -> remove any key according to the LRU algorithm
# volatile-random -> remove a random key with an expire set
# allkeys-random -> remove a random key, any key
# volatile-ttl -> remove the key with the nearest expire time (minor TTL)
# noeviction -> don't expire at all, just return an error on write operations
#
# Note: with any of the above policies, Redis will return an error on write
# operations, when there are no suitable keys for eviction.
#
# At the date of writing these commands are: set setnx setex append
# incr decr rpush lpush rpushx lpushx linsert lset rpoplpush sadd
# sinter sinterstore sunion sunionstore sdiff sdiffstore zadd zincrby
# zunionstore zinterstore hset hsetnx hmset hincrby incrby decrby
# getset mset msetnx exec sort
#
# The default is:
#
# maxmemory-policy noeviction

# LRU and minimal TTL algorithms are not precise algorithms but approximated
# algorithms (in order to save memory), so you can tune it for speed or
# accuracy. For default Redis will check five keys and pick the one that was
# used less recently, you can change the sample size using the following
# configuration directive.
#
# The default of 5 produces good enough results. 10 Approximates very closely
# true LRU but costs a bit more CPU. 3 is very fast but not very accurate.
#
# maxmemory-samples 5

############################## APPEND ONLY MODE ###############################

# By default Redis asynchronously dumps the dataset on disk. This mode is
# good enough in many applications, but an issue with the Redis process or
# a power outage may result into a few minutes of writes lost (depending on
# the configured save points).
#
# The Append Only File is an alternative persistence mode that provides
# much better durability. For instance using the default data fsync policy
# (see later in the config file) Redis can lose just one second of writes in a
# dramatic event like a server power outage, or a single write if something
# wrong with the Redis process itself happens, but the operating system is
# still running correctly.
#
# AOF and RDB persistence can be enabled at the same time without problems.
# If the AOF is enabled on startup Redis will load the AOF, that is the file
# with the better durability guarantees.
#
# Please check http://redis.io/topics/persistence for more information.

appendonly no

# The name of the append only file (default: "appendonly.aof")

appendfilename "appendonly.aof"

# The fsync() call tells the Operating System to actually write data on disk
# instead of waiting for more data in the output buffer. Some OS will really flush
# data on disk, some other OS will just try to do it ASAP.
#
# Redis supports three different modes:
#
# no: don't fsync, just let the OS flush the data when it wants. Faster.
# always: fsync after every write to the append only log. Slow, Safest.
# everysec: fsync only one time every second. Compromise.
#
# The default is "everysec", as that's usually the right compromise between
# speed and data safety. It's up to you to understand if you can relax this to
# "no" that will let the operating system flush the output buffer when
# it wants, for better performances (but if you can live with the idea of
# some data loss consider the default persistence mode that's snapshotting),
# or on the contrary, use "always" that's very slow but a bit safer than
# everysec.
#
# More details please check the following article:
# http://antirez.com/post/redis-persistence-demystified.html
#
# If unsure, use "everysec".

# appendfsync always
appendfsync everysec
# appendfsync no

# When the AOF fsync policy is set to always or everysec, and a background
# saving process (a background save or AOF log background rewriting) is
# performing a lot of I/O against the disk, in some Linux configurations
# Redis may block too long on the fsync() call. Note that there is no fix for
# this currently, as even performing fsync in a different thread will block
# our synchronous write(2) call.
#
# In order to mitigate this problem it's possible to use the following option
# that will prevent fsync() from being called in the main process while a
# BGSAVE or BGREWRITEAOF is in progress.
#
# This means that while another child is saving, the durability of Redis is
# the same as "appendfsync none". In practical terms, this means that it is
# possible to lose up to 30 seconds of log in the worst scenario (with the
# default Linux settings).
#
# If you have latency problems turn this to "yes". Otherwise leave it as
# "no" that is the safest pick from the point of view of durability.

no-appendfsync-on-rewrite no

# Automatic rewrite of the append only file.
# Redis is able to automatically rewrite the log file implicitly calling
# BGREWRITEAOF when the AOF log size grows by the specified percentage.
#
# This is how it works: Redis remembers the size of the AOF file after the
# latest rewrite (if no rewrite has happened since the restart, the size of
# the AOF at startup is used).
#
# This base size is compared to the current size. If the current size is
# bigger than the specified percentage, the rewrite is triggered. Also
# you need to specify a minimal size for the AOF file to be rewritten, this
# is useful to avoid rewriting the AOF file even if the percentage increase
# is reached but it is still pretty small.
#
# Specify a percentage of zero in order to disable the automatic AOF
# rewrite feature.

auto-aof-rewrite-percentage 100
auto-aof-rewrite-min-size 64mb

# An AOF file may be found to be truncated at the end during the Redis
# startup process, when the AOF data gets loaded back into memory.
# This may happen when the system where Redis is running
# crashes, especially when an ext4 filesystem is mounted without the
# data=ordered option (however this can't happen when Redis itself
# crashes or aborts but the operating system still works correctly).
#
# Redis can either exit with an error when this happens, or load as much
# data as possible (the default now) and start if the AOF file is found
# to be truncated at the end. The following option controls this behavior.
#
# If aof-load-truncated is set to yes, a truncated AOF file is loaded and
# the Redis server starts emitting a log to inform the user of the event.
# Otherwise if the option is set to no, the server aborts with an error
# and refuses to start. When the option is set to no, the user requires
# to fix the AOF file using the "redis-check-aof" utility before to restart
# the server.
#
# Note that if the AOF file will be found to be corrupted in the middle
# the server will still exit with an error. This option only applies when
# Redis will try to read more data from the AOF file but not enough bytes
# will be found.
aof-load-truncated yes

################################ LUA SCRIPTING ###############################

# Max execution time of a Lua script in milliseconds.
#
# If the maximum execution time is reached Redis will log that a script is
# still in execution after the maximum allowed time and will start to
# reply to queries with an error.
#
# When a long running script exceeds the maximum execution time only the
# SCRIPT KILL and SHUTDOWN NOSAVE commands are available. The first can be
# used to stop a script that did not yet called write commands. The second
# is the only way to shut down the server in the case a write command was
# already issued by the script but the user doesn't want to wait for the natural
# termination of the script.
#
# Set it to 0 or a negative value for unlimited execution without warnings.
lua-time-limit 5000

################################ REDIS CLUSTER ###############################
#
# ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
# WARNING EXPERIMENTAL: Redis Cluster is considered to be stable code, however
# in order to mark it as "mature" we need to wait for a non trivial percentage
# of users to deploy it in production.
# ++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++
#
# Normal Redis instances can't be part of a Redis Cluster; only nodes that are
# started as cluster nodes can. In order to start a Redis instance as a
# cluster node enable the cluster support uncommenting the following:
#
# cluster-enabled yes

# Every cluster node has a cluster configuration file. This file is not
# intended to be edited by hand. It is created and updated by Redis nodes.
# Every Redis Cluster node requires a different cluster configuration file.
# Make sure that instances running in the same system do not have
# overlapping cluster configuration file names.
#
# cluster-config-file nodes-6379.conf

# Cluster node timeout is the amount of milliseconds a node must be unreachable
# for it to be considered in failure state.
# Most other internal time limits are multiple of the node timeout.
#
# cluster-node-timeout 15000

# A slave of a failing master will avoid to start a failover if its data
# looks too old.
#
# There is no simple way for a slave to actually have a exact measure of
# its "data age", so the following two checks are performed:
#
# 1) If there are multiple slaves able to failover, they exchange messages
# in order to try to give an advantage to the slave with the best
# replication offset (more data from the master processed).
# Slaves will try to get their rank by offset, and apply to the start
# of the failover a delay proportional to their rank.
#
# 2) Every single slave computes the time of the last interaction with
# its master. This can be the last ping or command received (if the master
# is still in the "connected" state), or the time that elapsed since the
# disconnection with the master (if the replication link is currently down).
# If the last interaction is too old, the slave will not try to failover
# at all.
#
# The point "2" can be tuned by user. Specifically a slave will not perform
# the failover if, since the last interaction with the master, the time
# elapsed is greater than:
#
# (node-timeout * slave-validity-factor) + repl-ping-slave-period
#
# So for example if node-timeout is 30 seconds, and the slave-validity-factor
# is 10, and assuming a default repl-ping-slave-period of 10 seconds, the
# slave will not try to failover if it was not able to talk with the master
# for longer than 310 seconds.
#
# A large slave-validity-factor may allow slaves with too old data to failover
# a master, while a too small value may prevent the cluster from being able to
# elect a slave at all.
#
# For maximum availability, it is possible to set the slave-validity-factor
# to a value of 0, which means, that slaves will always try to failover the
# master regardless of the last time they interacted with the master.
# (However they'll always try to apply a delay proportional to their
# offset rank).
#
# Zero is the only value able to guarantee that when all the partitions heal
# the cluster will always be able to continue.
#
# cluster-slave-validity-factor 10

# Cluster slaves are able to migrate to orphaned masters, that are masters
# that are left without working slaves. This improves the cluster ability
# to resist to failures as otherwise an orphaned master can't be failed over
# in case of failure if it has no working slaves.
#
# Slaves migrate to orphaned masters only if there are still at least a
# given number of other working slaves for their old master. This number
# is the "migration barrier". A migration barrier of 1 means that a slave
# will migrate only if there is at least 1 other working slave for its master
# and so forth. It usually reflects the number of slaves you want for every
# master in your cluster.
#
# Default is 1 (slaves migrate only if their masters remain with at least
# one slave). To disable migration just set it to a very large value.
# A value of 0 can be set but is useful only for debugging and dangerous
# in production.
#
# cluster-migration-barrier 1

# By default Redis Cluster nodes stop accepting queries if they detect there
# is at least an hash slot uncovered (no available node is serving it).
# This way if the cluster is partially down (for example a range of hash slots
# are no longer covered) all the cluster becomes, eventually, unavailable.
# It automatically returns available as soon as all the slots are covered again.
#
# However sometimes you want the subset of the cluster which is working,
# to continue to accept queries for the part of the key space that is still
# covered. In order to do so, just set the cluster-require-full-coverage
# option to no.
#
# cluster-require-full-coverage yes

# In order to setup your cluster make sure to read the documentation
# available at http://redis.io web site.

################################## SLOW LOG ###################################

# The Redis Slow Log is a system to log queries that exceeded a specified
# execution time. The execution time does not include the I/O operations
# like talking with the client, sending the reply and so forth,
# but just the time needed to actually execute the command (this is the only
# stage of command execution where the thread is blocked and can not serve
# other requests in the meantime).
#
# You can configure the slow log with two parameters: one tells Redis
# what is the execution time, in microseconds, to exceed in order for the
# command to get logged, and the other parameter is the length of the
# slow log. When a new command is logged the oldest one is removed from the
# queue of logged commands.

# The following time is expressed in microseconds, so 1000000 is equivalent
# to one second. Note that a negative number disables the slow log, while
# a value of zero forces the logging of every command.
slowlog-log-slower-than 10000

# There is no limit to this length. Just be aware that it will consume memory.
# You can reclaim memory used by the slow log with SLOWLOG RESET.
slowlog-max-len 128

################################ LATENCY MONITOR ##############################

# The Redis latency monitoring subsystem samples different operations
# at runtime in order to collect data related to possible sources of
# latency of a Redis instance.
#
# Via the LATENCY command this information is available to the user that can
# print graphs and obtain reports.
#
# The system only logs operations that were performed in a time equal or
# greater than the amount of milliseconds specified via the
# latency-monitor-threshold configuration directive. When its value is set
# to zero, the latency monitor is turned off.
#
# By default latency monitoring is disabled since it is mostly not needed
# if you don't have latency issues, and collecting data has a performance
# impact, that while very small, can be measured under big load. Latency
# monitoring can easily be enabled at runtime using the command
# "CONFIG SET latency-monitor-threshold <milliseconds>" if needed.
latency-monitor-threshold 0

############################# EVENT NOTIFICATION ##############################

# Redis can notify Pub/Sub clients about events happening in the key space.
# This feature is documented at http://redis.io/topics/notifications
#
# For instance if keyspace events notification is enabled, and a client
# performs a DEL operation on key "foo" stored in the Database 0, two
# messages will be published via Pub/Sub:
#
# PUBLISH __keyspace@0__:foo del
# PUBLISH __keyevent@0__:del foo
#
# It is possible to select the events that Redis will notify among a set
# of classes. Every class is identified by a single character:
#
# K Keyspace events, published with __keyspace@<db>__ prefix.
# E Keyevent events, published with __keyevent@<db>__ prefix.
# g Generic commands (non-type specific) like DEL, EXPIRE, RENAME, ...
# $ String commands
# l List commands
# s Set commands
# h Hash commands
# z Sorted set commands
# x Expired events (events generated every time a key expires)
# e Evicted events (events generated when a key is evicted for maxmemory)
# A Alias for g$lshzxe, so that the "AKE" string means all the events.
#
# The "notify-keyspace-events" takes as argument a string that is composed
# of zero or multiple characters. The empty string means that notifications
# are disabled.
#
# Example: to enable list and generic events, from the point of view of the
# event name, use:
#
# notify-keyspace-events Elg
#
# Example 2: to get the stream of the expired keys subscribing to channel
# name __keyevent@0__:expired use:
#
# notify-keyspace-events Ex
#
# By default all notifications are disabled because most users don't need
# this feature and the feature has some overhead. Note that if you don't
# specify at least one of K or E, no events will be delivered.
notify-keyspace-events ""

############################### ADVANCED CONFIG ###############################

# Hashes are encoded using a memory efficient data structure when they have a
# small number of entries, and the biggest entry does not exceed a given
# threshold. These thresholds can be configured using the following directives.
hash-max-ziplist-entries 512
hash-max-ziplist-value 64

# Lists are also encoded in a special way to save a lot of space.
# The number of entries allowed per internal list node can be specified
# as a fixed maximum size or a maximum number of elements.
# For a fixed maximum size, use -5 through -1, meaning:
# -5: max size: 64 Kb <-- not recommended for normal workloads
# -4: max size: 32 Kb <-- not recommended
# -3: max size: 16 Kb <-- probably not recommended
# -2: max size: 8 Kb <-- good
# -1: max size: 4 Kb <-- good
# Positive numbers mean store up to _exactly_ that number of elements
# per list node.
# The highest performing option is usually -2 (8 Kb size) or -1 (4 Kb size),
# but if your use case is unique, adjust the settings as necessary.
list-max-ziplist-size -2

# Lists may also be compressed.
# Compress depth is the number of quicklist ziplist nodes from *each* side of
# the list to *exclude* from compression. The head and tail of the list
# are always uncompressed for fast push/pop operations. Settings are:
# 0: disable all list compression
# 1: depth 1 means "don't start compressing until after 1 node into the list,
# going from either the head or tail"
# So: [head]->node->node->...->node->[tail]
# [head], [tail] will always be uncompressed; inner nodes will compress.
# 2: [head]->[next]->node->node->...->node->[prev]->[tail]
# 2 here means: don't compress head or head->next or tail->prev or tail,
# but compress all nodes between them.
# 3: [head]->[next]->[next]->node->node->...->node->[prev]->[prev]->[tail]
# etc.
list-compress-depth 0

# Sets have a special encoding in just one case: when a set is composed
# of just strings that happen to be integers in radix 10 in the range
# of 64 bit signed integers.
# The following configuration setting sets the limit in the size of the
# set in order to use this special memory saving encoding.
set-max-intset-entries 512

# Similarly to hashes and lists, sorted sets are also specially encoded in
# order to save a lot of space. This encoding is only used when the length and
# elements of a sorted set are below the following limits:
zset-max-ziplist-entries 128
zset-max-ziplist-value 64

# HyperLogLog sparse representation bytes limit. The limit includes the
# 16 bytes header. When an HyperLogLog using the sparse representation crosses
# this limit, it is converted into the dense representation.
#
# A value greater than 16000 is totally useless, since at that point the
# dense representation is more memory efficient.
#
# The suggested value is ~ 3000 in order to have the benefits of
# the space efficient encoding without slowing down too much PFADD,
# which is O(N) with the sparse encoding. The value can be raised to
# ~ 10000 when CPU is not a concern, but space is, and the data set is
# composed of many HyperLogLogs with cardinality in the 0 - 15000 range.
hll-sparse-max-bytes 3000

# Active rehashing uses 1 millisecond every 100 milliseconds of CPU time in
# order to help rehashing the main Redis hash table (the one mapping top-level
# keys to values). The hash table implementation Redis uses (see dict.c)
# performs a lazy rehashing: the more operation you run into a hash table
# that is rehashing, the more rehashing "steps" are performed, so if the
# server is idle the rehashing is never complete and some more memory is used
# by the hash table.
#
# The default is to use this millisecond 10 times every second in order to
# actively rehash the main dictionaries, freeing memory when possible.
#
# If unsure:
# use "activerehashing no" if you have hard latency requirements and it is
# not a good thing in your environment that Redis can reply from time to time
# to queries with 2 milliseconds delay.
#
# use "activerehashing yes" if you don't have such hard requirements but
# want to free memory asap when possible.
activerehashing yes

# The client output buffer limits can be used to force disconnection of clients
# that are not reading data from the server fast enough for some reason (a
# common reason is that a Pub/Sub client can't consume messages as fast as the
# publisher can produce them).
#
# The limit can be set differently for the three different classes of clients:
#
# normal -> normal clients including MONITOR clients
# slave -> slave clients
# pubsub -> clients subscribed to at least one pubsub channel or pattern
#
# The syntax of every client-output-buffer-limit directive is the following:
#
# client-output-buffer-limit <class> <hard limit> <soft limit> <soft seconds>
#
# A client is immediately disconnected once the hard limit is reached, or if
# the soft limit is reached and remains reached for the specified number of
# seconds (continuously).
# So for instance if the hard limit is 32 megabytes and the soft limit is
# 16 megabytes / 10 seconds, the client will get disconnected immediately
# if the size of the output buffers reach 32 megabytes, but will also get
# disconnected if the client reaches 16 megabytes and continuously overcomes
# the limit for 10 seconds.
#
# By default normal clients are not limited because they don't receive data
# without asking (in a push way), but just after a request, so only
# asynchronous clients may create a scenario where data is requested faster
# than it can read.
#
# Instead there is a default limit for pubsub and slave clients, since
# subscribers and slaves receive data in a push fashion.
#
# Both the hard or the soft limit can be disabled by setting them to zero.
client-output-buffer-limit normal 0 0 0
client-output-buffer-limit slave 256mb 64mb 60
client-output-buffer-limit pubsub 32mb 8mb 60

# Redis calls an internal function to perform many background tasks, like
# closing connections of clients in timeout, purging expired keys that are
# never requested, and so forth.
#
# Not all tasks are performed with the same frequency, but Redis checks for
# tasks to perform according to the specified "hz" value.
#
# By default "hz" is set to 10. Raising the value will use more CPU when
# Redis is idle, but at the same time will make Redis more responsive when
# there are many keys expiring at the same time, and timeouts may be
# handled with more precision.
#
# The range is between 1 and 500, however a value over 100 is usually not
# a good idea. Most users should use the default of 10 and raise this up to
# 100 only in environments where very low latency is required.
hz 10

# When a child rewrites the AOF file, if the following option is enabled
# the file will be fsync-ed every 32 MB of data generated. This is useful
# in order to commit the file to the disk more incrementally and avoid
# big latency spikes.
aof-rewrite-incremental-fsync yes

wq保存第一個分片的配置文件

把第一個分片的配置文件復制兩份

第一份

?/golang/src/github.com/CodisLabs/codis/config/redis1.conf

把里面6379 都替換成6380

第二份

?/golang/src/github.com/CodisLabs/codis/config/redis2.conf

把里面6379 都替換成6381

exit ?退出docker

打鏡像

docker commit codis codis:3.2.1

打tag ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ??

docker tag codis:3.2.1 172.16.37.74:5000/codis:3.2.1

docker login?172.16.37.74:5000

172.16.37.74是我本地docker鏡像倉庫 ?輸入用戶名密碼

docker push?172.16.37.74:5000/codis:3.2.1

這樣鏡像就打好了

?然后 在本機嘗試拉起:

docker run ?-d -it ?--restart=always -v /etc/hosts:/etc/hosts -p 19000:19000 -p 6379:6379 -p 6380:6380 -p 11000:11000 -p 18080:18080 -p 11080:11080 -p 80:80 -p 8080:8080 -p 8090:8090 -p 9090:9090 -p 6381:6381 -p 6382:6382 -p 6383:6383 -p 6384:6384 --name=codis 172.16.37.74:5000/codis:3.2.1?/start.sh

不出意外應該能夠打開http://172.16.37.74:9090/

posted on 2016-11-21 21:49 雷獸 閱讀(...) 評論(...) 編輯 收藏

轉載于:https://www.cnblogs.com/sfissw/p/6087145.html

總結

以上是生活随笔為你收集整理的docker 安装redis第三方集群方案 codis的全部內容,希望文章能夠幫你解決所遇到的問題。

如果覺得生活随笔網站內容還不錯,歡迎將生活随笔推薦給好友。

亚洲精品视频一二三 | 中文字幕av电影下载 | 国产成人区 | 国产精品九九久久99视频 | 色综合天天综合 | 女人18毛片90分钟 | 99热高清 | 青草视频在线免费 | 特黄特色特刺激视频免费播放 | 午夜婷婷在线观看 | 免费av成人在线 | 久久99爱视频 | 国产亚洲免费的视频看 | 中文字幕资源站 | 欧美日韩午夜爽爽 | 亚洲一区精品二人人爽久久 | 久久久久久看片 | 免费三级av | 欧美另类调教 | 午夜精品久久 | 欧美激情视频一二区 | www.亚洲激情.com| 日本性生活一级片 | 青青草国产成人99久久 | av在线专区 | 九九九九精品九九九九 | 久久久久国产精品一区二区 | 在线观看小视频 | 国产999视频| 免费a一级 | 一二三精品视频 | 丁香花在线观看视频在线 | 五月天亚洲综合小说网 | 欧美在线99 | 日日干夜夜干 | 午夜精品福利一区二区三区蜜桃 | 黄色亚洲大片免费在线观看 | 国产久草在线观看 | 国产精品理论在线观看 | www.久艹| 激情婷婷在线观看 | 国产三级香港三韩国三级 | 国产精品久久久久久久久久尿 | 成人国产网址 | 一区二区三区免费在线观看视频 | 久草在线欧美 | 欧美精品你懂的 | 日本三级国产 | 精品免费一区二区三区 | 国产精品久久久久久婷婷天堂 | 视频成人 | 久久午夜色播影院免费高清 | 中文字幕在线观看第二页 | 欧美精品一区二区在线播放 | 狠狠躁夜夜躁人人爽超碰97香蕉 | 久草在线资源网 | 亚洲激色 | 91中文字幕| 综合激情婷婷 | 亚洲观看黄色网 | 91av免费观看 | 精品在线亚洲视频 | 亚洲综合一区二区精品导航 | 九九久久久久久久久激情 | 欧美99精品 | 国产精品一区二区免费在线观看 | 亚洲一级电影视频 | 午夜电影久久久 | 国产精品免费在线播放 | 亚洲高清av | 日韩精品不卡在线 | 天天色天天综合网 | 久久精品国产精品亚洲精品 | 国产亚洲精品中文字幕 | 欧美性极品xxxx做受 | 中文电影网| av解说在线观看 | 91av99 | 久久精品国亚洲 | 久久久久久久久久久免费视频 | 亚洲精品国产精品久久99 | 日韩免费高清在线观看 | 少妇超碰在线 | 国产成视频在线观看 | 成人av高清在线 | 一级电影免费在线观看 | 午夜av剧场 | 手机在线免费av | 国产综合视频在线观看 | 日韩中文字幕免费 | 国产资源中文字幕 | 色婷丁香 | 亚洲乱码中文字幕综合 | 国产在线视频一区二区三区 | av一区在线播放 | 久久精品国产第一区二区三区 | 91大神dom调教在线观看 | 久久成人精品 | 国产精品高潮在线观看 | 成年人免费看 | 日韩欧美不卡 | 色a网| 香蕉视频色 | 久久久午夜剧场 | 日韩成人免费观看 | 久久99这里只有精品 | 九九视频一区 | 亚洲少妇天堂 | 五月婷婷六月丁香 | 黄色免费大全 | 国产九色在线播放九色 | 中文字幕在线观看免费高清电影 | 激情五月婷婷综合 | 在线黄色av| 日韩中文字幕免费 | 国产精品国产自产拍高清av | 国产在线va | 91麻豆精品国产自产在线游戏 | 96视频免费在线观看 | 日本黄色特级片 | 天天插天天| 99热网站 | 免费观看视频黄 | a√天堂中文在线 | 久久久免费在线观看 | 欧美日韩国产精品一区 | 深夜福利视频一区二区 | 欧美性色黄大片在线观看 | av免费在线观看1 | 狠狠的干狠狠的操 | 国产区精品视频 | 国产精品久久久久久吹潮天美传媒 | 国产高清在线观看av | 黄色成人影视 | 色干干 | 日韩精品在线播放 | japanesefreesexvideo高潮 | 人人草网站 | 黄色小视频在线观看免费 | 婷婷午夜| 国产资源在线观看 | 国产视频一区二区在线 | 久久久久国产精品免费免费搜索 | 色婷婷激情四射 | 成人午夜精品福利免费 | 国产精品丝袜久久久久久久不卡 | 亚洲劲爆av| 天天操天天摸天天爽 | 久久成人精品电影 | 久久影视精品 | 国内外成人免费在线视频 | 9999在线| www.国产在线| 欧美做受高潮1 | 久久久久久久久久久高潮一区二区 | 视频高清 | 国产专区一 | 国产中文字幕国产 | 高清国产一区 | 日本激情中文字幕 | 欧美与欧洲交xxxx免费观看 | 黄色在线成人 | 在线观看黄a| 天天色天天 | 日韩欧美网址 | 狠狠躁日日躁狂躁夜夜躁 | 亚洲一区二区视频 | 亚洲国产精品一区二区久久hs | 丁香5月婷婷久久 | 亚洲欧美激情精品一区二区 | 久草91视频| 又黄又爽的免费高潮视频 | 在线视频1卡二卡三卡 | 欧美污在线观看 | 亚洲五月激情 | 少妇按摩av| 日本激情视频中文字幕 | 九九九九热精品免费视频点播观看 | 中文字幕一区二区三区在线视频 | 成人综合日日夜夜 | 国产黄色在线 | 狠狠综合久久av | 天天拍天天干 | 久久久久久久久久久国产精品 | 久久香蕉一区 | 91在线播放国产 | 片黄色毛片黄色毛片 | 黄网站色成年免费观看 | 国产成人精品在线 | 日韩久久久久久久久久 | 国产精品精品久久久 | 国产又粗又猛又色又黄网站 | 精品久久久久久久久久久久久久久久 | 欧美日韩一区二区免费在线观看 | 在线综合 亚洲 欧美在线视频 | 黄色1级毛片 | 在线观看日韩中文字幕 | 麻豆91精品视频 | 热re99久久精品国产66热 | 91中文视频| 99精品国产福利在线观看免费 | 亚洲综合精品在线 | 91在线免费观看国产 | 国产精品视频区 | 午夜色性片 | 久久久久久久久久久久久影院 | 亚洲精品视频第一页 | wwwwwww色| 亚洲精品国产精品国自产在线 | 天天久久夜夜 | 少妇av片| 国产伦理一区二区三区 | 日韩视频图片 | 这里只有精彩视频 | 久久精品伊人 | 国产成人精品综合 | 在线播放一区二区三区 | 在线免费观看黄色av | 97国产大学生情侣白嫩酒店 | 色av网站 | 久久久久中文字幕 | 久久综合久久综合九色 | 中文av字幕在线观看 | 91人人视频在线观看 | 人人澡av | 男女精品久久 | 美女视频又黄又免费 | 中文字幕在线不卡国产视频 | 色九九视频 | 免费在线色电影 | 91成人短视频在线观看 | 国产精品视频免费在线观看 | 亚洲va综合va国产va中文 | 久久久国产一区 | 国产在线观看免费观看 | 国产麻豆精品免费视频 | 免费国产亚洲视频 | 亚洲国产无| 亚洲精品国久久99热 | 日韩黄色一区 | 国产色视频网站 | 久草免费资源 | 5月丁香婷婷综合 | 九九精品久久 | 视频在线精品 | 日日夜夜精品网站 | 视频国产精品 | 欧美日韩精品在线免费观看 | 欧美性生交大片免网 | 久久久久久久久久毛片 | 99久久综合国产精品二区 | 九色自拍视频 | 久久99精品久久只有精品 | 亚洲一本视频 | 91探花国产综合在线精品 | 中文字幕在线观看av | 精品国自产在线观看 | 九九九九免费视频 | 中文字幕国语官网在线视频 | 亚洲成a人片77777kkkk1在线观看 | 久久久激情视频 | 精品久久亚洲 | 国产久草在线观看 | 久99久在线 | 成人午夜片av在线看 | 超碰在线亚洲 | 日韩在线二区 | 成年人三级网站 | 激情丁香5月 | 国产婷婷色 | 国产高清中文字幕 | 天天拍天天草 | 色窝资源| 国产三级香港三韩国三级 | 日韩大陆欧美高清视频区 | 51久久成人国产精品麻豆 | 婷婷精品 | 国产一区在线观看免费 | 国产中文字幕视频在线观看 | 在线电影日韩 | 91亚洲精品乱码久久久久久蜜桃 | 久久激情五月激情 | 国产一区二区高清视频 | 中文字幕免费高清 | 亚洲狠狠丁香婷婷综合久久久 | 成年人在线看片 | 久久综合精品国产一区二区三区 | 美女久久网站 | 又大又硬又黄又爽视频在线观看 | 亚洲精品视频在线观看视频 | 999视频网 | 九色91av| 日韩最新av | 久久久久久久久久久成人 | 亚洲最新合集 | 久9在线| 亚洲天天做 | 天天操天天色天天 | 免费在线一区二区 | 天天色天天色天天色 | 中文字幕免费在线看 | 欧美91av| 最近免费中文字幕大全高清10 | 免费网站色 | 91成人在线网站 | 在线观看视频在线 | 中文字幕精品一区二区三区电影 | 九九热视频在线播放 | 色综合久久88色综合天天 | 国产精品一区二区三区在线看 | 国产群p视频| 久久久片| 在线黄网站 | 国产精品一区二区美女视频免费看 | 手机看片1042| 高清中文字幕 | 午夜精品一区二区三区在线观看 | av在线免费网站 | 97超碰资源总站 | 久久一视频 | 成年人黄色免费网站 | 天天人人 | 黄色小网站在线观看 | 中文字幕一区在线观看视频 | 国产精品电影一区 | 久久成年人视频 | 成人国产电影在线观看 | 免费av福利 | 国产精品一区二区免费视频 | 日本在线观看一区 | 久久伊人精品一区二区三区 | 在线91网 | 欧美日韩国产在线精品 | 婷婷av网| 国产特黄色片 | 免费观看成人av | 国产精品久久久久免费观看 | 亚洲综合视频在线 | 日日爱视频 | 欧美性脚交 | 一区二区三区免费 | 麻豆精品传媒视频 | a级国产乱理伦片在线观看 亚洲3级 | 成年人免费在线观看网站 | 又黄又爽又无遮挡的视频 | av片子在线观看 | 精产嫩模国品一二三区 | 人人狠| 国产精品久久久久免费a∨ 欧美一级性生活片 | 国产成人中文字幕 | 欧美一区三区四区 | www.91国产| 免费看三级网站 | 女人18毛片90分钟 | 日韩av成人免费看 | 国产精品久久久久久久久久久久 | 亚洲综合婷婷 | 四虎免费av | 久久久久亚洲精品国产 | 日日干影院 | 欧洲激情综合 | 99久久毛片 | 国内精品在线观看视频 | 亚洲精选在线 | 免费高清国产 | 久久久影院一区二区三区 | 天天五月天色 | 国产vs久久 | 四虎国产精品免费 | 黄色综合| 国产成人高清 | 亚洲精品1234区 | 永久免费的啪啪网站免费观看浪潮 | 日韩动态视频 | 一区二区三区四区五区六区 | 超碰在线免费福利 | 亚洲涩涩涩涩涩涩 | 国产性天天综合网 | 91精品秘密在线观看 | 日本精品一区二区在线观看 | 99热官网 | av千婊在线免费观看 | 91视频首页 | 91精品综合在线观看 | 日韩成人免费电影 | 99精品免费久久久久久久久 | 国产a视频免费观看 | 久久久久欧美精品 | 丁香色天天 | 亚洲综合婷婷 | 久久91久久久久麻豆精品 | 国产一区二区精 | 在线免费中文字幕 | 人人网人人爽 | 亚州天堂 | 久久中文欧美 | 国产精品嫩草55av | 日批视频 | 又黄又爽又刺激 | 中文字幕91在线 | 日本精品久久久久影院 | 99高清视频有精品视频 | 黄色免费电影网站 | 免费在线激情电影 | 在线国产片 | 天天操操 | 午夜精品一区二区三区视频免费看 | 超碰免费观看 | 欧洲高潮三级做爰 | 久久精品爱爱视频 | 一区二区精品久久 | 精品国产一区二区三区四 | 99久久精品国产欧美主题曲 | 国产女教师精品久久av | 天天射网| 亚洲综合在线发布 | 黄色av高清 | 99视频在线免费播放 | 成人黄色片在线播放 | 欧美精品一区二区在线观看 | 国产日产精品一区二区三区四区的观看方式 | 日韩av一区二区三区在线观看 | 国产一级片免费观看 | 一级一片免费观看 | 精品久久久久一区二区国产 | 欧美激情精品久久久久久免费 | 中文字幕久久精品 | 人人爽人人av | 亚洲第一av在线 | 久久久久久网站 | 免费看一级特黄a大片 | 狠狠色狠狠色综合系列 | 狠狠色丁香婷婷综合 | 日韩动漫免费观看高清完整版在线观看 | 亚洲美女视频网 | 久久国产一二区 | av先锋影音少妇 | 午夜黄网| 一区二区 久久 | 337p西西人体大胆瓣开下部 | 精品嫩模福利一区二区蜜臀 | 一本一道波多野毛片中文在线 | 国产精品小视频网站 | 五月婷婷六月丁香在线观看 | 99久久超碰中文字幕伊人 | 日韩精品一区二区在线观看 | 精品特级毛片 | 黄在线| 欧美a级片网站 | 亚洲最大激情中文字幕 | 日韩欧美中文 | 成人av.com| 精品在线看 | 日韩av区 | 一级免费片 | 99在线观看视频网站 | 综合天堂av久久久久久久 | 黄色在线观看网站 | 国产欧美精品在线观看 | 日韩在线播放视频 | 99久久精品国产一区二区成人 | 91在线看免费 | 777xxx欧美 | 久久久免费观看视频 | 欧美亚洲免费在线一区 | 五月开心婷婷网 | 亚洲一级片 | 四虎成人精品在永久免费 | 91亚洲精品乱码久久久久久蜜桃 | a v在线视频 | 91成人免费观看视频 | 九九九视频精品 | 亚洲一区日韩在线 | 亚洲天堂视频在线 | 精品久久福利 | 亚洲永久字幕 | 国产在线不卡精品 | 粉嫩av一区二区三区四区 | 91精品亚洲影视在线观看 | 欧美夫妻性生活电影 | 狠狠躁夜夜躁人人爽超碰97香蕉 | 亚洲精品一区二区三区四区高清 | 国产九九在线 | 在线观看免费日韩 | .国产精品成人自产拍在线观看6 | 国产亚洲精品久久19p | 亚洲乱码在线观看 | 少妇bbbb| 激情丁香综合 | 日韩在线观看第一页 | 日本99久久 | 国产精品一区二区av日韩在线 | 五月婷婷一区二区三区 | 激情五月开心 | 中文字幕制服丝袜av久久 | 99热国产在线观看 | 国产视频二区三区 | 大荫蒂欧美视频另类xxxx | 久久久国产精品久久久 | 一区二区三区在线观看 | 日本久久综合网 | a黄色影院 | 免费视频三区 | 国产精品久久人 | 欧美日韩国产精品一区二区亚洲 | 国产视频一区在线免费观看 | 在线91观看 | 久久天天躁狠狠躁夜夜不卡公司 | 成年人免费在线观看网站 | 天天综合人人 | 亚洲色图色 | 欧美污污网站 | 日韩精品电影在线播放 | 免费在线观看成人av | 午夜999 | 国产午夜精品理论片在线 | 久久精品福利 | 狠狠干夜夜 | 午夜私人影院久久久久 | 97超碰成人在线 | 啪啪动态视频 | 日本中文字幕在线电影 | 99 色| 日本精品视频在线 | 在线观看中文字幕2021 | 日本久久不卡视频 | 欧美天天射 | 日本午夜免费福利视频 | 天天操天天射天天添 | 在线日韩中文字幕 | 国产免费又爽又刺激在线观看 | 91激情在线视频 | 久久er99热精品一区二区三区 | 欧美做受69| 色爱区综合激月婷婷 | 亚洲国产精品女人久久久 | 超碰97中文| 911国产 | 99免在线观看免费视频高清 | 国产一区二区三区网站 | 久久久久久久久网站 | 香蕉视频最新网址 | 午夜视频在线观看网站 | 日韩综合在线观看 | 日韩狠狠操| 久久久久久黄 | 五月天九九 | 日韩在线网址 | 国产精品福利小视频 | 青草视频在线 | 中文字幕第一页av | 欧美激情视频一二三区 | 欧美淫aaa免费观看 日韩激情免费视频 | 四虎成人精品永久免费av | 国产在线观 | 黄色小说视频网站 | 欧美激情综合色综合啪啪五月 | 美女黄网站视频免费 | 国产一区二区精品久久 | 玖玖爱国产在线 | 精品久久久一区二区 | 久久久亚洲网站 | 久久精品中文 | 91av在线电影| 日韩欧美一区二区三区在线 | 久久尤物电影视频在线观看 | 精品久久久999 | 天天视频亚洲 | 黄av免费在线观看 | 国产精品 日韩 | 视频91在线| 国产一区免费观看 | 久久久av电影 | 91av片 | 99精品视频一区二区 | 久久试看 | 亚洲激情一区二区三区 | 亚洲成人精品在线观看 | 玖玖爱免费视频 | 国产精品成人av在线 | 狠狠狠狠狠狠 | 亚洲黄色在线观看 | 在线免费色 | 97超级碰碰碰视频在线观看 | 天天干天天操天天入 | 中文字幕免费在线 | 国产高清无线码2021 | 国产精品区免费视频 | 天天搞夜夜骑 | 国产美女精品人人做人人爽 | 99精品免费 | 成年人黄色大片在线 | 婷婷综合视频 | 国产中文字幕一区二区三区 | 亚洲三级黄色 | 99草视频 | 国产三级精品在线 | 日韩成人黄色av | 成人在线播放免费观看 | 精品亚洲视频在线观看 | 成片免费观看视频大全 | 国内精品美女在线观看 | 国产香蕉视频在线播放 | 91久久一区二区 | 日本成址在线观看 | 中文字幕不卡在线88 | 欧美日韩高清一区二区 国产亚洲免费看 | 国产亚洲综合性久久久影院 | 18国产精品白浆在线观看免费 | 日韩欧美网址 | 成年人天堂com | 成人午夜剧场在线观看 | 亚洲欧洲精品久久 | 亚洲蜜桃在线 | 在线观看成人国产 | 色干综合 | av在线播放快速免费阴 | 欧美另类调教 | 91视频在线播放视频 | 欧美一区二区免费在线观看 | 91系列在线| 久久97精品 | 国产精品久久亚洲 | 国产精品久久久久久欧美 | 美女久久久久久久 | 日韩av成人在线观看 | 亚洲人精品午夜 | 日本中文一级片 | 日韩字幕在线观看 | 国内视频一区二区 | 久久久国产精品视频 | 久久国产精品久久久久 | 日日操狠狠干 | 美女视频黄在线观看 | 97精品国产97久久久久久免费 | 亚洲在线a | 天天操天天爱天天爽 | 女人18精品一区二区三区 | 91免费黄视频 | 色丁香色婷婷 | 黄色片视频免费 | 亚洲精选国产 | 国产精品视频久久久 | 久艹视频免费观看 | 欧美一级日韩三级 | 亚洲成a人片77777潘金莲 | 西西大胆免费视频 | 在线亚洲午夜片av大片 | 亚洲区另类春色综合小说 | 欧美日韩高清免费 | 欧美福利片在线观看 | 成人视屏免费看 | 日韩免费二区 | 女女av在线| 欧美亚洲国产日韩 | 成人黄大片视频在线观看 | 成人中文字幕av | 国产精品成人一区 | 日韩二区在线播放 | 久久综合九色99 | 国产成人精品国内自产拍免费看 | 天天天天天天干 | 免费看污在线观看 | 国产99爱| 久久久久 免费视频 | 毛片网在线观看 | 国产黄色一级片在线 | 丁香婷婷自拍 | 国内综合精品午夜久久资源 | 美女搞黄国产视频网站 | 久久精品99国产精品日本 | 精品福利视频在线 | 久久免费视频在线观看 | 少妇视频在线播放 | 欧美精品乱码99久久影院 | av在线最新| 97超碰在线资源 | 中文字幕网址 | 久久精品视频在线免费观看 | 亚洲最大成人免费网站 | 欧美日韩一区二区在线观看 | 欧美人牲| 久久久国产一区二区三区四区小说 | 日韩免费高清在线观看 | 在线观看视频一区二区三区 | 国产在线观看地址 | 毛片播放网站 | 一区二区视频在线观看免费 | 国产日本在线 | 亚洲精品乱码久久久久久写真 | 国产精品免费成人 | 成人av免费 | 狠狠狠狠狠干 | 国产一区二区高清视频 | 最近中文字幕完整高清 | 日韩av资源在线观看 | 2022久久国产露脸精品国产 | 亚洲国产美女精品久久久久∴ | 伊人久久婷婷 | 激情视频一区二区三区 | 欧美精品三级 | 97麻豆视频| 国产高清中文字幕 | 久久成人免费电影 | 欧美特一级片 | 日韩高清 一区 | 99视频国产在线 | 国产一级片在线播放 | 国产精品123 | 四虎成人免费影院 | 97人人艹| 中文字幕视频 | 成人av电影在线观看 | 99精品国产高清在线观看 | 男女激情免费网站 | 国产成人一二片 | 操操操日日| 欧美日韩在线免费观看 | 在线观看视频h | 天堂av网在线 | 人人玩人人添人人澡97 | 欧美日韩国产精品一区二区亚洲 | 午夜少妇一区二区三区 | 在线看黄色av | 人人爽人人爽人人爽人人爽 | 日日摸日日 | 久久久国产一区二区三区四区小说 | 中文字幕你懂的 | 亚洲另类在线视频 | 一本一道久久a久久综合蜜桃 | 亚洲精品字幕 | 色综合网在线 | 天天干,天天射,天天操,天天摸 | 精品女同一区二区三区在线观看 | 亚洲无吗av | 欧美日韩在线播放一区 | 成人在线视频你懂的 | 91av在线免费 | 国产精品美女久久久久久免费 | 在线免费观看的av | 国产日韩欧美在线看 | 日韩高清一区 | 久久综合9988久久爱 | 国产伦理一区二区三区 | 国产精品99久久久久人中文网介绍 | 午夜在线免费视频 | 日本丶国产丶欧美色综合 | 久久高清免费视频 | 国产一区二区在线观看视频 | 91在线看视频 | 超碰精品在线 | 最近免费中文视频 | 91尤物国产尤物福利在线播放 | 97人人模人人爽人人少妇 | 麻豆播放 | 欧美国产日韩一区二区三区 | 免费裸体视频网 | 亚洲涩涩涩涩涩涩 | 成人网在线免费视频 | 天天操天天干天天爽 | 久久激情日本aⅴ | 亚洲一区二区三区miaa149 | 国产精品视频全国免费观看 | 成人av午夜| 视频一区二区在线观看 | 97超碰在线久草超碰在线观看 | 狠狠狠色狠狠色综合 | 一级α片| 国产伦精品一区二区三区四区视频 | 丁香六月婷婷综合 | 一区二区三区动漫 | 97日日碰人人模人人澡分享吧 | 一区二区三区四区不卡 | 亚洲国内精品视频 | 日韩视频免费在线观看 | 久草爱| 日韩欧美高清在线观看 | 午夜av不卡 | 国产18精品乱码免费看 | 狠狠干五月天 | 国产精品午夜在线 | 久久免费在线观看 | 国产在线精品一区二区三区 | 国内精品久久久久影院一蜜桃 | 国内免费久久久久久久久久久 | 一区二区三区免费在线观看 | 色99之美女主播在线视频 | 欧美作爱视频 | 一区在线免费观看 | 国产精品久久久久久超碰 | 91在线免费观看网站 | 日韩精品字幕 | 日日干日日色 | 免费视频a| 国产精品资源网 | 免费国产一区二区 | 免费网站在线观看成人 | 国产97色 | 日韩欧美精品在线观看 | 久久久久女教师免费一区 | 久草电影在线 | 四虎影视欧美 | 日韩高清激情 | 91超碰在线播放 | www免费 | 免费三级黄色片 | 精品成人国产 | av噜噜噜在线播放 | 亚洲永久精品在线观看 | 欧美俄罗斯性视频 | 激情久久久久久久久久久久久久久久 | 人人射人人爱 | 国产一卡二卡四卡国 | .国产精品成人自产拍在线观看6 | 国产三级国产精品国产专区50 | 国产日韩在线一区 | 99国产精品视频免费观看一公开 | www.伊人色.com | 久久精品免费观看 | 激情偷乱人伦小说视频在线观看 | 日韩国产欧美在线视频 | 久久99精品久久久久婷婷 | 久久视频这里有久久精品视频11 | 欧美一级黄色片 | 视频在线观看99 | 国产一级二级三级在线观看 | 日韩久久久久久久久久 | 天天操天天插 | av大全免费在线观看 | 婷婷综合久久 | 国产亚洲视频在线 | 99久久这里有精品 | 欧美日韩不卡一区 | 久久精品99 | 国产精品欧美久久久久天天影视 | 涩涩伊人| 69国产精品成人在线播放 | 欧美国产日韩一区二区三区 | 日韩中文字幕国产 | 激情五月播播久久久精品 | 丁香婷婷射| 天天做天天爽 | 国产拍揄自揄精品视频麻豆 | 中文字幕日韩一区二区三区不卡 | 日韩午夜在线播放 | 精品一区二区免费视频 | 奇米网444 | 国产99在线播放 | 久操久| www.超碰97.com | 天天操天天舔天天爽 | 亚洲精品动漫成人3d无尽在线 | 九九九视频精品 | 国产小视频福利在线 | 在线看片日韩 | 日韩色区 | 狠狠狠色丁香婷婷综合久久五月 | 2000xxx影视 | 奇米影视777影音先锋 | 欧美福利视频一区 | 免费观看丰满少妇做爰 | 色中文字幕在线观看 | 久久综合久久综合这里只有精品 | 色婷婷综合久久久久中文字幕1 | 免费电影一区二区三区 | 欧美日韩国产一区二 | 色综合 久久精品 | 国产精品岛国久久久久久久久红粉 | 日韩午夜av | 国产视频在线观看免费 | 日韩黄色免费电影 | 免费观看性生交 | 伊人国产在线播放 | 三级av在线免费观看 | 欧美成人一区二区 | 国产精品美女视频 | 亚洲女人天堂成人av在线 | 亚洲精品视频二区 | 天天操比 | 日韩免费电影一区二区三区 | 在线视频观看你懂的 | 99久久精品免费看国产麻豆 | 91看片在线观看 | 亚洲成av人片 | 久久久91精品国产一区二区精品 | 91麻豆精品国产91久久久更新时间 | 午夜国产一区二区三区四区 | 国产精品久久久久久久久久东京 | 国产三级香港三韩国三级 | 天天色中文| 美女视频黄网站 | 国产精品久久久久aaaa | 国产专区精品视频 | 天堂在线v| 在线免费色| 中文字幕乱偷在线 | 日日弄天天弄美女bbbb | 六月激情婷婷 | 久久伦理影院 | 最新影院 | 国产91小视频 | 国产美女久久 | www.黄色在线| 久久精品www人人爽人人 | 精产嫩模国品一二三区 | 欧美黄色软件 | 欧美日本在线观看视频 | 欧美一区二区免费在线观看 | 国产亚洲精品久久久久久无几年桃 | 国产欧美综合视频 | 麻豆影视在线免费观看 | 日韩一区二区三 | 在线视频99 | 视频在线观看91 | 久久久久久蜜桃一区二区 | 五月亚洲婷婷 | 欧美精品一区二区免费 | 91porny九色在线播放 | 午夜视频免费播放 | 免费精品国产va自在自线 | 999一区二区三区 | 免费色视频在线 | 久久公开视频 | 精品五月天 | 欧美日韩免费观看一区=区三区 | 在线视频福利 | 九九九九热精品免费视频点播观看 | 国产破处在线播放 | 日本在线观看一区二区三区 | 精品久久一区二区三区 | 成人av中文字幕在线观看 | 三级av在线免费观看 | 999电影免费在线观看 | 日本一区二区三区免费看 | 色五月成人 | 国产伦理剧 | 精品欧美小视频在线观看 | 精品国产一区二 | 亚洲国产理论片 | 丁香六月色| 国产91精品欧美 | 免费在线观看黄色网 | 中文字幕久久亚洲 | 亚洲欧美日韩国产一区二区三区 | 视频成人永久免费视频 | 黄网站a | 免费观看的av网站 | 日韩免费av网址 | 久久99国产精品久久 | 久久人人爽人人人人片 | 日本美女xx| 亚洲色图色| 天天草天天草 | 国产va在线 | 日韩色中色 | 中文字幕av免费在线观看 | 五月激情站| 欧美91片| 激情婷婷久久 | 亚洲不卡av一区二区三区 | 涩涩伊人| 一 级 黄 色 片免费看的 | 国产精品网红直播 | 日韩av三区 | 免费视频三区 | av电影免费 | 亚洲激情久久 | 1区2区视频| 亚洲精品乱码久久 | 在线亚洲人成电影网站色www | 国产精品久久网站 | 97在线免费视频观看 | 手机在线观看国产精品 | 亚洲精品免费在线观看 | 99免在线观看免费视频高清 | 玖玖视频精品 | 在线观看韩国av | 亚洲va综合va国产va中文 | 国产福利在线免费观看 | 免费观看一级特黄欧美大片 | 在线免费黄色毛片 | 久久人人爽人人爽 | 午夜精品久久久久久久久久久久 | 国产精品国内免费一区二区三区 | 国产精品激情偷乱一区二区∴ | 在线亚洲播放 | 久久久久国产精品午夜一区 | 区一区二在线 | 成人一区二区三区在线 | 在线观看视频h | 久草免费在线观看视频 | 婷婷亚洲综合五月天小说 | 视频在线精品 | 色噜噜色噜噜 | 欧美日韩亚洲精品在线 | 九九热视频在线播放 | 中文字幕在线观看免费高清完整版 | 人人爱夜夜操 |