diff --git a/doc/INSTALL.md b/doc/INSTALL.md index 4596604..7354c7b 100644 --- a/doc/INSTALL.md +++ b/doc/INSTALL.md @@ -1,3 +1,6 @@ +# 一、硬部署 +无条件,可直接硬部署MYSQL与REDIS,即可使用项目。 + ## 01、安装MYSQL **一**、下载并安装mysql: @@ -45,9 +48,67 @@ flush privileges; exit ``` -**七**、在云服务上增加MySQL的端口 +**七**、在云服务上增加MySQL的端口(打开防火墙对应端口) + +## 02、安装REDIS +**一**、安装redis: + +``` +yum -y update +yum -y install redis +``` + +**二**、修改配置文件 + +``` +vi /etc/redis.conf +``` + +``` +protected-mode no +port 6379 +timeout 0 +save 900 1 +save 300 10 +save 60 10000 +rdbcompression yes +dbfilename dump.rdb +appendonly yes +appendfsync everysec +requirepass austin +``` + +**三**、启动redis + +``` +systemctl start redis +service redis start +``` + +**四**、检查redis状态 + +``` +sudo systemctl status redis +``` + +**五**、连接redis + +``` +# 默认端口号6379 +redis-cli + +# 验证密码 +AUTH austin +``` + +**六**、在云服务上增加Redis的端口(打开防火墙对应端口) -## 02、安装DOCKER和DOCKER-COMPOSE +--- + +# 二、DOCKER-COMPOSE方式部署 +为方便管理与部署,可以选择DOCKER-COMPOSE方式部署组件,同理除了MYSQL与REDIS,其余组件都是**可选**。 + +## 01、安装DOCKER和DOCKER-COMPOSE 首先我们需要安装GCC相关的环境: @@ -121,17 +182,104 @@ sudo ln -s /usr/local/bin/docker-compose /usr/bin/docker-compose docker-compose --version ``` - (Austin项目的中间件使用docker进行部署,文件内容可以参考项目中`docker`文件夹) -## 03、安装KAFKA +## 02、安装MySql -新建搭建kafka环境的`docker-compose.yml`文件,内容如下: +`docker-compose.yaml`文件如下 + +```yaml +version: '3' +services: + mysql: + image: mysql:5.7 + container_name: mysql + restart: always + ports: + - 3306:3306 + volumes: + - mysql-data:/var/lib/mysql + environment: + MYSQL_ROOT_PASSWORD: root123_A + TZ: Asia/Shanghai + command: --character-set-server=utf8mb4 --collation-server=utf8mb4_unicode_ci +volumes: + mysql-data: +``` ``` +docker-compose up -d + +docker ps +``` + +部署后,初始化SQL为./doc/sql/austin.sql,其余SQL安装对应组件才需要 + +**安装文件详见./doc/docker/mysql目录** + +## 03、安装REDIS + +新建一个文件夹`redis`,然后在该目录下创建出`data`文件夹、`redis.conf`文件和`docker-compose.yaml`文件 + +`redis.conf`文件的内容如下(后面的配置可在这更改,比如requirepass 我指定的密码为`austin`) + +``` +protected-mode no +port 6379 +timeout 0 +save 900 1 +save 300 10 +save 60 10000 +rdbcompression yes +dbfilename dump.rdb +dir /data +appendonly yes +appendfsync everysec +requirepass austin + +``` + +`docker-compose.yaml`的文件内容如下: + +```yaml version: '3' services: - zookepper: + redis: + image: redis:3.2 + container_name: redis + restart: always + ports: + - 6379:6379 + volumes: + - ./redis.conf:/usr/local/etc/redis/redis.conf:rw + - ./data:/data:rw + command: + /bin/bash -c "redis-server /usr/local/etc/redis/redis.conf" +``` + +配置的工作就完了,如果是云服务器,记得开redis端口**6379** + +``` +docker-compose up -d + +docker ps + +docker exec -it redis redis-cli + +auth austin + +``` + +**安装文件详见./doc/docker/redis目录** + +## 04、安装KAFKA(可选) + +新建搭建kafka环境的`docker-compose.yml`文件,内容如下: + +```yaml +version: '3' +services: + zookeeper: image: wurstmeister/zookeeper # 原镜像`wurstmeister/zookeeper` container_name: zookeeper # 容器名为'zookeeper' volumes: # 数据卷挂载路径设置,将本机目录映射到容器目录 @@ -148,19 +296,19 @@ services: KAFKA_BROKER_ID: 0 # 在kafka集群中,每个kafka都有一个BROKER_ID来区分自己 KAFKA_ADVERTISED_LISTENERS: PLAINTEXT://ip:9092 # TODO 将kafka的地址端口注册给zookeeper KAFKA_LISTENERS: PLAINTEXT://0.0.0.0:9092 # 配置kafka的监听端口 - KAFKA_ZOOKEEPER_CONNECT: zookeeper:2181 + KAFKA_ZOOKEEPER_CONNECT: zookeeper:2181 KAFKA_CREATE_TOPICS: "hello_world" KAFKA_HEAP_OPTS: -Xmx1G -Xms256M ports: # 映射端口 - "9092:9092" depends_on: # 解决容器依赖启动先后问题 - - zookepper + - zookeeper kafka-manager: - image: sheepkiller/kafka-manager # 原镜像`sheepkiller/kafka-manager` + image: kafkamanager/kafka-manager # 原镜像`sheepkiller/kafka-manager` container_name: kafka-manager # 容器名为'kafka-manager' environment: # 设置环境变量,相当于docker run命令中的-e - ZK_HOSTS: zookeeper:2181 + ZK_HOSTS: zookeeper:2181 APPLICATION_SECRET: xxxxx KAFKA_MANAGER_AUTH_ENABLED: "true" # 开启kafka-manager权限校验 KAFKA_MANAGER_USERNAME: admin # 登陆账户 @@ -208,65 +356,55 @@ $KAFKA_HOME/bin/kafka-topics.sh --create --topic austinRecall --partitions 1 --z $KAFKA_HOME/bin/kafka-topics.sh --zookeeper zookeeper:2181 --describe --topic austinBusiness ``` -## 04、安装REDIS - -首先,我们新建一个文件夹`redis`,然后在该目录下创建出`data`文件夹、`redis.conf`文件和`docker-compose.yaml`文件 - -`redis.conf`文件的内容如下(后面的配置可在这更改,比如requirepass 我指定的密码为`austin`) - -``` -protected-mode no -port 6379 -timeout 0 -save 900 1 -save 300 10 -save 60 10000 -rdbcompression yes -dbfilename dump.rdb -dir /data -appendonly yes -appendfsync everysec -requirepass austin +**安装文件详见./doc/docker/kafka目录** -``` +## 05、安装APOLLO(可选) -`docker-compose.yaml`的文件内容如下: +```yaml +version: '2.1' -``` -version: '3' services: - redis: - image: redis:latest - container_name: redis - restart: always + apollo-quick-start: + image: nobodyiam/apollo-quick-start + container_name: apollo-quick-start + depends_on: + apollo-db: + condition: service_healthy ports: - - 6379:6379 + - "8080:8080" + - "8090:8090" + - "8070:8070" + links: + - apollo-db + + apollo-db: + image: mysql:5.7 + container_name: apollo-db + environment: + TZ: Asia/Shanghai + MYSQL_ALLOW_EMPTY_PASSWORD: 'yes' + healthcheck: + test: ["CMD", "mysqladmin" ,"ping", "-h", "localhost"] + interval: 5s + timeout: 1s + retries: 10 + depends_on: + - apollo-dbdata + ports: + - "13306:3306" volumes: - - ./redis.conf:/usr/local/etc/redis/redis.conf:rw - - ./data:/data:rw - command: - /bin/bash -c "redis-server /usr/local/etc/redis/redis.conf " -``` - -配置的工作就完了,如果是云服务器,记得开redis端口**6379** - -启动Redis跟之前安装Kafka的时候就差不多啦 + - ./sql:/docker-entrypoint-initdb.d + volumes_from: + - apollo-dbdata -``` -docker-compose up -d - -docker ps - -docker exec -it redis redis-cli - -auth austin + apollo-dbdata: + image: alpine:latest + container_name: apollo-dbdata + volumes: + - /var/lib/mysql ``` -## 05、安装APOLLO - -部署Apollo跟之前一样直接用`docker-compose`就完事了,在GitHub已经给出了对应的教程和`docker-compose.yml`以及相关的文件,直接复制粘贴就完事咯。 - **PS: Apollo 的docker配置文件可以参考:docker/apollo/文件夹, 简单来说,在 docker/apollo/docker-quick-start/文件夹下执行docker-compose up -d 执行即可.** 目录结构最好保持一致: @@ -295,12 +433,13 @@ apollo配置样例可看example/apollo.properties文件的内容 动态线程池样例配置可看 dynamic-tp-apollo-dtp.yml 文件的内容 +**安装文件详见./doc/docker/apollo目录** ## 06、安装PROMETHEUS和GRAFANA(可选) 存放`docker-compose.yml`的信息: -``` +```yaml version: '2' networks: @@ -368,20 +507,30 @@ services: 新建prometheus的配置文件`prometheus.yml` -``` +```yaml global: - scrape_interval: 15s - evaluation_interval: 15s + scrape_interval: 1s + evaluation_interval: 1s scrape_configs: - job_name: 'prometheus' - static_configs: - - targets: ['ip:9090'] + static_configs: # TODO ip地址自己填我有相同的端口,因为是有两台机器,你们可以干掉相同的端口 + - targets: ['ip:9090'] - job_name: 'cadvisor' static_configs: - - targets: ['ip:8899'] + - targets: ['ip:8899'] - job_name: 'node' static_configs: - - targets: ['ip:9100'] + - targets: ['ip:9100'] + - job_name: 'cadvisor2' + static_configs: + - targets: ['ip:8899'] + - job_name: 'node2' + static_configs: + - targets: ['ip:9100'] + - job_name: 'austin' + metrics_path: '/actuator/prometheus' + static_configs: + - targets: ['ip:8888'] ``` (**这里要注意端口,按自己配置的来,ip也要填写为自己的**) @@ -434,6 +583,8 @@ import后就能直接看到高大上的监控页面了: ![](https://p3-juejin.byteimg.com/tos-cn-i-k3u1fbpfcp/26f4d6d83f4a441d85cb0a396cd0543c~tplv-k3u1fbpfcp-zoom-1.image) +**安装文件详见./doc/docker/prometheus目录** + ## 07、安装GRAYLOG(可选)-分布式日志收集框架 `docker-compose.yml`文件内容: @@ -500,32 +651,32 @@ networks: 最后配置`austin.grayLogIp`的ip即可实现分布式日志收集 -## 08、XXL-JOB - -文档:[https://www.xuxueli.com/xxl-job/#2.1%20%E5%88%9D%E5%A7%8B%E5%8C%96%E2%80%9C%E8%B0%83%E5%BA%A6%E6%95%B0%E6%8D%AE%E5%BA%93%E2%80%9D](https://www.xuxueli.com/xxl-job/#2.1%20%E5%88%9D%E5%A7%8B%E5%8C%96%E2%80%9C%E8%B0%83%E5%BA%A6%E6%95%B0%E6%8D%AE%E5%BA%93%E2%80%9D) - -xxl-job的部署我这边其实是依赖官网的文档的,步骤可以简单总结为: +**安装文件详见./doc/docker/graylog目录** -**1**、把xxl-job的仓库拉下来 +## 08、XXL-JOB(可选) -**2**、执行`/xxl-job/doc/db/tables_xxl_job.sql`的脚本(创建对应的库、创建表以及插入测试数据记录) - -**3**、如果是**本地**启动「调度中心」则在`xxl-job-admin`的`application.properties`更改相应的数据库配置,改完启动即可 - -**4**、如果是**云服务**启动「调度中心」,则可以选择拉取`docker`镜像进行部署,我拉取的是`2.30`版本,随后执行以下命令即可: - -```shell -docker pull xuxueli/xxl-job-admin:2.3.0 - -docker run -e PARAMS="--spring.datasource.url=jdbc:mysql://ip:3306/xxl_job?useUnicode=true&characterEncoding=UTF-8&autoReconnect=true&useSSL=false&zeroDateTimeBehavior=convertToNull --spring.datasource.username=root --spring.datasource.password=password " -p 6767:8080 --name xxl-job-admin -d xuxueli/xxl-job-admin:2.3.0 +`docker-compose.yaml`文件如下 +```yaml +version: '3' +services: + austin-xxl-job: + image: xuxueli/xxl-job-admin:2.3.0 + container_name: xxl-job-admin + restart: always + ports: + - "6767:8080" + environment: + PARAMS: '--spring.datasource.url=jdbc:mysql://ip:3306/xxl_job?useUnicode=true&characterEncoding=UTF-8&autoReconnect=true&useSSL=false&zeroDateTimeBehavior=convertToNull --spring.datasource.username=root --spring.datasource.password=root123_A' + # TODO 添加MySql网络,并更改ip ``` - -**注意**:第二条命令的**ip**和**password**需要更改为自己的,并且,我开的是**6767**端口 +**注意**:**ip**和**password**需要更改为自己的,并且,我开的是**6767**端口 ![](https://p3-juejin.byteimg.com/tos-cn-i-k3u1fbpfcp/180eabb4945e475494f3803c69318755~tplv-k3u1fbpfcp-zoom-1.image) -## 09、Flink +**安装文件详见./doc/docker/xxljob目录** + +## 09、Flink(可选) 部署Flink也是直接上docker-compose就完事了,值得注意的是:我们在部署的时候需要在配置文件里**指定时区** @@ -535,7 +686,7 @@ docker-compose.yml配置内容如下: version: "2.2" services: jobmanager: - image: flink:latest + image: flink:1.16.1 ports: - "8081:8081" command: jobmanager @@ -547,7 +698,7 @@ services: - CONTAINER_TIMEZONE=Asia/Shanghai - TZ=Asia/Shanghai taskmanager: - image: flink:latest + image: flink:1.16.1 depends_on: - jobmanager command: taskmanager @@ -560,10 +711,11 @@ services: - CONTAINER_TIMEZONE=Asia/Shanghai - TZ=Asia/Shanghai ``` +**安装文件详见./doc/docker/flink目录** -## 10、HIVE +## 10、HIVE(可选) -部署Flink也是直接上docker-compose就完事了 +部署Hive也是直接上docker-compose就完事了 1、把仓库拉到自己的服务器上 @@ -648,7 +800,9 @@ docker-compose exec hive-server bash /opt/hive/bin/beeline -u jdbc:hive2://localhost:10000 ``` -## 11、FLINK和HIVE融合 +**安装文件详见./doc/docker/hive目录** + +## 11、FLINK和HIVE融合(可选) 实时流处理的flink用的是docker-compose进行部署,而与hive融合的flink我这边是正常的姿势安装(主要是涉及的环境很多,用docker-compose就相对没那么方便了) @@ -881,23 +1035,154 @@ $KAFKA_HOME/bin/kafka-console-producer.sh --topic austinTraceLog --broker-list {"state":"1","businessId":"2","ids":[1,2,3],"logTimestamp":"123123"} ``` -## 12、安装METABASE +## 12、安装METABASE(可选) -部署`Metabase`很简单,也是使用`docker`进行安装部署,就两行命令(后续我会将其加入到`docker-compose`里面)。 +```yaml +version: '3' -```shell -docker pull metabase/metabase:latest +services: + metabase: + image: metabase/metabase + container_name: metabase + ports: + - "5001:3000" + restart: always ``` -```shell -docker run -d -p 5001:3000 --name metabase metabase/metabase +**安装文件详见./doc/docker/metabase目录** + +## 13、安装单机nacos(可选) + +`docker-compose.yaml`文件如下 + +```yaml +version: "3" +services: + nacos1: + container_name: nacos-server + hostname: nacos-server + image: nacos/nacos-server:v2.1.0 + environment: + - MODE=standalone + - PREFER_HOST_MODE=hostname + - SPRING_DATASOURCE_PLATFORM=mysql + - MYSQL_SERVICE_HOST=ip # TODO ip需设置 + - MYSQL_SERVICE_PORT=3306 + - MYSQL_SERVICE_USER=root + - MYSQL_SERVICE_PASSWORD=root123_A + - MYSQL_SERVICE_DB_NAME=nacos_config + - JVM_XMS=128m + - JVM_XMX=128m + - JVM_XMN=128m + volumes: + - /home/nacos/single-logs/nacos-server:/home/nacos/logs + - /home/nacos/init.d:/home/nacos/init.d + ports: + - 8848:8848 + - 9848:9848 + - 9849:9849 + restart: on-failure +``` + +**安装文件详见./doc/docker/nacos目录** + +## 14、安装单机rabbitmq(可选) + +`docker-compose.yaml`文件如下 + +```yaml +version: '3' +services: + rabbitmq: + image: registry.cn-hangzhou.aliyuncs.com/zhengqing/rabbitmq:3.7.8-management # 原镜像`rabbitmq:3.7.8-management` 【 注:该版本包含了web控制页面 】 + container_name: rabbitmq # 容器名为'rabbitmq' + hostname: my-rabbit + restart: unless-stopped # 指定容器退出后的重启策略为始终重启,但是不考虑在Docker守护进程启动时就已经停止了的容器 + environment: # 设置环境变量,相当于docker run命令中的-e + TZ: Asia/Shanghai + LANG: en_US.UTF-8 + RABBITMQ_DEFAULT_VHOST: my_vhost # 主机名 + RABBITMQ_DEFAULT_USER: admin # 登录账号 + RABBITMQ_DEFAULT_PASS: admin # 登录密码 + volumes: # 数据卷挂载路径设置,将本机目录映射到容器目录 + - "./rabbitmq/data:/var/lib/rabbitmq" + ports: # 映射端口 + - "5672:5672" + - "15672:15672" + ``` -完了之后,我们就可以打开`5001`端口到`Metabase`的后台了。 +**安装文件详见./doc/docker/rabbitmq目录** + +## 15、安装单机rocketmq(可选) + +`docker-compose.yaml`文件如下 +```yaml +version: '3.5' +services: + # mq服务 + rocketmq_server: + image: foxiswho/rocketmq:server + container_name: rocketmq_server + ports: + - 9876:9876 + volumes: + - ./rocketmq/rocketmq_server/logs:/opt/logs + - ./rocketmq/rocketmq_server/store:/opt/store + networks: + rocketmq: + aliases: + - rocketmq_server + + # mq中间件 + rocketmq_broker: + image: foxiswho/rocketmq:broker + container_name: rocketmq_broker + ports: + - 10909:10909 + - 10911:10911 + volumes: + - ./rocketmq/rocketmq_broker/logs:/opt/logs + - ./rocketmq/rocketmq_broker/store:/opt/store + - ./rocketmq/rocketmq_broker/conf/broker.conf:/etc/rocketmq/broker.conf + environment: + NAMESRV_ADDR: "rocketmq_server:9876" + JAVA_OPTS: " -Duser.home=/opt" + JAVA_OPT_EXT: "-server -Xms128m -Xmx128m -Xmn128m" + command: mqbroker -c /etc/rocketmq/broker.conf + depends_on: + - rocketmq_server + networks: + rocketmq: + aliases: + - rocketmq_broker + + # mq可视化控制台 + rocketmq_console_ng: + image: styletang/rocketmq-console-ng + container_name: rocketmq_console_ng + ports: + - 9002:8080 + environment: + JAVA_OPTS: "-Drocketmq.namesrv.addr=rocketmq_server:9876 -Dcom.rocketmq.sendMessageWithVIPChannel=false" + depends_on: + - rocketmq_server + networks: + rocketmq: + aliases: + - rocketmq_console_ng + +#容器通信network +networks: + rocketmq: + name: rocketmq + driver: bridge +``` +**安装文件详见./doc/docker/rocketmq目录** -## 13、未完待续 +## 未完待续 安装更详细的过程以及整个文章系列的更新思路都在公众号**Java3y**连载哟! diff --git a/doc/docker/apollo/docker-quick-start/docker-compose.yml b/doc/docker/apollo/docker-quick-start/docker-compose.yml index 10a0a3f..77aa9ac 100644 --- a/doc/docker/apollo/docker-quick-start/docker-compose.yml +++ b/doc/docker/apollo/docker-quick-start/docker-compose.yml @@ -15,7 +15,7 @@ services: - apollo-db apollo-db: - image: mysql:8.0 + image: mysql:5.7 container_name: apollo-db environment: TZ: Asia/Shanghai diff --git a/doc/docker/metabase/docker-compose.yml b/doc/docker/metabase/docker-compose.yml new file mode 100644 index 0000000..6d45248 --- /dev/null +++ b/doc/docker/metabase/docker-compose.yml @@ -0,0 +1,9 @@ +version: '3' + +services: + metabase: + image: metabase/metabase + container_name: metabase + ports: + - "5001:3000" + restart: always \ No newline at end of file diff --git a/doc/docker/mysql/docker-compose.yml b/doc/docker/mysql/docker-compose.yml index f8f14bc..408e83b 100644 --- a/doc/docker/mysql/docker-compose.yml +++ b/doc/docker/mysql/docker-compose.yml @@ -10,7 +10,7 @@ services: volumes: - mysql-data:/var/lib/mysql environment: - MYSQL_ROOT_PASSWORD: austin_pwd+- + MYSQL_ROOT_PASSWORD: root123_A TZ: Asia/Shanghai command: --character-set-server=utf8mb4 --collation-server=utf8mb4_unicode_ci volumes: diff --git a/doc/docker/redis/docker-compose.yaml b/doc/docker/redis/docker-compose.yaml index 7360562..5c50a19 100644 --- a/doc/docker/redis/docker-compose.yaml +++ b/doc/docker/redis/docker-compose.yaml @@ -2,7 +2,7 @@ version: '3' services: redis: - image: redis:latest + image: redis:3.2 container_name: redis restart: always ports: diff --git a/doc/docker/xxljob/docker-compose.yaml b/doc/docker/xxljob/docker-compose.yaml new file mode 100644 index 0000000..48e82b4 --- /dev/null +++ b/doc/docker/xxljob/docker-compose.yaml @@ -0,0 +1,11 @@ +version: '3' +services: + austin-xxl-job: + image: xuxueli/xxl-job-admin:2.3.0 + container_name: xxl-job-admin + restart: always + ports: + - "6767:8080" + environment: + PARAMS: '--spring.datasource.url=jdbc:mysql://ip:3306/xxl_job?useUnicode=true&characterEncoding=UTF-8&autoReconnect=true&useSSL=false&zeroDateTimeBehavior=convertToNull --spring.datasource.username=root --spring.datasource.password=root123_A' + # TODO 添加MySql网络,并更改ip \ No newline at end of file