部署方式:可以看这篇:
helm-charthttps://heidaodageshiwo.github.io/helm-chart/
1.当前文章是基于上一篇文章写的:
k8s Seata1.5.1_hunheidaode的博客-CSDN博客
在上面添加了些许改动,只不过是用helm部署了,其余的操作还是跟上篇博文一样子的。
2.环境:k8s 、nfs环境 或是动态存储
nfs:
3.下载seta helm部署yaml文件:文件位于:不知道的可看上篇博文。
4.我只是改动了values.yaml与templates下的deployment.yaml
deployment.yaml 改动了最后几行:
有原来hostpath改为NFS存储:
完整配置文件:
- apiVersion: apps/v1
- kind: Deployment
- metadata:
- {{- if .Values.namespace }}
- namespace: {{ .Values.namespace }}
- {{- end}}
- name: {{ include "seata-server.name" . }}
- labels:
- {{ include "seata-server.labels" . | indent 4 }}
- spec:
- replicas: {{ .Values.replicaCount }}
- selector:
- matchLabels:
- app.kubernetes.io/name: {{ include "seata-server.name" . }}
- app.kubernetes.io/instance: {{ .Release.Name }}
- template:
- metadata:
- labels:
- app.kubernetes.io/name: {{ include "seata-server.name" . }}
- app.kubernetes.io/instance: {{ .Release.Name }}
- spec:
- containers:
- - name: {{ .Chart.Name }}
- image: "{{ .Values.image.repository }}:{{ .Values.image.tag }}"
- imagePullPolicy: {{ .Values.image.pullPolicy }}
- ports:
- - name: http
- containerPort: 8091
- protocol: TCP
- {{- if .Values.volume }}
- volumeMounts:
- {{- range .Values.volume }}
- - name: {{ .name }}
- mountPath: {{ .mountPath }}
- {{- end}}
- {{- end}}
- {{- if .Values.env }}
- env:
- {{- if .Values.env.seataIp }}
- - name: SEATA_IP
- value: {{ .Values.env.seataIp | quote }}
- {{- end }}
- {{- if .Values.env.seataPort }}
- - name: SEATA_PORT
- value: {{ .Values.env.seataPort | quote }}
- {{- end }}
- {{- if .Values.env.seataEnv }}
- - name: SEATA_ENV
- value: {{ .Values.env.seataEnv }}
- {{- end }}
- {{- if .Values.env.seataConfigName }}
- - name: SEATA_CONFIG_NAME
- value: {{ .Values.env.seataConfigName }}
- {{- end }}
- {{- if .Values.env.serverNode }}
- - name: SERVER_NODE
- value: {{ .Values.env.serverNode | quote }}
- {{- end }}
- {{- if .Values.env.storeMode }}
- - name: STORE_MODE
- value: {{ .Values.env.storeMode }}
- {{- end }}
- {{- end }}
- {{- if .Values.volume }}
- volumes:
- {{- range .Values.volume }}
- - name: {{ .name }}
- nfs:
- server: {{ .nfsServers }}
- path: {{ .nfsPath }}
- {{- end}}
- {{- end}}
只是改动的这块:
- {{- if .Values.volume }}
- volumes:
- {{- range .Values.volume }}
- - name: {{ .name }}
- nfs:
- server: {{ .nfsServers }}
- path: {{ .nfsPath }}
- {{- end}}
- {{- end}}
values.yaml 完整:
- replicaCount: 1
-
- namespace: default
-
- image:
- repository: seataio/seata-server
- tag: 1.5.1
- pullPolicy: IfNotPresent
-
- service:
- type: NodePort
- port: 30091
- nodePort: 30091
-
- env:
- seataPort: "8091"
- storeMode: "file"
- seataIp: "192.168.56.211"
- seataConfigName: "file:/root/seata-config/registry/registry.conf"
-
- volume:
- - name: seata-config
- mountPath: /seata-server/resources
- nfsServers: 192.168.56.211
- nfsPath: /data/k8s/resource
env的环境之前配置过但是没有起作用,可以不用写:
seataConfigName:后面的内容
把/seata-server/resources 下的文件挂载到
/data/k8s/resource下
把上篇博文的文件下的文件修改后上传到 /data/k8s/resource文件下:
application.yaml 也与上篇博文一样:
- server:
- port: 7091
-
- spring:
- application:
- name: seata-server
-
- logging:
- config: classpath:logback-spring.xml
- file:
- path: ${user.home}/logs/seata
- extend:
- logstash-appender:
- destination: 127.0.0.1:4560
- kafka-appender:
- bootstrap-servers: 127.0.0.1:9092
- topic: logback_to_logstash
- console:
- user:
- username: seata
- password: seata
-
- seata:
- config:
- # support: nacos 、 consul 、 apollo 、 zk 、 etcd3
- type: nacos
- nacos:
- server-addr: 192.168.56.211:30683
- namespace: public
- group: SEATA_GROUP
- username: nacos
- password: nacos
- ##if use MSE Nacos with auth, mutex with username/password attribute
- #access-key: ""
- #secret-key: ""
- data-id: seata.properties
- registry:
- # support: nacos 、 eureka 、 redis 、 zk 、 consul 、 etcd3 、 sofa
- type: nacos
- preferred-networks: 192.168.*
- nacos:
- application: seata-server
- server-addr: 192.168.56.211:30683
- group: SEATA_GROUP
- namespace: public
- cluster: default
- username: nacos
- password: nacos
- ##if use MSE Nacos with auth, mutex with username/password attribute
- #access-key: ""
- #secret-key: ""
-
- store:
- # support: file 、 db 、 redis
- mode: db
- session:
- mode: db
- lock:
- mode: db
- file:
- dir: sessionStore
- max-branch-session-size: 16384
- max-global-session-size: 512
- file-write-buffer-cache-size: 16384
- session-reload-read-size: 100
- flush-disk-mode: async
- db:
- datasource: druid
- db-type: mysql
- driver-class-name: com.mysql.jdbc.Driver
- url: jdbc:mysql://192.168.56.211:31306/seata?rewriteBatchedStatements=true
- user: root
- password: s00J8De852
- min-conn: 5
- max-conn: 100
- global-table: global_table
- branch-table: branch_table
- lock-table: lock_table
- distributed-lock-table: distributed_lock
- query-limit: 100
- max-wait: 5000
- security:
- secretKey: SeataSecretKey0c382ef121d778043159209298fd40bf3850a017
- tokenValidityInMilliseconds: 1800000
- ignore:
- urls: /,/**/*.css,/**/*.js,/**/*.html,/**/*.map,/**/*.svg,/**/*.png,/**/*.ico,/console-fe/public/**,/api/v1/auth/login
5.直接启动
cd /root/seata
helm install seata ./seata-server
6.查看
7.访问nacos 看到已经被注册到nacos中。
8.seata控制台可视化界面:
当看到控制台一直报错时,我下载下seata1.5.1的源码下来,自己重新打包打镜像,搞了一下午!说实话各种报错!
搞一下午就是起不来!也没有报错!索性不搞了!
9.可视化界面展示
helm sevice.yaml文件修改:
- apiVersion: v1
- kind: Service
- metadata:
- name: {{ include "seata-server.fullname" . }}
- labels:
- {{ include "seata-server.labels" . | indent 4 }}
- spec:
- type: {{ .Values.service.type }}
- ports:
- - port: {{ .Values.service.port }}
- targetPort: {{ .Values.service.port }}
- protocol: TCP
- name: http
- selector:
- app.kubernetes.io/name: {{ include "seata-server.name" . }}
- app.kubernetes.io/instance: {{ .Release.Name }}
把targetPort: http 改为:
targetPort: {{ .Values.service.port }}
暴露出来:
然后:values.yaml
- replicaCount: 1
-
- namespace: default
-
- image:
- repository: seataio/seata-server
- #repository: library/seata-server
- tag: 1.5.1
- pullPolicy: IfNotPresent
-
- service:
- type: NodePort
- port: 7091
-
- env:
- seataPort: "8091"
- storeMode: "file"
- seataIp: "192.168.56.211"
- seataConfigName: "xxx"
-
- volume:
- - name: seata-config
- mountPath: /seata-server/resources
- nfsServers: 192.168.56.211
- nfsPath: /data/k8s/resource
启动:
访问:ip:32542端口即可:
但是有错误 列表出不来:
我又在自己的本地启动:控制台还是包mysql的错误但是不影响
也注册到nacos中了:
本地访问也是报错的:
这个问题不知道咋回事!,日志太多,看不到具体报错!但是确实是部署起来了!
10.可视化界面问题补充(如果你也是出现了这样的错误,无非就是你也是使用k8s部署的nacos或者是使用helm部署的nacos,你就会出现上面我说的这种情况!怎么解决呢!)
本人已经解决:具体请看nacos issues
helm 部署 nacos2.1.0 代码不能访问,ui可以访问 · Issue #9075 · alibaba/nacos · GitHub
这是我从部署到访问nacos到注册的问题,大家可以参考(前提:你是使用k8s或者是helm部署的nacos,绝对会出现这个问题!)
11.最终解决方案:
本人已经试过好多个方法:
1.首先把mysql换成了5.7:
2。目前helm部署seata并且能提供对java的访问。这块我还没有解决。
比较欣慰的是 k8s部署seata,java可以访问。这个问题我解决了。
我目前这种方式在csdn你是找不到的。
有人说官方不是提供了部署文档么,端口都代理出来了。那你们就去用用试试,用代码去连接试试。能成功的话就不用看我写的这个了。不行的话那就往下看呗。
完整k8s部署 yaml:
- apiVersion: v1
- kind: Service
- metadata:
- name: seata-server
- namespace: default
- labels:
- k8s-app: seata-server
- spec:
- type: ClusterIP
- ports:
- - port: 8091
- targetPort: 8091
- protocol: TCP
- name: http
- selector:
- k8s-app: seata-server
- sessionAffinity: None
- ---
- apiVersion: apps/v1
- kind: Deployment
- metadata:
- name: seata-server
- namespace: default
- labels:
- k8s-app: seata-server
- spec:
- replicas: 1
- selector:
- matchLabels:
- k8s-app: seata-server
- template:
- metadata:
- labels:
- k8s-app: seata-server
- spec:
- containers:
- - name: seata-server
- image: docker.io/seataio/seata-server:1.5.1
- imagePullPolicy: IfNotPresent
- ports:
- - name: http
- containerPort: 8091
- protocol: TCP
- env:
- - name: SEATA_IP
- value: 192.168.56.211
- - name: SEATA_PORT
- value: "31113"
- volumeMounts:
- - name: seata-config
- mountPath: /seata-server/resources
- volumes:
- - name: seata-config
- hostPath:
- path: /root/resources
- ---
- apiVersion: v1
- kind: Service
- metadata:
- name: seata-server-testzhangqiang
- spec:
- selector:
- k8s-app: seata-server
- ports:
- - name: client-port1
- port: 7091
- protocol: TCP
- targetPort: 7091
- nodePort: 31005
- - name: bus-port1
- port: 31113
- protocol: TCP
- targetPort: 31113
- nodePort: 31113
- type: NodePort
- ---
主要改动是参照helm ENV传环境变量的方式进行部署。
为什么要代理31113这个端口,为什么env里面要写31113这个端口。
因为:
你重启项目的时候 项目是连接这个端口的。
如果你是windows启动的,那么你平常连接的端口是:8091 端口。
但是你是k8s部署的nacos,你本机有8091端口么。没有!所以代码他连接不上!
所以要把这个31113端口写进去,并代理出来,让本地代码能够访问这个端口。
你们代码还有配置需要有以下改动:
nacos .seata配置修改为:
查看注册信息:
env:
- name: SEATA_IP
value: 192.168.56.211
- name: SEATA_PORT
value: "31113"
这个192.168.56.211是虚拟机ip
检查注册地址与ip 一定要对应上:
代码:
我是严格按照官网依赖来的。nacos2.1 seata1.5.1(我使用mysql5.7数据库,mysql8的数据库搞了2周都没搞起来)
application.yml(feign配置要添加不然 会 read time out 新版本是这样配置,搞了好久)
- # 数据源
- spring:
- datasource:
- username: root
- password: 123456
- url: jdbc:mysql://192.168.56.213:31306/seata_order?characterEncoding=utf8&useSSL=false&serverTimezone=UTC
- driver-class-name: com.mysql.jdbc.Driver
- type: com.alibaba.druid.pool.DruidDataSource
-
- #初始化时运行sql脚本
- schema: classpath:sql/schema.sql
- initialization-mode: never
- application:
- name: alibaba-order-seata
- #设置mybatis
- mybatis:
- mapper-locations: classpath:com/xx/order/mapper/*Mapper.xml
- #config-location: classpath:mybatis-config.xml
- typeAliasesPackage: com.xx.order.pojo
- configuration:
- mapUnderscoreToCamelCase: true
- server:
- port: 8072
-
- feign:
- client:
- config:
- default:
- readTimeout: 10000
- connectTimeout: 10000
bootstrap.yml(nacos配置要放到这里面)
- # 数据源
- spring:
- cloud:
- nacos:
- server-addr: 192.168.56.211:31000
- discovery:
- server-addr: 192.168.56.211:31000
- password: nacos
- username: nacos
- enabled: true
- seata:
- tx-service-group: zq
- registry:
- nacos:
- group: SEATA_GROUP
- username: nacos
- password: nacos
- application: seata-server
- # server-addr: 192.168.56.211:31000
- server-addr: 192.168.56.211:31000
- cluster: default
- type: nacos
- config:
- type: nacos
- nacos:
- password: nacos
- username: nacos
- # server-addr: 192.168.56.211:31000
- server-addr: 192.168.56.211:31000
- group: SEATA_GROUP
- service:
- grouplist:
- default: 192.168.56.211:31113
- disable-global-transaction: false
- # vgroup-mapping:
- # default_tx_group: default
- vgroup-mapping:
- default_tx_group: zq
注意这个配置:
大功告成了!
你们可以去试试了!
本机启动日志:
-
- . ____ _ __ _ _
- /\\ / ___'_ __ _ _(_)_ __ __ _ \ \ \ \
- ( ( )\___ | '_ | '_| | '_ \/ _` | \ \ \ \
- \\/ ___)| |_)| | | | | || (_| | ) ) ) )
- ' |____| .__|_| |_|_| |_\__, | / / / /
- =========|_|==============|___/=/_/_/_/
- :: Spring Boot :: (v2.3.12.RELEASE)
- 2022-09-06 17:00:48.843 INFO 132 --- [ main] c.t.order.AlibabaOrderSeataApplication : No active profile set, falling back to default profiles: default
- 2022-09-06 17:00:49.895 INFO 132 --- [ main] o.s.cloud.context.scope.GenericScope : BeanFactory id=f0658dad-eccb-3a82-9f89-e373def31818
- 2022-09-06 17:00:49.901 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'io.seata.spring.boot.autoconfigure.SeataCoreAutoConfiguration' of type [io.seata.spring.boot.autoconfigure.SeataCoreAutoConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:49.902 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'springApplicationContextProvider' of type [io.seata.spring.boot.autoconfigure.provider.SpringApplicationContextProvider] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:49.903 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'io.seata.spring.boot.autoconfigure.SeataAutoConfiguration' of type [io.seata.spring.boot.autoconfigure.SeataAutoConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:49.958 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'failureHandler' of type [io.seata.tm.api.DefaultFailureHandlerImpl] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:49.974 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'springCloudAlibabaConfiguration' of type [io.seata.spring.boot.autoconfigure.properties.SpringCloudAlibabaConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:49.978 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'seataProperties' of type [io.seata.spring.boot.autoconfigure.properties.SeataProperties] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:49.981 INFO 132 --- [ main] i.s.s.b.a.SeataAutoConfiguration : Automatically configure Seata
- 2022-09-06 17:00:50.061 INFO 132 --- [ main] io.seata.config.ConfigurationFactory : load Configuration from :Spring Configuration
- 2022-09-06 17:00:50.076 INFO 132 --- [ main] i.seata.config.nacos.NacosConfiguration : Nacos check auth with userName/password.
- 2022-09-06 17:00:50.128 INFO 132 --- [ main] c.a.n.p.a.s.c.ClientAuthPluginManager : [ClientAuthPluginManager] Load ClientAuthService com.alibaba.nacos.client.auth.impl.NacosClientAuthServiceImpl success.
- 2022-09-06 17:00:50.128 INFO 132 --- [ main] c.a.n.p.a.s.c.ClientAuthPluginManager : [ClientAuthPluginManager] Load ClientAuthService com.alibaba.nacos.client.auth.ram.RamClientAuthServiceImpl success.
- 2022-09-06 17:00:53.548 INFO 132 --- [ main] i.s.s.a.GlobalTransactionScanner : Initializing Global Transaction Clients ...
- 2022-09-06 17:00:53.629 INFO 132 --- [ main] i.s.core.rpc.netty.NettyClientBootstrap : NettyClientBootstrap has started
- 2022-09-06 17:00:53.652 INFO 132 --- [ main] c.a.n.p.a.s.c.ClientAuthPluginManager : [ClientAuthPluginManager] Load ClientAuthService com.alibaba.nacos.client.auth.impl.NacosClientAuthServiceImpl success.
- 2022-09-06 17:00:53.652 INFO 132 --- [ main] c.a.n.p.a.s.c.ClientAuthPluginManager : [ClientAuthPluginManager] Load ClientAuthService com.alibaba.nacos.client.auth.ram.RamClientAuthServiceImpl success.
- 2022-09-06 17:00:53.987 INFO 132 --- [ main] i.s.c.r.netty.NettyClientChannelManager : will connect to 192.168.56.211:31113
- 2022-09-06 17:00:54.530 INFO 132 --- [ main] i.s.core.rpc.netty.NettyPoolableFactory : NettyPool create channel to transactionRole:TMROLE,address:192.168.56.211:31113,msg:< RegisterTMRequest{applicationId='alibaba-order-seata', transactionServiceGroup='zq'} >
- 2022-09-06 17:00:55.836 INFO 132 --- [ main] i.s.c.rpc.netty.TmNettyRemotingClient : register TM success. client version:1.5.1, server version:1.5.1,channel:[id: 0xccc3ed55, L:/192.168.56.1:56381 - R:/192.168.56.211:31113]
- 2022-09-06 17:00:55.845 INFO 132 --- [ main] i.s.core.rpc.netty.NettyPoolableFactory : register success, cost 69 ms, version:1.5.1,role:TMROLE,channel:[id: 0xccc3ed55, L:/192.168.56.1:56381 - R:/192.168.56.211:31113]
- 2022-09-06 17:00:55.846 INFO 132 --- [ main] i.s.s.a.GlobalTransactionScanner : Transaction Manager Client is initialized. applicationId[alibaba-order-seata] txServiceGroup[zq]
- 2022-09-06 17:00:55.864 INFO 132 --- [ main] io.seata.rm.datasource.AsyncWorker : Async Commit Buffer Limit: 10000
- 2022-09-06 17:00:55.865 INFO 132 --- [ main] i.s.rm.datasource.xa.ResourceManagerXA : ResourceManagerXA init ...
- 2022-09-06 17:00:55.874 INFO 132 --- [ main] i.s.core.rpc.netty.NettyClientBootstrap : NettyClientBootstrap has started
- 2022-09-06 17:00:55.874 INFO 132 --- [ main] i.s.s.a.GlobalTransactionScanner : Resource Manager is initialized. applicationId[alibaba-order-seata] txServiceGroup[zq]
- 2022-09-06 17:00:55.874 INFO 132 --- [ main] i.s.s.a.GlobalTransactionScanner : Global Transaction Clients are initialized.
- 2022-09-06 17:00:55.877 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'io.seata.spring.boot.autoconfigure.SeataDataSourceAutoConfiguration' of type [io.seata.spring.boot.autoconfigure.SeataDataSourceAutoConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:56.006 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'com.alibaba.cloud.seata.feign.SeataFeignClientAutoConfiguration$FeignBeanPostProcessorConfiguration' of type [com.alibaba.cloud.seata.feign.SeataFeignClientAutoConfiguration$FeignBeanPostProcessorConfiguration] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:56.010 INFO 132 --- [ main] trationDelegate$BeanPostProcessorChecker : Bean 'seataFeignObjectWrapper' of type [com.alibaba.cloud.seata.feign.SeataFeignObjectWrapper] is not eligible for getting processed by all BeanPostProcessors (for example: not eligible for auto-proxying)
- 2022-09-06 17:00:56.369 INFO 132 --- [ main] o.s.b.w.embedded.tomcat.TomcatWebServer : Tomcat initialized with port(s): 8072 (http)
- 2022-09-06 17:00:56.381 INFO 132 --- [ main] o.apache.catalina.core.StandardService : Starting service [Tomcat]
- 2022-09-06 17:00:56.381 INFO 132 --- [ main] org.apache.catalina.core.StandardEngine : Starting Servlet engine: [Apache Tomcat/9.0.46]
- 2022-09-06 17:00:56.612 INFO 132 --- [ main] o.a.c.c.C.[Tomcat].[localhost].[/] : Initializing Spring embedded WebApplicationContext
- 2022-09-06 17:00:56.612 INFO 132 --- [ main] w.s.c.ServletWebServerApplicationContext : Root WebApplicationContext: initialization completed in 7748 ms
- 2022-09-06 17:00:56.854 INFO 132 --- [ main] c.a.d.s.b.a.DruidDataSourceAutoConfigure : Init DruidDataSource
- Loading class `com.mysql.jdbc.Driver'. This is deprecated. The new driver class is `com.mysql.cj.jdbc.Driver'. The driver is automatically registered via the SPI and manual loading of the driver class is generally unnecessary.
- 2022-09-06 17:00:56.991 INFO 132 --- [ main] com.alibaba.druid.pool.DruidDataSource : {dataSource-1} inited
- 2022-09-06 17:00:57.284 INFO 132 --- [ main] i.s.c.r.netty.NettyClientChannelManager : will connect to 192.168.56.211:31113
- 2022-09-06 17:00:57.284 INFO 132 --- [ main] i.s.c.rpc.netty.RmNettyRemotingClient : RM will register :jdbc:mysql://192.168.56.213:31306/seata_order
- 2022-09-06 17:00:57.285 INFO 132 --- [ main] i.s.core.rpc.netty.NettyPoolableFactory : NettyPool create channel to transactionRole:RMROLE,address:192.168.56.211:31113,msg:< RegisterRMRequest{resourceIds='jdbc:mysql://192.168.56.213:31306/seata_order', applicationId='alibaba-order-seata', transactionServiceGroup='zq'} >
- 2022-09-06 17:00:57.300 INFO 132 --- [ main] i.s.c.rpc.netty.RmNettyRemotingClient : register RM success. client version:1.5.1, server version:1.5.1,channel:[id: 0xe1b055dc, L:/192.168.56.1:56386 - R:/192.168.56.211:31113]
- 2022-09-06 17:00:57.301 INFO 132 --- [ main] i.s.core.rpc.netty.NettyPoolableFactory : register success, cost 10 ms, version:1.5.1,role:RMROLE,channel:[id: 0xe1b055dc, L:/192.168.56.1:56386 - R:/192.168.56.211:31113]
- 2022-09-06 17:00:57.745 INFO 132 --- [ main] o.s.c.openfeign.FeignClientFactoryBean : For 'alibaba-stock-seata' URL not provided. Will try picking an instance via load-balancing.
- 2022-09-06 17:00:57.847 INFO 132 --- [ main] i.s.s.a.GlobalTransactionScanner : Bean[com.xx.order.service.impl.OrderServiceImpl] with name [orderServiceImpl] would use interceptor [io.seata.spring.annotation.GlobalTransactionalInterceptor]
- 2022-09-06 17:00:57.933 WARN 132 --- [ main] c.n.c.sources.URLConfigurationSource : No URLs will be polled as dynamic configuration sources.
- 2022-09-06 17:00:57.933 INFO 132 --- [ main] c.n.c.sources.URLConfigurationSource : To enable URLs as dynamic configuration sources, define System property archaius.configurationSource.additionalUrls or make config.properties available on classpath.
- 2022-09-06 17:00:57.944 WARN 132 --- [ main] c.n.c.sources.URLConfigurationSource : No URLs will be polled as dynamic configuration sources.
- 2022-09-06 17:00:57.944 INFO 132 --- [ main] c.n.c.sources.URLConfigurationSource : To enable URLs as dynamic configuration sources, define System property archaius.configurationSource.additionalUrls or make config.properties available on classpath.
- 2022-09-06 17:00:58.110 INFO 132 --- [ main] o.s.s.concurrent.ThreadPoolTaskExecutor : Initializing ExecutorService 'applicationTaskExecutor'
- 2022-09-06 17:00:58.951 INFO 132 --- [ main] o.s.s.c.ThreadPoolTaskScheduler : Initializing ExecutorService 'Nacos-Watch-Task-Scheduler'
- 2022-09-06 17:00:59.958 INFO 132 --- [ main] c.a.n.p.a.s.c.ClientAuthPluginManager : [ClientAuthPluginManager] Load ClientAuthService com.alibaba.nacos.client.auth.impl.NacosClientAuthServiceImpl success.
- 2022-09-06 17:00:59.958 INFO 132 --- [ main] c.a.n.p.a.s.c.ClientAuthPluginManager : [ClientAuthPluginManager] Load ClientAuthService com.alibaba.nacos.client.auth.ram.RamClientAuthServiceImpl success.
- 2022-09-06 17:01:00.268 INFO 132 --- [ main] o.s.b.w.embedded.tomcat.TomcatWebServer : Tomcat started on port(s): 8072 (http) with context path ''
- 2022-09-06 17:01:00.287 INFO 132 --- [ main] c.a.c.n.registry.NacosServiceRegistry : nacos registry, DEFAULT_GROUP alibaba-order-seata 172.20.10.3:8072 register finished
- 2022-09-06 17:01:01.047 INFO 132 --- [ main] c.t.order.AlibabaOrderSeataApplication : Started AlibabaOrderSeataApplication in 14.933 seconds (JVM running for 16.759)
- 2022-09-06 17:01:13.060 INFO 132 --- [nio-8072-exec-1] o.a.c.c.C.[Tomcat].[localhost].[/] : Initializing Spring DispatcherServlet 'dispatcherServlet'
- 2022-09-06 17:01:13.060 INFO 132 --- [nio-8072-exec-1] o.s.web.servlet.DispatcherServlet : Initializing Servlet 'dispatcherServlet'
- 2022-09-06 17:01:13.071 INFO 132 --- [nio-8072-exec-1] o.s.web.servlet.DispatcherServlet : Completed initialization in 11 ms
- 2022-09-06 17:01:13.142 INFO 132 --- [nio-8072-exec-1] io.seata.tm.TransactionManagerHolder : TransactionManager Singleton io.seata.tm.DefaultTransactionManager@30a3de9b
- 2022-09-06 17:01:13.160 INFO 132 --- [nio-8072-exec-1] i.seata.tm.api.DefaultGlobalTransaction : Begin new global transaction [192.168.56.211:31113:3657226080688205940]
- assssssssssssssss
- 2022-09-06 17:01:14.196 INFO 132 --- [nio-8072-exec-1] c.netflix.config.ChainedDynamicProperty : Flipping property: alibaba-stock-seata.ribbon.ActiveConnectionsLimit to use NEXT property: niws.loadbalancer.availabilityFilteringRule.activeConnectionsLimit = 2147483647
- 2022-09-06 17:01:14.231 INFO 132 --- [nio-8072-exec-1] c.netflix.loadbalancer.BaseLoadBalancer : Client: alibaba-stock-seata instantiated a LoadBalancer: DynamicServerListLoadBalancer:{NFLoadBalancer:name=alibaba-stock-seata,current list of Servers=[],Load balancer stats=Zone stats: {},Server stats: []}ServerList:null
- 2022-09-06 17:01:14.240 INFO 132 --- [nio-8072-exec-1] c.n.l.DynamicServerListLoadBalancer : Using serverListUpdater PollingServerListUpdater
- 2022-09-06 17:01:14.336 INFO 132 --- [nio-8072-exec-1] c.netflix.config.ChainedDynamicProperty : Flipping property: alibaba-stock-seata.ribbon.ActiveConnectionsLimit to use NEXT property: niws.loadbalancer.availabilityFilteringRule.activeConnectionsLimit = 2147483647
- 2022-09-06 17:01:14.338 INFO 132 --- [nio-8072-exec-1] c.n.l.DynamicServerListLoadBalancer : DynamicServerListLoadBalancer for client alibaba-stock-seata initialized: DynamicServerListLoadBalancer:{NFLoadBalancer:name=alibaba-stock-seata,current list of Servers=[172.20.10.3:8073],Load balancer stats=Zone stats: {unknown=[Zone:unknown; Instance count:1; Active connections count: 0; Circuit breaker tripped count: 0; Active connections per server: 0.0;]
- },Server stats: [[Server:172.20.10.3:8073; Zone:UNKNOWN; Total Requests:0; Successive connection failure:0; Total blackout seconds:0; Last connection made:Thu Jan 01 08:00:00 CST 1970; First connection made: Thu Jan 01 08:00:00 CST 1970; Active Connections:0; total failure count in last (1000) msecs:0; average resp time:0.0; 90 percentile resp time:0.0; 95 percentile resp time:0.0; min resp time:0.0; max resp time:0.0; stddev resp time:0.0]
- ]}ServerList:com.alibaba.cloud.nacos.ribbon.NacosServerList@64034cde
- 2022-09-06 17:01:15.252 INFO 132 --- [erListUpdater-0] c.netflix.config.ChainedDynamicProperty : Flipping property: alibaba-stock-seata.ribbon.ActiveConnectionsLimit to use NEXT property: niws.loadbalancer.availabilityFilteringRule.activeConnectionsLimit = 2147483647
- 2022-09-06 17:01:15.348 INFO 132 --- [nio-8072-exec-1] i.seata.tm.api.DefaultGlobalTransaction : Suspending current transaction, xid = 192.168.56.211:31113:3657226080688205940
- 2022-09-06 17:01:15.348 INFO 132 --- [nio-8072-exec-1] i.seata.tm.api.DefaultGlobalTransaction : [192.168.56.211:31113:3657226080688205940] commit status: Committed
- 2022-09-06 17:01:16.086 INFO 132 --- [ch_RMROLE_1_1_8] i.s.c.r.p.c.RmBranchCommitProcessor : rm client handle branch commit process:xid=192.168.56.211:31113:3657226080688205940,branchId=3657226080688205942,branchType=AT,resourceId=jdbc:mysql://192.168.56.213:31306/seata_order,applicationData={"skipCheckLock":true}
- 2022-09-06 17:01:16.101 INFO 132 --- [ch_RMROLE_1_1_8] io.seata.rm.AbstractRMHandler : Branch committing: 192.168.56.211:31113:3657226080688205940 3657226080688205942 jdbc:mysql://192.168.56.213:31306/seata_order {"skipCheckLock":true}
- 2022-09-06 17:01:16.102 INFO 132 --- [ch_RMROLE_1_1_8] io.seata.rm.AbstractRMHandler : Branch commit result: PhaseTwo_Committed
- 2022-09-06 17:17:57.315 WARN 132 --- [MetaChecker_1_1] c.a.druid.pool.DruidAbstractDataSource : discard long time none received connection. , jdbcUrl : jdbc:mysql://192.168.56.213:31306/seata_order?characterEncoding=utf8&useSSL=false&serverTimezone=UTC, version : 1.2.3, lastPacketReceivedIdleMillis : 119997
结尾:已经把helm的部署给解决了。
部署方式:
或者是查看:helm-charthttps://heidaodageshiwo.github.io/helm-chart/
添加仓库 [root@master ~]# helm repo ls NAME URL myrepo https://heidaodageshiwo.github.io/helm-chart [root@master ~]# 更新仓库 pull helm包或者是直接去github下载即可 [root@master ~]# helm search repo myrepo NAME CHART VERSION APP VERSION DESCRIPTION myrepo/helloworld 0.1.0 1.16.0 A Helm chart for Kubernetes myrepo/nacos 0.1.5 1.0 A Helm chart for Kubernetes myrepo/nginx 13.1.7 1.23.1 NGINX Open Source is a web server that can be a... myrepo/seata-server 1.0.0 1.0 Seata Server [root@master ~]# mkdir seatahelmtest [root@master ~]# cd seatahelmtest/ [root@master seatahelmtest]# ll 总用量 0 [root@master seatahelmtest]# helm pull myrepo/seata-server [root@master seatahelmtest]# ls seata-server-1.0.0.tgz [root@master seatahelmtest]# tar -zxvf seata-server-1.0.0.tgz seata-server/Chart.yaml seata-server/values.yaml seata-server/templates/NOTES.txt seata-server/templates/_helpers.tpl seata-server/templates/deployment.yaml seata-server/templates/service.yaml seata-server/templates/tests/test-connection.yaml seata-server/.helmignore seata-server/node.yaml seata-server/servicesss.yaml [root@master seatahelmtest]# ls seata-server seata-server-1.0.0.tgz [root@master seatahelmtest]# ll 总用量 4 drwxr-xr-x 3 root root 119 9月 7 16:37 seata-server -rw-r--r-- 1 root root 2636 9月 7 16:37 seata-server-1.0.0.tgz [root@master seatahelmtest]# ls seata-server seata-server-1.0.0.tgz [root@master seatahelmtest]# helm install seata ./seata-server NAME: seata LAST DEPLOYED: Wed Sep 7 16:39:17 2022 NAMESPACE: default STATUS: deployed REVISION: 1 NOTES: 1. Get the application URL by running these commands: export NODE_PORT=$(kubectl get --namespace default -o jsonpath="{.spec.ports[0].nodePort}" services seata-seata-server) export NODE_IP=$(kubectl get nodes --namespace default -o jsonpath="{.items[0].status.addresses[0].address}") echo http://$NODE_IP:$NODE_PORT [root@master seatahelmtest]#
1.首先把代理的配置文件给挂载出来。我是用的是nfs: https://blog.csdn.net/hunheidaode/article/details/126623672 2.需要修改values.yaml里面的ip地址:我的是192.168.56.211 3.我是用的是mysql5.7 mysql8的一直报错 4.可以安装了
界面访问 ip:31005 seata可视化界面:
需要注意这个ip与端口
地址要一样即可。 代码的连接方式在博客中:k8s helm Seata1.5.1_hunheidaode的博客-CSDN博客 我也已经提到过了。这里就不贴了。
如有问题请在博客中发表评论。