设为首页 收藏本站
查看: 1407|回复: 0

[经验分享] openstack结合glusterfs存储 其二(部署以及优化)

[复制链接]

尚未签到

发表于 2019-2-1 12:41:42 | 显示全部楼层 |阅读模式
  3、部署O+G环境
3.1、每台机器上安装gfs组件
  # yum install -y glusterfs-server
  
3.2、在YUN21上安装packstack,并部署openstack
  
  首先更新每台机器
  # yum update -y && reboot
  这里在更新的时候由于系统的原因(和最简化系统对比),没有在Centos软件源中找到google为开头的两个软件,可以通过wget的方式把系统镜像源里边的这两个软件下载到本地并安装,之后再次运行命令。
  
  # yum install -y openstack-packstack
  
  # packstack --gen-answer-file answers.txt
  
  # vi answers.txt
  修改密码
  CONFIG_KEYSTONE_ADMIN_PW=openstack
  
  CONFIG_PROVISION_DEMO=n
  
  修改网络
  CONFIG_NOVA_COMPUTE_PRIVIF=eth1
  
  CONFIG_NOVA_NETWORK_PUBIF=eth0
  
  CONFIG_NOVA_NETWORK_PRIVIF=eth1
  
  to
  
  CONFIG_NOVA_COMPUTE_PRIVIF=eth1
  
  CONFIG_NOVA_NETWORK_PUBIF=eth2
  
  CONFIG_NOVA_NETWORK_PRIVIF=eth1
  (上边网卡的修改在不同的物理环境中,也就是物理机不同的环境中是不一样的,要视情况而修改,两个参数有PRIVIF”的表示内部网络,中间一个有“PUBIF”标识的表示外部网络,也就是分配浮动IP的网络。)
  添加计算节点
  CONFIG_COMPUTE_HOSTS=192.168.0.121
  
  to
  
  CONFIG_COMPUTE_HOSTS=192.168.0.121,192.168.0.122,192.168.0.123,192.168.0.124
  
  # packstack --answer-file answers.txt
  
  配置网桥
  [root@YUN21 ~]# cp /etc/sysconfig/network-scripts/ifcfg-eth2 ifcfg-eth2.bak
  
  [root@YUN21 ~]# cp /etc/sysconfig/network-scripts/ifcfg-eth2 /etc/sysconfig/network-scripts/ifcfg-br-ex
  
  [root@YUN21 ~]# vi /etc/sysconfig/network-scripts/ifcfg-eth2
  DEVICE=eth2
  HWADDR=xx:xx:xx:xx:xx:xx
  TYPE=OVSPort
  OVS_BRIDGE=br-ex
  DEVICETYPE=ovs
  ONBOOT=yes
  
  [root@YUN21 ~]# vi /etc/sysconfig/network-scripts/ifcfg-br-ex
  DEVICE=br-ex
  TYPE=OVSBridge
  DEVICETYPE=ovs
  ONBOOT=yes
  BOOTPROTO=static
  IPADDR=192.168.0.121
  NETMASK=255.255.255.128
  GATEWAY=10.231.29.1
  
  [root@YUN21 ~]# vi /etc/neutron/plugin.ini
  添加
  network_vlan_ranges = physnet1
  bridge_mappings = physnet1:br-ex
  
  [root@YUN21 ~]# service network restart
  
  [root@YUN21 ~]# ifconfig
  br-ex     Link encap:Ethernet  HWaddr xx:xx:xx:xx:xx:xx  
            inet addr:192.168.0.121  Bcast:192.168.0.127  Mask:255.255.255.0
            inet6 addr: fe80::49b:36ff:fed3:bb5e/64 Scope:Link
            UP BROADCAST RUNNING  MTU:1500  Metric:1
            RX packets:1407 errors:0 dropped:0 overruns:0 frame:0
            TX packets:856 errors:0 dropped:0 overruns:0 carrier:0
            collisions:0 txqueuelen:0
            RX bytes:309542 (302.2 KiB)  TX bytes:171147 (167.1 KiB)
  
  eth1      Link encap:Ethernet  HWaddr xx:xx:xx:xx:xx:xx  
            inet addr:10.0.0.21  Bcast:10.0.0.255  Mask:255.255.255.0
            inet6 addr: fe80::6e92:bfff:fe0b:de45/64 Scope:Link
            UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
            RX packets:142 errors:0 dropped:0 overruns:0 frame:0
            TX packets:14 errors:0 dropped:0 overruns:0 carrier:0
            collisions:0 txqueuelen:1000
            RX bytes:10730 (10.4 KiB)  TX bytes:1128 (1.1 KiB)
            Memory:dfa20000-dfa3ffff
  
  eth2      Link encap:Ethernet  HWaddr xx:xx:xx:xx:xx:xx  
            inet6 addr: fe80::6e92:bfff:fe0b:de44/64 Scope:Link
            UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
            RX packets:176062 errors:0 dropped:0 overruns:0 frame:0
            TX packets:80147 errors:0 dropped:0 overruns:0 carrier:0
            collisions:0 txqueuelen:1000
            RX bytes:231167565 (220.4 MiB)  TX bytes:9536425 (9.0 MiB)
            Memory:dfa00000-dfa1ffff
  
  eth4      Link encap:Ethernet  HWaddr xx:xx:xx:xx:xx:xx  
            inet addr:20.0.0.21  Bcast:20.0.0.255  Mask:255.255.255.0
            inet6 addr: fe80::7a24:afff:fe85:3a32/64 Scope:Link
            UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
            RX packets:0 errors:0 dropped:0 overruns:0 frame:0
            TX packets:9 errors:0 dropped:0 overruns:0 carrier:0
            collisions:0 txqueuelen:1000
            RX bytes:0 (0.0 b)  TX bytes:670 (670.0 b)
            Interrupt:68 Memory:fa000000-fa7fffff
  
  当名为br-ex的网卡的IP替换为eth2网卡的IP时才可以接下来的操作
  
3.3、创建gfs卷并挂载
  
  [root@YUN21 ~]# service glusterd status
  glusterd (pid  3124) is running...
  
  在任意一台机器说那个操作,把其他机器加入到存储池
  [root@YUN21 ~]# gluster peer probe 20.0.0.22
  peer probe: success.
  [root@YUN21 ~]# gluster peer probe 20.0.0.23
  peer probe: success.
  [root@YUN21 ~]# gluster peer probe 20.0.0.24
  peer probe: success.
  
  [root@YUN21 ~]# gluster peer status
  Number of Peers: 3
  
  Hostname: 20.0.0.22
  Uuid: 434fc5dd-22c9-49c8-9e42-4962279cdca6
  State: Peer in Cluster (Connected)
  
  Hostname: 20.0.0.23
  Uuid: a3c6770a-0b3b-4dc5-ad94-37e8c06da3b5
  State: Peer in Cluster (Connected)
  
  Hostname: 20.0.0.24
  Uuid: 13905ea7-0c32-4be0-9708-b6788033070c
  State: Peer in Cluster (Connected)
  
  每个机器上创建组成卷的二级目录
  # mkdir /gv0/brick
  # mkdir /gv1/brick
  # mkdir /gv2/brick
  
  
  创建nova
  [root@YUN21 ~]# gluster volume create nova replica 2 20.0.0.21:/gv0/brick/ 20.0.0.22:/gv0/brick/ 20.0.0.23:/gv0/brick/ 20.0.0.24:/gv0/brick/
  volume create: nova: success: please start the volume to access data
  
  [root@YUN21 ~]# gluster volume start nova
  volume start: nova: success
  
  [root@YUN21 ~]# gluster volume status nova
  Status of volume: nova
  Gluster process      Port Online Pid
  ------------------------------------------------------------------------------
  Brick 20.0.0.21:/gv0/brick    49152 Y 7672
  Brick 20.0.0.22:/gv0/brick    49152 Y 30221
  Brick 20.0.0.23:/gv0/brick    49152 Y 30432
  Brick 20.0.0.24:/gv0/brick    49152 Y 22918
  NFS Server on localhost     2049 Y 7687
  Self-heal Daemon on localhost    N/A Y 7693
  NFS Server on 20.0.0.24     2049 Y 22933
  Self-heal Daemon on 20.0.0.24    N/A Y 22938
  NFS Server on 20.0.0.22     2049 Y 30236
  Self-heal Daemon on 20.0.0.22    N/A Y 30242
  NFS Server on 20.0.0.23     2049 Y 30447
  Self-heal Daemon on 20.0.0.23    N/A Y 30453
  
  Task Status of Volume nova
  ------------------------------------------------------------------------------
  There are no active volume tasks
  
  每台机器上都配置自动挂载
  [root@YUN21 ~]# echo "20.0.0.21:/nova /var/lib/nova/instances/ glusterfs defaults,_netdev 0 0" >> /etc/fstab
  [root@YUN21 ~]# mount -a
  [root@YUN21 ~]# mount
  /dev/mapper/vg_YUN21-lv_root on / type ext4 (rw)
  proc on /proc type proc (rw)
  sysfs on /sys type sysfs (rw)
  devpts on /dev/pts type devpts (rw,gid=5,mode=620)
  tmpfs on /dev/shm type tmpfs (rw)
  /dev/sda1 on /boot type ext4 (rw)
  /dev/mapper/vg_YUN21-lv_gv0 on /gv0 type xfs (rw,nobarrier)
  /dev/mapper/vg_YUN21-lv_gv1 on /gv1 type xfs (rw,nobarrier)
  /dev/mapper/vg_YUN21-lv_gv2 on /gv2 type xfs (rw,nobarrier)
  none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
  /srv/loopback-device/swiftloopback on /srv/node/swiftloopback type ext4 (rw,noatime,nodiratime,loop=/dev/loop1,nobarrier,user_xattr)
  gvfs-fuse-daemon on /root/.gvfs type fuse.gvfs-fuse-daemon (rw,nosuid,nodev)
  20.0.0.21:/nova on /var/lib/nova/instances type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072)
  
  [root@YUN22 ~]# echo "20.0.0.22:/nova /var/lib/nova/instances/ glusterfs defaults,_netdev 0 0" >> /etc/fstab
  [root@YUN22 ~]# mount -a && mount
  /dev/mapper/vg_YUN13-lv_root on / type ext4 (rw)
  proc on /proc type proc (rw)
  sysfs on /sys type sysfs (rw)
  devpts on /dev/pts type devpts (rw,gid=5,mode=620)
  tmpfs on /dev/shm type tmpfs (rw)
  /dev/sda1 on /boot type ext4 (rw)
  /dev/mapper/vg_YUN13-lv_gv0 on /gv0 type xfs (rw)
  /dev/mapper/vg_YUN13-lv_gv1 on /gv1 type xfs (rw)
  /dev/mapper/vg_YUN13-lv_gv2 on /gv2 type xfs (rw)
  /dev/mapper/vg_YUN13-lv_gv3 on /gv3 type xfs (rw)
  none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
  20.0.0.22:/nova on /var/lib/nova/instances type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072)
  
  [root@YUN23 ~]# echo "20.0.0.23:/nova /var/lib/nova/instances/ glusterfs defaults,_netdev 0 0" >> /etc/fstab
  [root@YUN23 ~]# mount -a && mount
  /dev/mapper/vg_YUN23-lv_root on / type ext4 (rw)
  proc on /proc type proc (rw)
  sysfs on /sys type sysfs (rw)
  devpts on /dev/pts type devpts (rw,gid=5,mode=620)
  tmpfs on /dev/shm type tmpfs (rw)
  /dev/sda1 on /boot type ext4 (rw)
  /dev/mapper/vg_YUN23-lv_gv0 on /gv0 type xfs (rw)
  /dev/mapper/vg_YUN23-lv_gv1 on /gv1 type xfs (rw)
  /dev/mapper/vg_YUN23-lv_gv2 on /gv2 type xfs (rw)
  /dev/mapper/vg_YUN23-lv_gv3 on /gv3 type xfs (rw)
  none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
  20.0.0.23:/nova on /var/lib/nova/instances type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072)
  
  [root@YUN24 ~]# echo "20.0.0.24:/nova /var/lib/nova/instances/ glusterfs defaults,_netdev 0 0" >> /etc/fstab
  
  [root@YUN24 ~]# mount -a && mount
  /dev/mapper/vg_YUN17-lv_root on / type ext4 (rw)
  proc on /proc type proc (rw)
  sysfs on /sys type sysfs (rw)
  devpts on /dev/pts type devpts (rw,gid=5,mode=620)
  tmpfs on /dev/shm type tmpfs (rw)
  /dev/sda1 on /boot type ext4 (rw)
  /dev/mapper/vg_YUN17-lv_gv0 on /gv0 type xfs (rw)
  /dev/mapper/vg_YUN17-lv_gv1 on /gv1 type xfs (rw)
  /dev/mapper/vg_YUN17-lv_gv2 on /gv2 type xfs (rw)
  none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
  20.0.0.24:/nova on /var/lib/nova/instances type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072
  
  每台机器上操作
  # vi .bash_profile
  在最后添加
  export PS1='[\u@\h \W]\$'
  mount -a
  这个操作是为了保证在系统重启之后,glusterfs卷可以完成自动挂载
  
  查看并修改目录权限
  [root@YUN21 ~]#ll -d /var/lib/nova/instances/
  drwxr-xr-x 3 root root 46 Dec 25 17:18 /var/lib/nova/instances/
  [root@YUN21 ~]#chown -R nova:nova /var/lib/nova/instances/
  [root@YUN21 ~]#ll -d /var/lib/nova/instances/
  drwxr-xr-x 3 nova nova 46 Dec 25 17:18 /var/lib/nova/instances/
  其他三台机器依次做修改
  
  
  各台机器上重启服务
  #service openstack-nova-compute restart
  Stopping openstack-nova-compute:                           [  OK  ]
  Starting openstack-nova-compute:                           [  OK  ]
  
3.4、进入dash,上传镜像,创建网络,修改配额
  创建实例成功
3.5、创建glancecinder卷并挂载
  
  [root@YUN21 ~]#gluster volume create glance replica 2 20.0.0.21:/gv1/brick 20.0.0.22:/gv1/brick 20.0.0.23:/gv1/brick 20.0.0.24:/gv1/brick
  volume create: glance: success: please start the volume to access data
  
  [root@YUN21 ~]#gluster volume create cinder replica 2 20.0.0.21:/gv2/brick 20.0.0.22:/gv2/brick 20.0.0.23:/gv2/brick 20.0.0.24:/gv2/brick
  volume create: cinder: success: please start the volume to access data
  
  [root@YUN21 ~]#gluster volume start glance
  volume start: glance: success
  
  [root@YUN21 ~]#gluster volume start cinder
  volume start: cinder: success
  
  [root@YUN21 ~]#gluster volume status glance
  Status of volume: glance
  Gluster process      Port Online Pid
  ------------------------------------------------------------------------------
  Brick 20.0.0.21:/gv1/brick    49153 Y 18269
  Brick 20.0.0.22:/gv1/brick    49153 Y 39924
  Brick 20.0.0.23:/gv1/brick    49153 Y 40300
  Brick 20.0.0.24:/gv1/brick    49153 Y 30920
  NFS Server on localhost     2049 Y 18374
  Self-heal Daemon on localhost    N/A Y 18389
  NFS Server on 20.0.0.24     2049 Y 31005
  Self-heal Daemon on 20.0.0.24    N/A Y 31015
  NFS Server on 20.0.0.22     2049 Y 40010
  Self-heal Daemon on 20.0.0.22    N/A Y 40020
  NFS Server on 20.0.0.23     2049 Y 40385
  Self-heal Daemon on 20.0.0.23    N/A Y 40395
  
  Task Status of Volume glance
  ------------------------------------------------------------------------------
  There are no active volume tasks
  
  [root@YUN21 ~]#gluster volume status cinder
  Status of volume: cinder
  Gluster process      Port Online Pid
  ------------------------------------------------------------------------------
  Brick 20.0.0.21:/gv2/brick    49154 Y 18362
  Brick 20.0.0.22:/gv2/brick    49154 Y 39993
  Brick 20.0.0.23:/gv2/brick    49154 Y 40369
  Brick 20.0.0.24:/gv2/brick    49154 Y 30989
  NFS Server on localhost     2049 Y 18374
  Self-heal Daemon on localhost    N/A Y 18389
  NFS Server on 20.0.0.24     2049 Y 31005
  Self-heal Daemon on 20.0.0.24    N/A Y 31015
  NFS Server on 20.0.0.23     2049 Y 40385
  Self-heal Daemon on 20.0.0.23    N/A Y 40395
  NFS Server on 20.0.0.22     2049 Y 40010
  Self-heal Daemon on 20.0.0.22    N/A Y 40020
  
  Task Status of Volume cinder
  ------------------------------------------------------------------------------
  There are no active volume tasks
  
  配置glance卷和cinder卷自动挂载(只需要在YUN21上操作)
  [root@YUN21 ~]#echo "20.0.0.21:/glance /var/lib/glance/images/ glusterfs defaults,_netdev 0 0" >> /etc/fstab
  [root@YUN21 ~]#mount -a && mount
  /dev/mapper/vg_YUN21-lv_root on / type ext4 (rw)
  proc on /proc type proc (rw)
  sysfs on /sys type sysfs (rw)
  devpts on /dev/pts type devpts (rw,gid=5,mode=620)
  tmpfs on /dev/shm type tmpfs (rw)
  /dev/sda1 on /boot type ext4 (rw)
  /dev/mapper/vg_YUN21-lv_gv0 on /gv0 type xfs (rw,nobarrier)
  /dev/mapper/vg_YUN21-lv_gv1 on /gv1 type xfs (rw,nobarrier)
  /dev/mapper/vg_YUN21-lv_gv2 on /gv2 type xfs (rw,nobarrier)
  /srv/loopback-device/swiftloopback on /srv/node/swiftloopback type ext4 (rw,noatime,nodiratime,nobarrier,user_xattr,nobarrier,loop=/dev/loop0)
  none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
  20.0.0.21:/nova on /var/lib/nova/instances type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072)
  20.0.0.21:/glance on /var/lib/glance/images type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072)
  
  [root@YUN21 ~]#service openstack-glance-api restart
  Stopping openstack-glance-api:                             [  OK  ]
  Starting openstack-glance-api:                             [  OK  ]
  
  修改目录权限
  [root@YUN21 ~]#ll -d /var/lib/glance/images/
  drwxr-xr-x 3 root root 46 Dec 25 18:20 /var/lib/glance/images/
  [root@YUN21 ~]#chown -R glance:glance /var/lib/glance/images/
  [root@YUN21 ~]#ll -d /var/lib/glance/images/
  drwxr-xr-x 3 glance glance 46 Dec 25 18:20 /var/lib/glance/images/
  
  配置cinder
  [root@YUN21 ~]#vi /etc/cinder/share.conf
  20.0.0.21:/cinder
  
  [root@YUN21 ~]#chmod 0640 /etc/cinder/share.conf
  
  [root@YUN21 ~]#chown root:cinder /etc/cinder/share.conf
  
  [root@YUN21 ~]#vi /etc/cinder/cinder.conf
  volume_driver=cinder.volume.drivers.lvm.LVMISCSIDriver
  
  to
  
  volume_driver=cinder.volume.drivers.glusterfs.GlusterfsDriver
  
  添加
  glusterfs_shares_config=/etc/cinder/share.conf
  
  glusterfs_mount_point_base=/var/lib/cinder/volumes
  
  [root@YUN21 ~]#for i in api scheduler volume; do sudo service openstack-cinder-${i} restart;done
  Stopping openstack-cinder-api:                             [  OK  ]
  Starting openstack-cinder-api:                             [  OK  ]
  Stopping openstack-cinder-scheduler:                       [  OK  ]
  Starting openstack-cinder-scheduler:                       [  OK  ]
  Stopping openstack-cinder-volume:                          [  OK  ]
  Starting openstack-cinder-volume:                          [  OK  ]
  
  [root@YUN21 ~]#mount
  /dev/mapper/vg_YUN21-lv_root on / type ext4 (rw)
  proc on /proc type proc (rw)
  sysfs on /sys type sysfs (rw)
  devpts on /dev/pts type devpts (rw,gid=5,mode=620)
  tmpfs on /dev/shm type tmpfs (rw)
  /dev/sda1 on /boot type ext4 (rw)
  /dev/mapper/vg_YUN21-lv_gv0 on /gv0 type xfs (rw,nobarrier)
  /dev/mapper/vg_YUN21-lv_gv1 on /gv1 type xfs (rw,nobarrier)
  /dev/mapper/vg_YUN21-lv_gv2 on /gv2 type xfs (rw,nobarrier)
  /srv/loopback-device/swiftloopback on /srv/node/swiftloopback type ext4 (rw,noatime,nodiratime,nobarrier,user_xattr,nobarrier,loop=/dev/loop0)
  none on /proc/sys/fs/binfmt_misc type binfmt_misc (rw)
  20.0.0.21:/nova on /var/lib/nova/instances type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072)
  20.0.0.21:/glance on /var/lib/glance/images type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072)
  20.0.0.21:/cinder on /var/lib/cinder/volumes/6c05f25454fce4801c6aae690faff3dc type fuse.glusterfs (rw,default_permissions,allow_other,max_read=131072)
  
4、系统优化
  增加云主机带宽
  在控制节点上
  [root@YUN21 ~]#vi /etc/neutron/dhcp_agent.ini
  dnsmasq_config_file = /etc/neutron/dnsmasq-neutron.conf
  
  [root@YUN21 ~]#vi /etc/neutron/dnsmasq-neutron.conf
  dhcp-option-force=26,1400
  
  [root@YUN21 ~]#ethtool -K eth1 tso off
  [root@YUN21 ~]#ethtool -K eth2 tso off
  [root@YUN21 ~]#ethtool -K eth4 tso off
  [root@YUN21 ~]#ethtool -K eth1 gro off
  [root@YUN21 ~]#ethtool -K eth2 gro off
  [root@YUN21 ~]#ethtool -K eth4 gro off
  
  [root@YUN21 ~]#vi /etc/rc.d/rc.local
  ethtool -K eth1 tso off
  ethtool -K eth2 tso off
  ethtool -K eth4 tso
  
  对于创建的虚机也依照上边的方式关闭对应网卡的tsogro服务




运维网声明 1、欢迎大家加入本站运维交流群:群②:261659950 群⑤:202807635 群⑦870801961 群⑧679858003
2、本站所有主题由该帖子作者发表,该帖子作者与运维网享有帖子相关版权
3、所有作品的著作权均归原作者享有,请您和我们一样尊重他人的著作权等合法权益。如果您对作品感到满意,请购买正版
4、禁止制作、复制、发布和传播具有反动、淫秽、色情、暴力、凶杀等内容的信息,一经发现立即删除。若您因此触犯法律,一切后果自负,我们对此不承担任何责任
5、所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其内容的准确性、可靠性、正当性、安全性、合法性等负责,亦不承担任何法律责任
6、所有作品仅供您个人学习、研究或欣赏,不得用于商业或者其他用途,否则,一切后果均由您自己承担,我们对此不承担任何法律责任
7、如涉及侵犯版权等问题,请您及时通知我们,我们将立即采取措施予以解决
8、联系人Email:admin@iyunv.com 网址:www.yunweiku.com

所有资源均系网友上传或者通过网络收集,我们仅提供一个展示、介绍、观摩学习的平台,我们不对其承担任何法律责任,如涉及侵犯版权等问题,请您及时通知我们,我们将立即处理,联系人Email:kefu@iyunv.com,QQ:1061981298 本贴地址:https://www.yunweiku.com/thread-670456-1-1.html 上篇帖子: Cloudstack+Glusterfs+Kvm 集群(笔记) 下篇帖子: GlusterFS文件占用空间大小显示不统一问题分析
您需要登录后才可以回帖 登录 | 立即注册

本版积分规则

扫码加入运维网微信交流群X

扫码加入运维网微信交流群

扫描二维码加入运维网微信交流群,最新一手资源尽在官方微信交流群!快快加入我们吧...

扫描微信二维码查看详情

客服E-mail:kefu@iyunv.com 客服QQ:1061981298


QQ群⑦:运维网交流群⑦ QQ群⑧:运维网交流群⑧ k8s群:运维网kubernetes交流群


提醒:禁止发布任何违反国家法律、法规的言论与图片等内容;本站内容均来自个人观点与网络等信息,非本站认同之观点.


本站大部分资源是网友从网上搜集分享而来,其版权均归原作者及其网站所有,我们尊重他人的合法权益,如有内容侵犯您的合法权益,请及时与我们联系进行核实删除!



合作伙伴: 青云cloud

快速回复 返回顶部 返回列表