TL; DR: Umumiy sharh maqolasi - konteynerlarda ilovalarni ishga tushirish uchun muhitlarni solishtirish bo'yicha qo'llanma. Docker va boshqa shunga o'xshash tizimlarning imkoniyatlari ko'rib chiqiladi.
Bularning barchasi qayerdan kelganligi haqida bir oz tarix
ΠΡΡΠΎΡΠΈΡ
Ilovani izolyatsiya qilishning birinchi mashhur usuli bu chroot. Xuddi shu nomdagi tizim chaqiruvi ildiz katalogini o'zgartirishni ta'minlaydi - shuning uchun uni chaqirgan dasturga kirishni ta'minlaydi, faqat ushbu katalog ichidagi fayllarga kirishni ta'minlaydi. Ammo agar dastur ichida superfoydalanuvchi huquqlari berilsa, u chrootdan "qochib" qolishi va asosiy operatsion tizimga kirishi mumkin. Bundan tashqari, ildiz katalogini o'zgartirishdan tashqari, boshqa resurslar (RAM, protsessor), shuningdek, tarmoqqa kirish cheklanmaydi.
Keyingi yo'l - operatsion tizim yadrosi mexanizmlaridan foydalangan holda konteyner ichida to'liq operatsion tizimni ishga tushirish. Turli xil operatsion tizimlarda bu usul har xil deb ataladi, ammo mohiyati bir xil - bir nechta mustaqil operatsion tizimlar ishlaydi, ularning har biri asosiy operatsion tizimni boshqaradigan bir yadroda ishlaydi. Bunga Linux uchun FreeBSD Jails, Solaris Zones, OpenVZ va LXC kiradi. Izolyatsiya nafaqat disk maydoni uchun, balki boshqa resurslar uchun ham ta'minlanadi, xususan, har bir konteynerda protsessor vaqti, RAM, tarmoq o'tkazish qobiliyati bo'yicha cheklovlar bo'lishi mumkin. Chroot bilan solishtirganda, konteynerni tark etish qiyinroq, chunki konteynerdagi superfoydalanuvchi faqat konteynerning ichki qismiga kirish huquqiga ega, ammo konteyner ichidagi operatsion tizimni yangilab turish zarurati va eski yadrodan foydalanish tufayli. versiyalari (Linux uchun, kamroq darajada FreeBSD uchun tegishli), yadro izolyatsiyasi tizimini "buzib o'tish" va asosiy operatsion tizimga kirish imkoniyati nolga teng bo'lmagan.
Konteynerda to'liq operatsion tizimni ishga tushirish o'rniga (insializatsiya tizimi, paketlar menejeri va boshqalar bilan) ilovalarni darhol ishga tushirish mumkin, asosiysi ilovalarni ushbu imkoniyat bilan ta'minlash (kerakli kutubxonalar mavjudligi va). boshqa fayllar). Ushbu g'oya konteynerlashtirilgan ilovalarni virtualizatsiya qilish uchun asos bo'lib xizmat qildi, uning eng ko'zga ko'ringan va taniqli vakili Docker hisoblanadi. Oldingi tizimlar bilan solishtirganda, izolyatsiyalashning yanada moslashuvchan mexanizmlari, konteynerlar o'rtasidagi virtual tarmoqlarni o'rnatilgan qo'llab-quvvatlashi va konteyner ichidagi dasturning holati ko'p sonli jismoniy serverlardan konteynerlarni ishga tushirish uchun yagona yaxlit muhitni yaratish qobiliyatiga olib keldi. resurslarni qo'lda boshqarish zarurati.
Docker
Docker - bu eng mashhur ilovalarni konteynerlashtirish dasturi. Go tilida yozilgan bo'lib, u Linux yadrosining odatiy imkoniyatlaridan - guruhlar, nomlar bo'shliqlari, imkoniyatlar va boshqalardan, shuningdek, diskdagi joyni tejash uchun Aufs fayl tizimlaridan va shunga o'xshashlardan foydalanadi.
Manba: wikimedia
arxitektura
1.11 versiyasidan oldin Docker konteynerlar bilan barcha operatsiyalarni bajaradigan yagona xizmat sifatida ishlagan: konteynerlar uchun rasmlarni yuklab olish, konteynerlarni ishga tushirish, API so'rovlarini qayta ishlash. 1.11 versiyasidan boshlab Docker bir-biri bilan o'zaro aloqada bo'lgan bir nechta qismlarga bo'lingan: konteyner, konteynerlarning butun hayotiy tsiklini boshqarish uchun (disk maydonini ajratish, rasmlarni yuklab olish, tarmoqqa ulanish, ishga tushirish, konteynerlarning holatini kuzatish) va runC , Linux yadrosining guruhlari va boshqa xususiyatlaridan foydalanishga asoslangan konteyner ish vaqtlari. Docker xizmatining o'zi qoladi, ammo endi u faqat konteynerga uzatiladigan API so'rovlarini qayta ishlash uchun xizmat qiladi.
O'rnatish va sozlash
Docker-ni o'rnatishning eng sevimli usuli docker-machine bo'lib, u uzoq serverlarda (jumladan, turli xil bulutlarda) docker-ni to'g'ridan-to'g'ri o'rnatish va sozlashdan tashqari, uzoq serverlarning fayl tizimlari bilan ishlashga imkon beradi va turli xil buyruqlarni ham bajarishi mumkin.
Biroq, 2018 yildan beri loyiha deyarli ishlab chiqilmadi, shuning uchun biz uni ko'pchilik Linux distributivlari uchun odatiy tarzda o'rnatamiz - omborni qo'shish va kerakli paketlarni o'rnatish.
Ushbu usul, masalan, Ansible yoki boshqa shunga o'xshash tizimlar yordamida avtomatlashtirilgan o'rnatish uchun ham qo'llaniladi, ammo men buni ushbu maqolada ko'rib chiqmayman.
O'rnatish Centos 7 da amalga oshiriladi, men virtual mashinadan server sifatida foydalanaman, o'rnatish uchun quyidagi buyruqlarni bajaring:
# yum install -y yum-utils
# yum-config-manager --add-repo https://download.docker.com/linux/centos/docker-ce.repo
# yum install docker-ce docker-ce-cli containerd.io
O'rnatishdan so'ng siz xizmatni ishga tushirishingiz kerak, uni autoloadga qo'ying:
# systemctl enable docker
# systemctl start docker
# firewall-cmd --zone=public --add-port=2377/tcp --permanent
Bundan tashqari, siz docker guruhini yaratishingiz mumkin, uning foydalanuvchilari sudosiz docker bilan ishlashlari, jurnalni o'rnatishlari, tashqi tomondan API-ga kirishni yoqishlari, xavfsizlik devorini nozik sozlashni unutmang (ruxsat berilmagan hamma narsa Yuqoridagi va quyida keltirilgan misollarda taqiqlangan - soddalik va vizualizatsiya uchun men buni o'tkazib yubordim), lekin bu erda batafsilroq gapirmayman.
Boshqa xususiyatlar
Yuqoridagi docker mashinasiga qo'shimcha ravishda, shuningdek, docker registri, konteynerlar uchun tasvirlarni saqlash vositasi, shuningdek docker compose - ilovalarni konteynerlarda joylashtirishni avtomatlashtirish vositasi mavjud, YAML fayllari konteynerlarni qurish va sozlash uchun ishlatiladi va boshqa tegishli narsalar (masalan, tarmoqlar, ma'lumotlarni saqlash uchun doimiy fayl tizimlari).
Bundan tashqari, CICD uchun quvurlarni tashkil qilish uchun ham foydalanish mumkin. Yana bir qiziqarli xususiyat klaster rejimida ishlaydi, ya'ni to'da rejimi deb ataladigan (1.12 versiyasidan oldin u docker to'dasi deb nomlanardi), bu konteynerlarni ishga tushirish uchun bir nechta serverlardan bitta infratuzilmani yig'ish imkonini beradi. Barcha serverlar ustida virtual tarmoqni qo'llab-quvvatlash mavjud, o'rnatilgan yuk balansi mavjud, shuningdek, konteynerlar uchun sirlarni qo'llab-quvvatlash mavjud.
Docker compose-dan YAML fayllari kichik o'zgartirishlar bilan bunday klasterlar uchun ishlatilishi mumkin, turli maqsadlar uchun kichik va o'rta klasterlarga texnik xizmat ko'rsatishni to'liq avtomatlashtiradi. Katta klasterlar uchun Kubernetes afzalroq, chunki to'da rejimiga texnik xizmat ko'rsatish xarajatlari Kubernetesnikidan ko'proq bo'lishi mumkin. RunC ga qo'shimcha ravishda, konteynerlar uchun ijro muhiti sifatida, masalan, o'rnatishingiz mumkin
Docker bilan ishlash
O'rnatish va konfiguratsiyadan so'ng biz ishlab chiqish guruhi uchun GitLab va Docker Registry-ni o'rnatadigan klaster yaratishga harakat qilamiz. Serverlar sifatida men uchta virtual mashinadan foydalanaman, ularda qo'shimcha ravishda GlusterFS tarqatilgan FS-ni joylashtiraman, uni docker hajmlarini saqlash joyi sifatida ishlataman, masalan, docker registrining xavfsiz versiyasini ishga tushirish uchun. Ishlash uchun asosiy komponentlar: Docker Registry, Postgresql, Redis, GitLab Swarm tepasida GitLab Runner-ni qo'llab-quvvatlaydi. Postgresql klasterlash bilan ishga tushiriladi
GlusterFS-ni barcha serverlarda joylashtirish uchun (ular tugun1, tugun2, tugun3 deb ataladi) paketlarni o'rnatishingiz, xavfsizlik devorini yoqishingiz, kerakli kataloglarni yaratishingiz kerak:
# yum -y install centos-release-gluster7
# yum -y install glusterfs-server
# systemctl enable glusterd
# systemctl start glusterd
# firewall-cmd --add-service=glusterfs --permanent
# firewall-cmd --reload
# mkdir -p /srv/gluster
# mkdir -p /srv/docker
# echo "$(hostname):/docker /srv/docker glusterfs defaults,_netdev 0 0" >> /etc/fstab
O'rnatishdan so'ng, GlusterFS-ni sozlash bo'yicha ishlarni bitta tugundan davom ettirish kerak, masalan, 1-tugun:
# gluster peer probe node2
# gluster peer probe node3
# gluster volume create docker replica 3 node1:/srv/gluster node2:/srv/gluster node3:/srv/gluster force
# gluster volume start docker
Keyin olingan hajmni o'rnatishingiz kerak (buyruq barcha serverlarda bajarilishi kerak):
# mount /srv/docker
Swarm rejimi serverlardan birida sozlangan, u Lider bo'ladi, qolganlari klasterga qo'shilishlari kerak, shuning uchun birinchi serverda buyruqni bajarish natijasi qolganlarida ko'chirilishi va bajarilishi kerak.
Dastlabki klasterni sozlash, men tugun 1 da buyruqni bajaraman:
# docker swarm init
Swarm initialized: current node (a5jpfrh5uvo7svzz1ajduokyq) is now a manager.
To add a worker to this swarm, run the following command:
docker swarm join --token SWMTKN-1-0c5mf7mvzc7o7vjk0wngno2dy70xs95tovfxbv4tqt9280toku-863hyosdlzvd76trfptd4xnzd xx.xx.xx.xx:2377
To add a manager to this swarm, run 'docker swarm join-token manager' and follow the instructions.
# docker swarm join-token manager
Ikkinchi buyruqning natijasini nusxalash, tugun 2 va tugun 3 da bajaring:
# docker swarm join --token SWMTKN-x-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx-xxxxxxxxx xx.xx.xx.xx:2377
This node joined a swarm as a manager.
Bu serverlarning dastlabki konfiguratsiyasini yakunlaydi, keling, xizmatlarni sozlashni boshlaylik, agar boshqacha ko'rsatilmagan bo'lsa, bajariladigan buyruqlar tugun 1 dan ishga tushiriladi.
Avvalo, konteynerlar uchun tarmoqlar yaratamiz:
# docker network create --driver=overlay etcd
# docker network create --driver=overlay pgsql
# docker network create --driver=overlay redis
# docker network create --driver=overlay traefik
# docker network create --driver=overlay gitlab
Keyin biz serverlarni belgilaymiz, bu ba'zi xizmatlarni serverlarga ulash uchun kerak:
# docker node update --label-add nodename=node1 node1
# docker node update --label-add nodename=node2 node2
# docker node update --label-add nodename=node3 node3
Keyinchalik, Traefik va Stolonga kerak bo'lgan etcd ma'lumotlarini, KV xotirasini saqlash uchun kataloglarni yaratamiz. Postgresql-ga o'xshab, ular serverlarga bog'langan konteynerlar bo'ladi, shuning uchun biz ushbu buyruqni barcha serverlarda bajaramiz:
# mkdir -p /srv/etcd
Keyin etcd ni sozlash uchun fayl yarating va uni qo'llang:
00etcd.yml
version: '3.7'
services:
etcd1:
image: quay.io/coreos/etcd:latest
hostname: etcd1
command:
- etcd
- --name=etcd1
- --data-dir=/data.etcd
- --advertise-client-urls=http://etcd1:2379
- --listen-client-urls=http://0.0.0.0:2379
- --initial-advertise-peer-urls=http://etcd1:2380
- --listen-peer-urls=http://0.0.0.0:2380
- --initial-cluster=etcd1=http://etcd1:2380,etcd2=http://etcd2:2380,etcd3=http://etcd3:2380
- --initial-cluster-state=new
- --initial-cluster-token=etcd-cluster
networks:
- etcd
volumes:
- etcd1vol:/data.etcd
deploy:
replicas: 1
placement:
constraints: [node.labels.nodename == node1]
etcd2:
image: quay.io/coreos/etcd:latest
hostname: etcd2
command:
- etcd
- --name=etcd2
- --data-dir=/data.etcd
- --advertise-client-urls=http://etcd2:2379
- --listen-client-urls=http://0.0.0.0:2379
- --initial-advertise-peer-urls=http://etcd2:2380
- --listen-peer-urls=http://0.0.0.0:2380
- --initial-cluster=etcd1=http://etcd1:2380,etcd2=http://etcd2:2380,etcd3=http://etcd3:2380
- --initial-cluster-state=new
- --initial-cluster-token=etcd-cluster
networks:
- etcd
volumes:
- etcd2vol:/data.etcd
deploy:
replicas: 1
placement:
constraints: [node.labels.nodename == node2]
etcd3:
image: quay.io/coreos/etcd:latest
hostname: etcd3
command:
- etcd
- --name=etcd3
- --data-dir=/data.etcd
- --advertise-client-urls=http://etcd3:2379
- --listen-client-urls=http://0.0.0.0:2379
- --initial-advertise-peer-urls=http://etcd3:2380
- --listen-peer-urls=http://0.0.0.0:2380
- --initial-cluster=etcd1=http://etcd1:2380,etcd2=http://etcd2:2380,etcd3=http://etcd3:2380
- --initial-cluster-state=new
- --initial-cluster-token=etcd-cluster
networks:
- etcd
volumes:
- etcd3vol:/data.etcd
deploy:
replicas: 1
placement:
constraints: [node.labels.nodename == node3]
volumes:
etcd1vol:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/etcd"
etcd2vol:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/etcd"
etcd3vol:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/etcd"
networks:
etcd:
external: true
# docker stack deploy --compose-file 00etcd.yml etcd
Biroz vaqt o'tgach, etcd klasteri ko'tarilganligini tekshiramiz:
# docker exec $(docker ps | awk '/etcd/ {print $1}') etcdctl member list
ade526d28b1f92f7: name=etcd1 peerURLs=http://etcd1:2380 clientURLs=http://etcd1:2379 isLeader=false
bd388e7810915853: name=etcd3 peerURLs=http://etcd3:2380 clientURLs=http://etcd3:2379 isLeader=false
d282ac2ce600c1ce: name=etcd2 peerURLs=http://etcd2:2380 clientURLs=http://etcd2:2379 isLeader=true
# docker exec $(docker ps | awk '/etcd/ {print $1}') etcdctl cluster-health
member ade526d28b1f92f7 is healthy: got healthy result from http://etcd1:2379
member bd388e7810915853 is healthy: got healthy result from http://etcd3:2379
member d282ac2ce600c1ce is healthy: got healthy result from http://etcd2:2379
cluster is healthy
Postgresql uchun katalog yarating, barcha serverlarda buyruqni bajaring:
# mkdir -p /srv/pgsql
Keyin Postgresql-ni sozlash uchun fayl yarating:
01pgsql.yml
version: '3.7'
services:
pgsentinel:
image: sorintlab/stolon:master-pg10
command:
- gosu
- stolon
- stolon-sentinel
- --cluster-name=stolon-cluster
- --store-backend=etcdv3
- --store-endpoints=http://etcd1:2379,http://etcd2:2379,http://etcd3:2379
- --log-level=debug
networks:
- etcd
- pgsql
deploy:
replicas: 3
update_config:
parallelism: 1
delay: 30s
order: stop-first
failure_action: pause
pgkeeper1:
image: sorintlab/stolon:master-pg10
hostname: pgkeeper1
command:
- gosu
- stolon
- stolon-keeper
- --pg-listen-address=pgkeeper1
- --pg-repl-username=replica
- --uid=pgkeeper1
- --pg-su-username=postgres
- --pg-su-passwordfile=/run/secrets/pgsql
- --pg-repl-passwordfile=/run/secrets/pgsql_repl
- --data-dir=/var/lib/postgresql/data
- --cluster-name=stolon-cluster
- --store-backend=etcdv3
- --store-endpoints=http://etcd1:2379,http://etcd2:2379,http://etcd3:2379
networks:
- etcd
- pgsql
environment:
- PGDATA=/var/lib/postgresql/data
volumes:
- pgkeeper1:/var/lib/postgresql/data
secrets:
- pgsql
- pgsql_repl
deploy:
replicas: 1
placement:
constraints: [node.labels.nodename == node1]
pgkeeper2:
image: sorintlab/stolon:master-pg10
hostname: pgkeeper2
command:
- gosu
- stolon
- stolon-keeper
- --pg-listen-address=pgkeeper2
- --pg-repl-username=replica
- --uid=pgkeeper2
- --pg-su-username=postgres
- --pg-su-passwordfile=/run/secrets/pgsql
- --pg-repl-passwordfile=/run/secrets/pgsql_repl
- --data-dir=/var/lib/postgresql/data
- --cluster-name=stolon-cluster
- --store-backend=etcdv3
- --store-endpoints=http://etcd1:2379,http://etcd2:2379,http://etcd3:2379
networks:
- etcd
- pgsql
environment:
- PGDATA=/var/lib/postgresql/data
volumes:
- pgkeeper2:/var/lib/postgresql/data
secrets:
- pgsql
- pgsql_repl
deploy:
replicas: 1
placement:
constraints: [node.labels.nodename == node2]
pgkeeper3:
image: sorintlab/stolon:master-pg10
hostname: pgkeeper3
command:
- gosu
- stolon
- stolon-keeper
- --pg-listen-address=pgkeeper3
- --pg-repl-username=replica
- --uid=pgkeeper3
- --pg-su-username=postgres
- --pg-su-passwordfile=/run/secrets/pgsql
- --pg-repl-passwordfile=/run/secrets/pgsql_repl
- --data-dir=/var/lib/postgresql/data
- --cluster-name=stolon-cluster
- --store-backend=etcdv3
- --store-endpoints=http://etcd1:2379,http://etcd2:2379,http://etcd3:2379
networks:
- etcd
- pgsql
environment:
- PGDATA=/var/lib/postgresql/data
volumes:
- pgkeeper3:/var/lib/postgresql/data
secrets:
- pgsql
- pgsql_repl
deploy:
replicas: 1
placement:
constraints: [node.labels.nodename == node3]
postgresql:
image: sorintlab/stolon:master-pg10
command: gosu stolon stolon-proxy --listen-address 0.0.0.0 --cluster-name stolon-cluster --store-backend=etcdv3 --store-endpoints http://etcd1:2379,http://etcd2:2379,http://etcd3:2379
networks:
- etcd
- pgsql
deploy:
replicas: 3
update_config:
parallelism: 1
delay: 30s
order: stop-first
failure_action: rollback
volumes:
pgkeeper1:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/pgsql"
pgkeeper2:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/pgsql"
pgkeeper3:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/pgsql"
secrets:
pgsql:
file: "/srv/docker/postgres"
pgsql_repl:
file: "/srv/docker/replica"
networks:
etcd:
external: true
pgsql:
external: true
Biz sirlarni yaratamiz, faylni qo'llaymiz:
# </dev/urandom tr -dc 234567890qwertyuopasdfghjkzxcvbnmQWERTYUPASDFGHKLZXCVBNM | head -c $(((RANDOM%3)+15)) > /srv/docker/replica
# </dev/urandom tr -dc 234567890qwertyuopasdfghjkzxcvbnmQWERTYUPASDFGHKLZXCVBNM | head -c $(((RANDOM%3)+15)) > /srv/docker/postgres
# docker stack deploy --compose-file 01pgsql.yml pgsql
Biroz vaqt o'tgach (buyruqning chiqishiga qarang docker xizmati lsbarcha xizmatlar ko'tarildi) Postgresql klasterini ishga tushiring:
# docker exec $(docker ps | awk '/pgkeeper/ {print $1}') stolonctl --cluster-name=stolon-cluster --store-backend=etcdv3 --store-endpoints=http://etcd1:2379,http://etcd2:2379,http://etcd3:2379 init
Postgresql klasterining tayyorligini tekshirish:
# docker exec $(docker ps | awk '/pgkeeper/ {print $1}') stolonctl --cluster-name=stolon-cluster --store-backend=etcdv3 --store-endpoints=http://etcd1:2379,http://etcd2:2379,http://etcd3:2379 status
=== Active sentinels ===
ID LEADER
26baa11d false
74e98768 false
a8cb002b true
=== Active proxies ===
ID
4d233826
9f562f3b
b0c79ff1
=== Keepers ===
UID HEALTHY PG LISTENADDRESS PG HEALTHY PG WANTEDGENERATION PG CURRENTGENERATION
pgkeeper1 true pgkeeper1:5432 true 2 2
pgkeeper2 true pgkeeper2:5432 true 2 2
pgkeeper3 true pgkeeper3:5432 true 3 3
=== Cluster Info ===
Master Keeper: pgkeeper3
===== Keepers/DB tree =====
pgkeeper3 (master)
ββpgkeeper2
ββpgkeeper1
Biz traefikni konteynerlarga tashqaridan kirishni ochish uchun sozlaymiz:
03traefik.yml
version: '3.7'
services:
traefik:
image: traefik:latest
command: >
--log.level=INFO
--providers.docker=true
--entryPoints.web.address=:80
--providers.providersThrottleDuration=2
--providers.docker.watch=true
--providers.docker.swarmMode=true
--providers.docker.swarmModeRefreshSeconds=15s
--providers.docker.exposedbydefault=false
--accessLog.bufferingSize=0
--api=true
--api.dashboard=true
--api.insecure=true
networks:
- traefik
ports:
- 80:80
volumes:
- /var/run/docker.sock:/var/run/docker.sock
deploy:
replicas: 3
placement:
constraints:
- node.role == manager
preferences:
- spread: node.id
labels:
- traefik.enable=true
- traefik.http.routers.traefik.rule=Host(`traefik.example.com`)
- traefik.http.services.traefik.loadbalancer.server.port=8080
- traefik.docker.network=traefik
networks:
traefik:
external: true
# docker stack deploy --compose-file 03traefik.yml traefik
Biz Redis Cluster-ni ishga tushiramiz, buning uchun biz barcha tugunlarda saqlash katalogini yaratamiz:
# mkdir -p /srv/redis
05redis.yml
version: '3.7'
services:
redis-master:
image: 'bitnami/redis:latest'
networks:
- redis
ports:
- '6379:6379'
environment:
- REDIS_REPLICATION_MODE=master
- REDIS_PASSWORD=xxxxxxxxxxx
deploy:
mode: global
restart_policy:
condition: any
volumes:
- 'redis:/opt/bitnami/redis/etc/'
redis-replica:
image: 'bitnami/redis:latest'
networks:
- redis
ports:
- '6379'
depends_on:
- redis-master
environment:
- REDIS_REPLICATION_MODE=slave
- REDIS_MASTER_HOST=redis-master
- REDIS_MASTER_PORT_NUMBER=6379
- REDIS_MASTER_PASSWORD=xxxxxxxxxxx
- REDIS_PASSWORD=xxxxxxxxxxx
deploy:
mode: replicated
replicas: 3
update_config:
parallelism: 1
delay: 10s
restart_policy:
condition: any
redis-sentinel:
image: 'bitnami/redis:latest'
networks:
- redis
ports:
- '16379'
depends_on:
- redis-master
- redis-replica
entrypoint: |
bash -c 'bash -s <<EOF
"/bin/bash" -c "cat <<EOF > /opt/bitnami/redis/etc/sentinel.conf
port 16379
dir /tmp
sentinel monitor master-node redis-master 6379 2
sentinel down-after-milliseconds master-node 5000
sentinel parallel-syncs master-node 1
sentinel failover-timeout master-node 5000
sentinel auth-pass master-node xxxxxxxxxxx
sentinel announce-ip redis-sentinel
sentinel announce-port 16379
EOF"
"/bin/bash" -c "redis-sentinel /opt/bitnami/redis/etc/sentinel.conf"
EOF'
deploy:
mode: global
restart_policy:
condition: any
volumes:
redis:
driver: local
driver_opts:
type: 'none'
o: 'bind'
device: "/srv/redis"
networks:
redis:
external: true
# docker stack deploy --compose-file 05redis.yml redis
Docker registrini qo'shing:
06registry.yml
version: '3.7'
services:
registry:
image: registry:2.6
networks:
- traefik
volumes:
- registry_data:/var/lib/registry
deploy:
replicas: 1
placement:
constraints: [node.role == manager]
restart_policy:
condition: on-failure
labels:
- traefik.enable=true
- traefik.http.routers.registry.rule=Host(`registry.example.com`)
- traefik.http.services.registry.loadbalancer.server.port=5000
- traefik.docker.network=traefik
volumes:
registry_data:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/docker/registry"
networks:
traefik:
external: true
# mkdir /srv/docker/registry
# docker stack deploy --compose-file 06registry.yml registry
Va nihoyat - GitLab:
08gitlab-runner.yml
version: '3.7'
services:
gitlab:
image: gitlab/gitlab-ce:latest
networks:
- pgsql
- redis
- traefik
- gitlab
ports:
- 22222:22
environment:
GITLAB_OMNIBUS_CONFIG: |
postgresql['enable'] = false
redis['enable'] = false
gitlab_rails['registry_enabled'] = false
gitlab_rails['db_username'] = "gitlab"
gitlab_rails['db_password'] = "XXXXXXXXXXX"
gitlab_rails['db_host'] = "postgresql"
gitlab_rails['db_port'] = "5432"
gitlab_rails['db_database'] = "gitlab"
gitlab_rails['db_adapter'] = 'postgresql'
gitlab_rails['db_encoding'] = 'utf8'
gitlab_rails['redis_host'] = 'redis-master'
gitlab_rails['redis_port'] = '6379'
gitlab_rails['redis_password'] = 'xxxxxxxxxxx'
gitlab_rails['smtp_enable'] = true
gitlab_rails['smtp_address'] = "smtp.yandex.ru"
gitlab_rails['smtp_port'] = 465
gitlab_rails['smtp_user_name'] = "[email protected]"
gitlab_rails['smtp_password'] = "xxxxxxxxx"
gitlab_rails['smtp_domain'] = "example.com"
gitlab_rails['gitlab_email_from'] = '[email protected]'
gitlab_rails['smtp_authentication'] = "login"
gitlab_rails['smtp_tls'] = true
gitlab_rails['smtp_enable_starttls_auto'] = true
gitlab_rails['smtp_openssl_verify_mode'] = 'peer'
external_url 'http://gitlab.example.com/'
gitlab_rails['gitlab_shell_ssh_port'] = 22222
volumes:
- gitlab_conf:/etc/gitlab
- gitlab_logs:/var/log/gitlab
- gitlab_data:/var/opt/gitlab
deploy:
mode: replicated
replicas: 1
placement:
constraints:
- node.role == manager
labels:
- traefik.enable=true
- traefik.http.routers.gitlab.rule=Host(`gitlab.example.com`)
- traefik.http.services.gitlab.loadbalancer.server.port=80
- traefik.docker.network=traefik
gitlab-runner:
image: gitlab/gitlab-runner:latest
networks:
- gitlab
volumes:
- gitlab_runner_conf:/etc/gitlab
- /var/run/docker.sock:/var/run/docker.sock
deploy:
mode: replicated
replicas: 1
placement:
constraints:
- node.role == manager
volumes:
gitlab_conf:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/docker/gitlab/conf"
gitlab_logs:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/docker/gitlab/logs"
gitlab_data:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/docker/gitlab/data"
gitlab_runner_conf:
driver: local
driver_opts:
type: none
o: bind
device: "/srv/docker/gitlab/runner"
networks:
pgsql:
external: true
redis:
external: true
traefik:
external: true
gitlab:
external: true
# mkdir -p /srv/docker/gitlab/conf
# mkdir -p /srv/docker/gitlab/logs
# mkdir -p /srv/docker/gitlab/data
# mkdir -p /srv/docker/gitlab/runner
# docker stack deploy --compose-file 08gitlab-runner.yml gitlab
Klaster va xizmatlarning yakuniy holati:
# docker service ls
ID NAME MODE REPLICAS IMAGE PORTS
lef9n3m92buq etcd_etcd1 replicated 1/1 quay.io/coreos/etcd:latest
ij6uyyo792x5 etcd_etcd2 replicated 1/1 quay.io/coreos/etcd:latest
fqttqpjgp6pp etcd_etcd3 replicated 1/1 quay.io/coreos/etcd:latest
hq5iyga28w33 gitlab_gitlab replicated 1/1 gitlab/gitlab-ce:latest *:22222->22/tcp
dt7s6vs0q4qc gitlab_gitlab-runner replicated 1/1 gitlab/gitlab-runner:latest
k7uoezno0h9n pgsql_pgkeeper1 replicated 1/1 sorintlab/stolon:master-pg10
cnrwul4r4nse pgsql_pgkeeper2 replicated 1/1 sorintlab/stolon:master-pg10
frflfnpty7tr pgsql_pgkeeper3 replicated 1/1 sorintlab/stolon:master-pg10
x7pqqchi52kq pgsql_pgsentinel replicated 3/3 sorintlab/stolon:master-pg10
mwu2wl8fti4r pgsql_postgresql replicated 3/3 sorintlab/stolon:master-pg10
9hkbe2vksbzb redis_redis-master global 3/3 bitnami/redis:latest *:6379->6379/tcp
l88zn8cla7dc redis_redis-replica replicated 3/3 bitnami/redis:latest *:30003->6379/tcp
1utp309xfmsy redis_redis-sentinel global 3/3 bitnami/redis:latest *:30002->16379/tcp
oteb824ylhyp registry_registry replicated 1/1 registry:2.6
qovrah8nzzu8 traefik_traefik replicated 3/3 traefik:latest *:80->80/tcp, *:443->443/tcp
Yana nimani yaxshilash mumkin? Traefik-ni https konteynerlari bilan ishlash uchun sozlashni unutmang, Postgresql va Redis uchun tls shifrlashni qo'shing. Ammo umuman olganda, siz uni ishlab chiquvchilarga PoC sifatida berishingiz mumkin. Keling, Docker-ga alternativalarni ko'rib chiqaylik.
podman
Podkalar bo'yicha guruhlangan konteynerlarni ishga tushirish uchun yana bir taniqli dvigatel (podkalar, birgalikda joylashtirilgan konteynerlar guruhlari). Docker-dan farqli o'laroq, u konteynerlarni ishga tushirish uchun hech qanday xizmatni talab qilmaydi, barcha ishlar libpod kutubxonasi orqali amalga oshiriladi. Shuningdek, Go-da yozilgan, runC kabi konteynerlarni ishga tushirish uchun OCI-ga mos ish vaqti kerak.
Umuman olganda, Podman bilan ishlash Dockernikiga o'xshaydi, chunki siz buni shunday qila olasiz (ko'pchilik buni sinab ko'rganlar, shu jumladan ushbu maqola muallifi tomonidan da'vo qilingan):
$ alias docker=podman
va siz ishlashni davom ettirishingiz mumkin. Umuman olganda, Podman bilan bog'liq vaziyat juda qiziq, chunki agar Kubernetesning dastlabki versiyalari Docker bilan ishlagan bo'lsa, taxminan 2015 yildan boshlab, konteyner dunyosi standartlashtirilgandan keyin (OCI - Open Container Initiative) va Docker konteyner va runC ga ajratilgandan keyin. , Kubernetes-da ishlash uchun Docker-ga muqobil ishlab chiqilmoqda: CRI-O. Bu borada Podman Kubernetes tamoyillari asosida qurilgan Docker-ga muqobil, shu jumladan konteynerlarni guruhlash, lekin loyihaning asosiy maqsadi Docker uslubidagi konteynerlarni qo'shimcha xizmatlarsiz ishga tushirishdir. Aniq sabablarga ko'ra, to'da rejimi yo'q, chunki ishlab chiquvchilar aniq aytishadiki, agar sizga klaster kerak bo'lsa, Kubernetes-ni oling.
sozlama
Centos 7-ga o'rnatish uchun faqat Qo'shimchalar omborini faollashtiring va keyin hamma narsani buyruq bilan o'rnating:
# yum -y install podman
Boshqa xususiyatlar
Podman systemd uchun birliklarni yaratishi mumkin, shu bilan serverni qayta ishga tushirgandan so'ng konteynerlarni ishga tushirish muammosini hal qiladi. Bundan tashqari, systemd konteynerda pid 1 sifatida to'g'ri ishlashi e'lon qilinadi. Konteynerlarni qurish uchun alohida qurilish vositasi mavjud, shuningdek, uchinchi tomon vositalari - docker-compose analoglari mavjud, ular ham Kubernetesga mos keladigan konfiguratsiya fayllarini yaratadi, shuning uchun Podmandan Kubernetesga o'tish imkon qadar sodda.
Podman bilan ishlash
To'da rejimi yo'qligi sababli (agar klaster kerak bo'lsa, u Kubernetesga o'tishi kerak), biz uni alohida idishlarga yig'amiz.
Podman-compose dasturini o'rnating:
# yum -y install python3-pip
# pip3 install podman-compose
Podman uchun olingan konfiguratsiya fayli biroz farq qiladi, masalan, biz alohida hajmlar bo'limini to'g'ridan-to'g'ri xizmatlar bo'limiga ko'chirishimiz kerak edi.
gitlab-podman.yml
version: '3.7'
services:
gitlab:
image: gitlab/gitlab-ce:latest
hostname: gitlab.example.com
restart: unless-stopped
environment:
GITLAB_OMNIBUS_CONFIG: |
gitlab_rails['gitlab_shell_ssh_port'] = 22222
ports:
- "80:80"
- "22222:22"
volumes:
- /srv/podman/gitlab/conf:/etc/gitlab
- /srv/podman/gitlab/data:/var/opt/gitlab
- /srv/podman/gitlab/logs:/var/log/gitlab
networks:
- gitlab
gitlab-runner:
image: gitlab/gitlab-runner:alpine
restart: unless-stopped
depends_on:
- gitlab
volumes:
- /srv/podman/gitlab/runner:/etc/gitlab-runner
- /var/run/docker.sock:/var/run/docker.sock
networks:
- gitlab
networks:
gitlab:
# podman-compose -f gitlab-runner.yml -d up
Ish natijasi:
# podman ps
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
da53da946c01 docker.io/gitlab/gitlab-runner:alpine run --user=gitlab... About a minute ago Up About a minute ago 0.0.0.0:22222->22/tcp, 0.0.0.0:80->80/tcp root_gitlab-runner_1
781c0103c94a docker.io/gitlab/gitlab-ce:latest /assets/wrapper About a minute ago Up About a minute ago 0.0.0.0:22222->22/tcp, 0.0.0.0:80->80/tcp root_gitlab_1
Keling, u systemd va kubernetlar uchun nimani yaratishini ko'rib chiqaylik, buning uchun podning nomi yoki identifikatorini bilib olishimiz kerak:
# podman pod ls
POD ID NAME STATUS CREATED # OF CONTAINERS INFRA ID
71fc2b2a5c63 root Running 11 minutes ago 3 db40ab8bf84b
Kubernetes:
# podman generate kube 71fc2b2a5c63
# Generation of Kubernetes YAML is still under development!
#
# Save the output of this file and use kubectl create -f to import
# it into Kubernetes.
#
# Created with podman-1.6.4
apiVersion: v1
kind: Pod
metadata:
creationTimestamp: "2020-07-29T19:22:40Z"
labels:
app: root
name: root
spec:
containers:
- command:
- /assets/wrapper
env:
- name: PATH
value: /opt/gitlab/embedded/bin:/opt/gitlab/bin:/assets:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
- name: TERM
value: xterm
- name: HOSTNAME
value: gitlab.example.com
- name: container
value: podman
- name: GITLAB_OMNIBUS_CONFIG
value: |
gitlab_rails['gitlab_shell_ssh_port'] = 22222
- name: LANG
value: C.UTF-8
image: docker.io/gitlab/gitlab-ce:latest
name: rootgitlab1
ports:
- containerPort: 22
hostPort: 22222
protocol: TCP
- containerPort: 80
hostPort: 80
protocol: TCP
resources: {}
securityContext:
allowPrivilegeEscalation: true
capabilities: {}
privileged: false
readOnlyRootFilesystem: false
volumeMounts:
- mountPath: /var/opt/gitlab
name: srv-podman-gitlab-data
- mountPath: /var/log/gitlab
name: srv-podman-gitlab-logs
- mountPath: /etc/gitlab
name: srv-podman-gitlab-conf
workingDir: /
- command:
- run
- --user=gitlab-runner
- --working-directory=/home/gitlab-runner
env:
- name: PATH
value: /usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin
- name: TERM
value: xterm
- name: HOSTNAME
- name: container
value: podman
image: docker.io/gitlab/gitlab-runner:alpine
name: rootgitlab-runner1
resources: {}
securityContext:
allowPrivilegeEscalation: true
capabilities: {}
privileged: false
readOnlyRootFilesystem: false
volumeMounts:
- mountPath: /etc/gitlab-runner
name: srv-podman-gitlab-runner
- mountPath: /var/run/docker.sock
name: var-run-docker.sock
workingDir: /
volumes:
- hostPath:
path: /srv/podman/gitlab/runner
type: Directory
name: srv-podman-gitlab-runner
- hostPath:
path: /var/run/docker.sock
type: File
name: var-run-docker.sock
- hostPath:
path: /srv/podman/gitlab/data
type: Directory
name: srv-podman-gitlab-data
- hostPath:
path: /srv/podman/gitlab/logs
type: Directory
name: srv-podman-gitlab-logs
- hostPath:
path: /srv/podman/gitlab/conf
type: Directory
name: srv-podman-gitlab-conf
status: {}
systemd:
# podman generate systemd 71fc2b2a5c63
# pod-71fc2b2a5c6346f0c1c86a2dc45dbe78fa192ea02aac001eb8347ccb8c043c26.service
# autogenerated by Podman 1.6.4
# Thu Jul 29 15:23:28 EDT 2020
[Unit]
Description=Podman pod-71fc2b2a5c6346f0c1c86a2dc45dbe78fa192ea02aac001eb8347ccb8c043c26.service
Documentation=man:podman-generate-systemd(1)
Requires=container-781c0103c94aaa113c17c58d05ddabf8df4bf39707b664abcf17ed2ceff467d3.service container-da53da946c01449f500aa5296d9ea6376f751948b17ca164df438b7df6607864.service
Before=container-781c0103c94aaa113c17c58d05ddabf8df4bf39707b664abcf17ed2ceff467d3.service container-da53da946c01449f500aa5296d9ea6376f751948b17ca164df438b7df6607864.service
[Service]
Restart=on-failure
ExecStart=/usr/bin/podman start db40ab8bf84bf35141159c26cb6e256b889c7a98c0418eee3c4aa683c14fccaa
ExecStop=/usr/bin/podman stop -t 10 db40ab8bf84bf35141159c26cb6e256b889c7a98c0418eee3c4aa683c14fccaa
KillMode=none
Type=forking
PIDFile=/var/run/containers/storage/overlay-containers/db40ab8bf84bf35141159c26cb6e256b889c7a98c0418eee3c4aa683c14fccaa/userdata/conmon.pid
[Install]
WantedBy=multi-user.target
# container-da53da946c01449f500aa5296d9ea6376f751948b17ca164df438b7df6607864.service
# autogenerated by Podman 1.6.4
# Thu Jul 29 15:23:28 EDT 2020
[Unit]
Description=Podman container-da53da946c01449f500aa5296d9ea6376f751948b17ca164df438b7df6607864.service
Documentation=man:podman-generate-systemd(1)
RefuseManualStart=yes
RefuseManualStop=yes
BindsTo=pod-71fc2b2a5c6346f0c1c86a2dc45dbe78fa192ea02aac001eb8347ccb8c043c26.service
After=pod-71fc2b2a5c6346f0c1c86a2dc45dbe78fa192ea02aac001eb8347ccb8c043c26.service
[Service]
Restart=on-failure
ExecStart=/usr/bin/podman start da53da946c01449f500aa5296d9ea6376f751948b17ca164df438b7df6607864
ExecStop=/usr/bin/podman stop -t 10 da53da946c01449f500aa5296d9ea6376f751948b17ca164df438b7df6607864
KillMode=none
Type=forking
PIDFile=/var/run/containers/storage/overlay-containers/da53da946c01449f500aa5296d9ea6376f751948b17ca164df438b7df6607864/userdata/conmon.pid
[Install]
WantedBy=multi-user.target
# container-781c0103c94aaa113c17c58d05ddabf8df4bf39707b664abcf17ed2ceff467d3.service
# autogenerated by Podman 1.6.4
# Thu Jul 29 15:23:28 EDT 2020
[Unit]
Description=Podman container-781c0103c94aaa113c17c58d05ddabf8df4bf39707b664abcf17ed2ceff467d3.service
Documentation=man:podman-generate-systemd(1)
RefuseManualStart=yes
RefuseManualStop=yes
BindsTo=pod-71fc2b2a5c6346f0c1c86a2dc45dbe78fa192ea02aac001eb8347ccb8c043c26.service
After=pod-71fc2b2a5c6346f0c1c86a2dc45dbe78fa192ea02aac001eb8347ccb8c043c26.service
[Service]
Restart=on-failure
ExecStart=/usr/bin/podman start 781c0103c94aaa113c17c58d05ddabf8df4bf39707b664abcf17ed2ceff467d3
ExecStop=/usr/bin/podman stop -t 10 781c0103c94aaa113c17c58d05ddabf8df4bf39707b664abcf17ed2ceff467d3
KillMode=none
Type=forking
PIDFile=/var/run/containers/storage/overlay-containers/781c0103c94aaa113c17c58d05ddabf8df4bf39707b664abcf17ed2ceff467d3/userdata/conmon.pid
[Install]
WantedBy=multi-user.target
Afsuski, ishlaydigan konteynerlardan tashqari, systemd uchun yaratilgan birlik boshqa hech narsa qilmaydi (masalan, bunday xizmat qayta ishga tushirilganda eski konteynerlarni tozalash), shuning uchun siz bunday narsalarni o'zingiz qo'shishingiz kerak bo'ladi.
Aslida, Podman konteynerlar nima ekanligini sinab ko'rish, docker-compose uchun eski konfiguratsiyalarni o'tkazish va keyin kerak bo'lsa, klasterda Kubernetesga o'tish yoki Docker-ga foydalanish uchun qulayroq alternativani olish uchun etarli.
rkt
Loyiha
Flash
Ko'proq
topilmalar
Kubernetes bilan vaziyat juda qiziq: bir tomondan, Docker bilan siz klasterni yig'ishingiz mumkin (to'da rejimida), uning yordamida siz hatto mijozlar uchun ishlab chiqarish muhitini ham ishga tushirishingiz mumkin, bu ayniqsa kichik jamoalar (3-5 kishi) uchun to'g'ri keladi. ), yoki kichik umumiy yuk bilan , yoki Kubernetesni o'rnatishning nozikliklarini tushunish istagi yo'qligi, shu jumladan yuqori yuklar uchun.
Podman to'liq muvofiqlikni ta'minlamaydi, lekin u bitta muhim afzalliklarga ega - Kubernetes bilan moslik, shu jumladan qo'shimcha vositalar (buildah va boshqalar). Shuning uchun men ish uchun vositani tanlashga quyidagicha yondashaman: kichik jamoalar uchun yoki cheklangan byudjet bilan - Docker (mumkin bo'lgan to'da rejimi bilan), o'zim uchun shaxsiy localhostda ishlab chiqish uchun - Podman o'rtoqlar va boshqalar uchun. - Kubernetes.
Men Docker bilan vaziyat kelajakda o'zgarmasligiga ishonchim komil emas, axir ular kashshoflar, shuningdek, asta-sekin standartlashmoqda, lekin Podman, barcha kamchiliklari bilan (faqat Linuxda ishlaydi, klasterlash, yig'ish yo'q) va boshqa harakatlar uchinchi tomon qarorlari) kelajak aniqroq, shuning uchun men barchani ushbu topilmalarni sharhlarda muhokama qilishga taklif qilaman.
PS 3 avgustda biz ishga tushiramiz "
Chiqarilishdan oldin oldindan buyurtma narxi: 5000 rubl. "Docker Video Course" dasturini topish mumkin
Manba: www.habr.com