forked from mastodon/mastodon
-
Notifications
You must be signed in to change notification settings - Fork 0
/
docker-compose.yml
215 lines (204 loc) · 7.16 KB
/
docker-compose.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
version: '3'
services:
db:
restart: always
image: postgres:9.6-alpine
shm_size: 256mb
## 2022-11-19 - Lou - try to stop Postgres from OOMing
#command: postgres -c max_connections=300 -c shared_buffers=1GB -c effective_cache_size=3GB -c maintenance_work_mem=256MB -c checkpoint_completion_target=0.9 -c wal_buffers=16MB -c default_statistics_target=100 -c random_page_cost=1.1 -c effective_io_concurrency=200 -c work_mem=1747kB -c min_wal_size=2GB -c max_wal_size=8GB -c max_worker_processes=4 -c max_parallel_workers_per_gather=2
#command: postgres -c max_connections=300 -c shared_buffers=512MB -c effective_cache_size=1536MB -c maintenance_work_mem=128MB -c checkpoint_completion_target=0.9 -c wal_buffers=16MB -c default_statistics_target=100 -c random_page_cost=1.1 -c effective_io_concurrency=200 -c work_mem=873kB -c min_wal_size=2GB -c max_wal_size=8GB -c max_worker_processes=4 -c max_parallel_workers_per_gather=2
## 2022-11-21 - rabbit - tuning via packbats
command: postgres -c max_connections=300 -c shared_buffers=768MB -c effective_cache_size=2304MB -c maintenance_work_mem=192MB -c checkpoint_completion_target=0.9 -c wal_buffers=16MB -c default_statistics_target=100 -c random_page_cost=1.1 -c effective_io_concurrency=200 -c work_mem=1310kB -c min_wal_size=2GB -c max_wal_size=8GB -c max_worker_processes=4 -c max_parallel_workers_per_gather=2
networks:
- internal_network
healthcheck:
test: ['CMD', 'pg_isready', '-U', 'postgres']
volumes:
- ./postgres:/var/lib/postgresql/data
environment:
- 'POSTGRES_HOST_AUTH_METHOD=trust'
redis:
restart: always
image: redis:6-alpine
networks:
- internal_network
healthcheck:
test: ['CMD', 'redis-cli', 'ping']
volumes:
- ./redis:/data
web:
build: .
image: tootsuite/mastodon:v3.5.5
restart: always
env_file: .env.production
## 2022-11-14 - Nix - Added MAX_CONCURRENCY
## 2022-11-18 - Lou - MAX_CONCURRENCY -> WEB_CONCURRENCY
environment:
- MAX_THREADS=10
- WEB_CONCURRENCY=3
command: bash -c "rm -f /mastodon/tmp/pids/server.pid; bundle exec rails s -p 3000"
networks:
- external_network
- internal_network
healthcheck:
# prettier-ignore
test: ['CMD-SHELL', 'wget -q --spider --proxy=off localhost:3000/health || exit 1']
ports:
- '127.0.0.1:3000:3000'
depends_on:
- db
- redis
# - es
volumes:
- ./public/system:/mastodon/public/system
streaming:
build: .
image: tootsuite/mastodon:v3.5.5
restart: always
env_file: .env.production
command: node ./streaming
networks:
- external_network
- internal_network
healthcheck:
# prettier-ignore
test: ['CMD-SHELL', 'wget -q --spider --proxy=off localhost:4000/api/v1/streaming/health || exit 1']
ports:
- '127.0.0.1:4000:4000'
depends_on:
- db
- redis
sidekiq-low-volume:
build: .
image: tootsuite/mastodon:v3.5.5
restart: always
env_file: .env.production
### 2022-11-06 lou: increased number of threads, because the queue runs full.
### setting DB_POOL and MAX_THREADS in .env.production didn't change the thread number
### so adding "environment", and additionally "-c " to "command"
### 2022-11-07 nora: decreased MAX_THREADS and -c, to split default and pull queues
### to their own processes
### 2022-11-07 nora: split low volume queues into a low thread sidekiq
### 2022-11-10 secretpeej: change env var to DB_POOL, apply priorities from mastodon default sidekiq.yml
environment:
- DB_POOL=15
command: bundle exec sidekiq -c 15 -q mailers,2 -q scheduler,1
depends_on:
- db
- redis
networks:
- external_network
- internal_network
volumes:
- ./public/system:/mastodon/public/system
# # sidekiq-default-1:
# build: .
# image: tootsuite/mastodon:v3.5.5
# restart: always
# env_file: .env.production
# ### 2022-11-07 nora: split default queue into seperate process
# ### 2022-11-07 nora: add fallbacks so these workers can work on push and pull when default is done
# ### 2022-11-10 secretpeej: change env var to DB_POOL, reduce concurrency to 25
# environment:
# - DB_POOL=25
# command: bundle exec sidekiq -c 25 -q default -q push -q pull
# depends_on:
# - db
# - redis
# networks:
# - external_network
# - internal_network
# volumes:
# - ./public/system:/mastodon/public/system
sidekiq-default-2:
build: .
image: tootsuite/mastodon:v3.5.5
restart: always
env_file: .env.production
### 2022-11-07 nora: split default queue into seperate process
### 2022-11-07 nora: add fallbacks so these workers can work on push and pull when default is done
### 2022-11-10 secretpeej: change env var to DB_POOL, reduce concurrency to 25, adjust queue order
environment:
- DB_POOL=25
command: bundle exec sidekiq -c 25 -q default -q pull -q push
depends_on:
- db
- redis
networks:
- external_network
- internal_network
volumes:
- ./public/system:/mastodon/public/system
sidekiq-pull:
build: .
image: tootsuite/mastodon:v3.5.5
restart: always
env_file: .env.production
### 2022-11-07 nora: split pull queue into seperate process
### 2022-11-07 nora: add fallbacks so these workers can work on default and push when pull is done
environment:
- DB_POOL=25
command: bundle exec sidekiq -c 25 -q pull -q default -q push
depends_on:
- db
- redis
networks:
- external_network
- internal_network
volumes:
- ./public/system:/mastodon/public/system
sidekiq-push:
build: .
image: tootsuite/mastodon:v3.5.5
restart: always
env_file: .env.production
### 2022-11-10 secretpeej: change env var to DB_POOL, reduce concurrency to 25
environment:
- DB_POOL=25
command: bundle exec sidekiq -c 25 -q push -q default -q pull
depends_on:
- db
- redis
networks:
- external_network
- internal_network
volumes:
- ./public/system:/mastodon/public/system
# es:
# restart: always
# image: docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.10
# environment:
# - "ES_JAVA_OPTS=-Xms512m -Xmx512m"
# - "cluster.name=es-mastodon"
# - "discovery.type=single-node"
# - "bootstrap.memory_lock=true"
# networks:
# - internal_network
# healthcheck:
# test: ["CMD-SHELL", "curl --silent --fail localhost:9200/_cluster/health || exit 1"]
# volumes:
# - ./elasticsearch:/usr/share/elasticsearch/data
# ulimits:
# memlock:
# soft: -1
# hard: -1
## Uncomment to enable federation with tor instances along with adding the following ENV variables
## http_proxy=http://privoxy:8118
## ALLOW_ACCESS_TO_HIDDEN_SERVICE=true
# tor:
# image: sirboops/tor
# networks:
# - external_network
# - internal_network
#
# privoxy:
# image: sirboops/privoxy
# volumes:
# - ./priv-config:/opt/config
# networks:
# - external_network
# - internal_network
networks:
external_network:
internal_network:
internal: true