Last active
March 9, 2021 16:44
-
-
Save starofsky/c1c83e556fbfd0dfe336d652fc08c5f5 to your computer and use it in GitHub Desktop.
Log Sentry on Google Cloud
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
root@server:/opt/sentry/onpremise# ./install.sh | |
▶ Defining variables and helpers ... | |
▶ Parsing command line ... | |
▶ Setting up error handling ... | |
▶ Checking minimum requirements ... | |
▶ Creating volumes for persistent storage ... | |
Created sentry-data. | |
Created sentry-postgres. | |
Created sentry-redis. | |
Created sentry-zookeeper. | |
Created sentry-kafka. | |
Created sentry-clickhouse. | |
Created sentry-symbolicator. | |
▶ Ensuring files from examples ... | |
Creating sentry/sentry.conf.py... | |
Creating sentry/config.yml... | |
Creating sentry/requirements.txt... | |
Creating symbolicator/config.yml... | |
Creating relay/config.yml... | |
▶ Generating secret key ... | |
Secret key written to sentry/config.yml | |
▶ Replacing TSDB ... | |
▶ Fetching and updating Docker images ... | |
--no-ansi option is deprecated and will be removed in future versions. | |
Some service image(s) must be built from source by running: | |
docker-compose build sentry-cleanup snuba-cleanup symbolicator-cleanup | |
nightly: Pulling from getsentry/sentry | |
Digest: sha256:5a5ee326323d46ec730e0739acdb593f7cbb4cb396135e7b7ce6d623c4535ada | |
Status: Image is up to date for getsentry/sentry:nightly | |
docker.io/getsentry/sentry:nightly | |
▶ Building and tagging Docker images ... | |
--no-ansi option is deprecated and will be removed in future versions. | |
smtp uses an image, skipping | |
memcached uses an image, skipping | |
redis uses an image, skipping | |
postgres uses an image, skipping | |
zookeeper uses an image, skipping | |
kafka uses an image, skipping | |
clickhouse uses an image, skipping | |
geoipupdate uses an image, skipping | |
snuba-api uses an image, skipping | |
snuba-consumer uses an image, skipping | |
snuba-outcomes-consumer uses an image, skipping | |
snuba-sessions-consumer uses an image, skipping | |
snuba-transactions-consumer uses an image, skipping | |
snuba-replacer uses an image, skipping | |
snuba-subscription-consumer-events uses an image, skipping | |
snuba-subscription-consumer-transactions uses an image, skipping | |
symbolicator uses an image, skipping | |
web uses an image, skipping | |
cron uses an image, skipping | |
worker uses an image, skipping | |
ingest-consumer uses an image, skipping | |
post-process-forwarder uses an image, skipping | |
subscription-consumer-events uses an image, skipping | |
subscription-consumer-transactions uses an image, skipping | |
relay uses an image, skipping | |
nginx uses an image, skipping | |
Building snuba-cleanup | |
Sending build context to Docker daemon 3.072kB | |
Step 1/5 : ARG BASE_IMAGE | |
Step 2/5 : FROM ${BASE_IMAGE} | |
---> 81e63e225173 | |
Step 3/5 : RUN apt-get update && apt-get install -y --no-install-recommends cron && rm -r /var/lib/apt/lists/* | |
---> Running in 04f6d17910ed | |
Get:1 http://security.debian.org/debian-security buster/updates InRelease [65.4 kB] | |
Get:2 http://deb.debian.org/debian buster InRelease [122 kB] | |
Get:3 http://deb.debian.org/debian buster-updates InRelease [51.9 kB] | |
Get:4 http://security.debian.org/debian-security buster/updates/main amd64 Packages [268 kB] | |
Get:5 http://deb.debian.org/debian buster/main amd64 Packages [7907 kB] | |
Get:6 http://deb.debian.org/debian buster-updates/main amd64 Packages [9504 B] | |
Fetched 8423 kB in 1s (5780 kB/s) | |
Reading package lists... | |
Reading package lists... | |
Building dependency tree... | |
Reading state information... | |
The following additional packages will be installed: | |
lsb-base sensible-utils | |
Suggested packages: | |
anacron logrotate checksecurity | |
Recommended packages: | |
default-mta | mail-transport-agent | |
The following NEW packages will be installed: | |
cron lsb-base sensible-utils | |
0 upgraded, 3 newly installed, 0 to remove and 2 not upgraded. | |
Need to get 143 kB of archives. | |
After this operation, 383 kB of additional disk space will be used. | |
Get:1 http://deb.debian.org/debian buster/main amd64 sensible-utils all 0.0.12 [15.8 kB] | |
Get:2 http://deb.debian.org/debian buster/main amd64 lsb-base all 10.2019051400 [28.4 kB] | |
Get:3 http://deb.debian.org/debian buster/main amd64 cron amd64 3.0pl1-134+deb10u1 [99.0 kB] | |
debconf: delaying package configuration, since apt-utils is not installed | |
Fetched 143 kB in 0s (0 B/s) | |
Selecting previously unselected package sensible-utils. | |
(Reading database ... | |
(Reading database ... 5% | |
(Reading database ... 10% | |
(Reading database ... 15% | |
(Reading database ... 20% | |
(Reading database ... 25% | |
(Reading database ... 30% | |
(Reading database ... 35% | |
(Reading database ... 40% | |
(Reading database ... 45% | |
(Reading database ... 50% | |
(Reading database ... 55% | |
(Reading database ... 60% | |
(Reading database ... 65% | |
(Reading database ... 70% | |
(Reading database ... 75% | |
(Reading database ... 80% | |
(Reading database ... 85% | |
(Reading database ... 90% | |
(Reading database ... 95% | |
(Reading database ... 100% | |
(Reading database ... 6840 files and directories currently installed.) | |
Preparing to unpack .../sensible-utils_0.0.12_all.deb ... | |
Unpacking sensible-utils (0.0.12) ... | |
Selecting previously unselected package lsb-base. | |
Preparing to unpack .../lsb-base_10.2019051400_all.deb ... | |
Unpacking lsb-base (10.2019051400) ... | |
Selecting previously unselected package cron. | |
Preparing to unpack .../cron_3.0pl1-134+deb10u1_amd64.deb ... | |
Unpacking cron (3.0pl1-134+deb10u1) ... | |
Setting up lsb-base (10.2019051400) ... | |
Setting up sensible-utils (0.0.12) ... | |
Setting up cron (3.0pl1-134+deb10u1) ... | |
Adding group `crontab' (GID 101) ... | |
Done. | |
invoke-rc.d: could not determine current runlevel | |
invoke-rc.d: policy-rc.d denied execution of start. | |
Removing intermediate container 04f6d17910ed | |
---> 072f6edf9b77 | |
Step 4/5 : COPY entrypoint.sh /entrypoint.sh | |
---> 6de60178d8d0 | |
Step 5/5 : ENTRYPOINT ["/entrypoint.sh"] | |
---> Running in 1deeb5ce2da1 | |
Removing intermediate container 1deeb5ce2da1 | |
---> 17eba8f62477 | |
Successfully built 17eba8f62477 | |
Successfully tagged snuba-cleanup-onpremise-local:latest | |
Building symbolicator-cleanup | |
Sending build context to Docker daemon 3.072kB | |
Step 1/5 : ARG BASE_IMAGE | |
Step 2/5 : FROM ${BASE_IMAGE} | |
---> ee6e43b3fef9 | |
Step 3/5 : RUN apt-get update && apt-get install -y --no-install-recommends cron && rm -r /var/lib/apt/lists/* | |
---> Running in 771aacb410a9 | |
Get:1 http://security.debian.org/debian-security stretch/updates InRelease [53.0 kB] | |
Ign:2 http://deb.debian.org/debian stretch InRelease | |
Get:3 http://deb.debian.org/debian stretch-updates InRelease [93.6 kB] | |
Get:4 http://deb.debian.org/debian stretch Release [118 kB] | |
Get:5 http://deb.debian.org/debian stretch Release.gpg [2410 B] | |
Get:6 http://security.debian.org/debian-security stretch/updates/main amd64 Packages [660 kB] | |
Get:7 http://deb.debian.org/debian stretch-updates/main amd64 Packages [2596 B] | |
Get:8 http://deb.debian.org/debian stretch/main amd64 Packages [7080 kB] | |
Fetched 8009 kB in 1s (5947 kB/s) | |
Reading package lists... | |
Reading package lists... | |
Building dependency tree... | |
Reading state information... | |
Suggested packages: | |
anacron logrotate checksecurity | |
Recommended packages: | |
exim4 | postfix | mail-transport-agent | |
The following NEW packages will be installed: | |
cron | |
0 upgraded, 1 newly installed, 0 to remove and 2 not upgraded. | |
Need to get 95.4 kB of archives. | |
After this operation, 257 kB of additional disk space will be used. | |
Get:1 http://deb.debian.org/debian stretch/main amd64 cron amd64 3.0pl1-128+deb9u1 [95.4 kB] | |
debconf: delaying package configuration, since apt-utils is not installed | |
Fetched 95.4 kB in 0s (0 B/s) | |
Selecting previously unselected package cron. | |
(Reading database ... | |
(Reading database ... 5% | |
(Reading database ... 10% | |
(Reading database ... 15% | |
(Reading database ... 20% | |
(Reading database ... 25% | |
(Reading database ... 30% | |
(Reading database ... 35% | |
(Reading database ... 40% | |
(Reading database ... 45% | |
(Reading database ... 50% | |
(Reading database ... 55% | |
(Reading database ... 60% | |
(Reading database ... 65% | |
(Reading database ... 70% | |
(Reading database ... 75% | |
(Reading database ... 80% | |
(Reading database ... 85% | |
(Reading database ... 90% | |
(Reading database ... 95% | |
(Reading database ... 100% | |
(Reading database ... 6661 files and directories currently installed.) | |
Preparing to unpack .../cron_3.0pl1-128+deb9u1_amd64.deb ... | |
Unpacking cron (3.0pl1-128+deb9u1) ... | |
Setting up cron (3.0pl1-128+deb9u1) ... | |
Adding group `crontab' (GID 101) ... | |
Done. | |
update-rc.d: warning: start and stop actions are no longer supported; falling back to defaults | |
invoke-rc.d: could not determine current runlevel | |
invoke-rc.d: policy-rc.d denied execution of start. | |
Removing intermediate container 771aacb410a9 | |
---> c4cfa2d2e459 | |
Step 4/5 : COPY entrypoint.sh /entrypoint.sh | |
---> f9a3bd806a82 | |
Step 5/5 : ENTRYPOINT ["/entrypoint.sh"] | |
---> Running in fc0ebbc9a4c6 | |
Removing intermediate container fc0ebbc9a4c6 | |
---> eafaeaca32b6 | |
Successfully built eafaeaca32b6 | |
Successfully tagged symbolicator-cleanup-onpremise-local:latest | |
Building sentry-cleanup | |
Sending build context to Docker daemon 3.072kB | |
Step 1/5 : ARG BASE_IMAGE | |
Step 2/5 : FROM ${BASE_IMAGE} | |
---> 32009fa5d23c | |
Step 3/5 : RUN apt-get update && apt-get install -y --no-install-recommends cron && rm -r /var/lib/apt/lists/* | |
---> Running in 58504f63b406 | |
Get:1 http://security.debian.org/debian-security buster/updates InRelease [65.4 kB] | |
Get:2 http://deb.debian.org/debian buster InRelease [122 kB] | |
Get:3 http://deb.debian.org/debian buster-updates InRelease [51.9 kB] | |
Get:4 http://security.debian.org/debian-security buster/updates/main amd64 Packages [268 kB] | |
Get:5 http://deb.debian.org/debian buster/main amd64 Packages [7907 kB] | |
Get:6 http://deb.debian.org/debian buster-updates/main amd64 Packages [9504 B] | |
Fetched 8423 kB in 1s (6004 kB/s) | |
Reading package lists... | |
Reading package lists... | |
Building dependency tree... | |
Reading state information... | |
The following additional packages will be installed: | |
lsb-base sensible-utils | |
Suggested packages: | |
anacron logrotate checksecurity | |
Recommended packages: | |
default-mta | mail-transport-agent | |
The following NEW packages will be installed: | |
cron lsb-base sensible-utils | |
0 upgraded, 3 newly installed, 0 to remove and 2 not upgraded. | |
Need to get 143 kB of archives. | |
After this operation, 383 kB of additional disk space will be used. | |
Get:1 http://deb.debian.org/debian buster/main amd64 sensible-utils all 0.0.12 [15.8 kB] | |
Get:2 http://deb.debian.org/debian buster/main amd64 lsb-base all 10.2019051400 [28.4 kB] | |
Get:3 http://deb.debian.org/debian buster/main amd64 cron amd64 3.0pl1-134+deb10u1 [99.0 kB] | |
debconf: delaying package configuration, since apt-utils is not installed | |
Fetched 143 kB in 0s (9634 kB/s) | |
Selecting previously unselected package sensible-utils. | |
(Reading database ... | |
(Reading database ... 5% | |
(Reading database ... 10% | |
(Reading database ... 15% | |
(Reading database ... 20% | |
(Reading database ... 25% | |
(Reading database ... 30% | |
(Reading database ... 35% | |
(Reading database ... 40% | |
(Reading database ... 45% | |
(Reading database ... 50% | |
(Reading database ... 55% | |
(Reading database ... 60% | |
(Reading database ... 65% | |
(Reading database ... 70% | |
(Reading database ... 75% | |
(Reading database ... 80% | |
(Reading database ... 85% | |
(Reading database ... 90% | |
(Reading database ... 95% | |
(Reading database ... 100% | |
(Reading database ... 11935 files and directories currently installed.) | |
Preparing to unpack .../sensible-utils_0.0.12_all.deb ... | |
Unpacking sensible-utils (0.0.12) ... | |
Selecting previously unselected package lsb-base. | |
Preparing to unpack .../lsb-base_10.2019051400_all.deb ... | |
Unpacking lsb-base (10.2019051400) ... | |
Selecting previously unselected package cron. | |
Preparing to unpack .../cron_3.0pl1-134+deb10u1_amd64.deb ... | |
Unpacking cron (3.0pl1-134+deb10u1) ... | |
Setting up lsb-base (10.2019051400) ... | |
Setting up sensible-utils (0.0.12) ... | |
Setting up cron (3.0pl1-134+deb10u1) ... | |
Adding group `crontab' (GID 101) ... | |
Done. | |
invoke-rc.d: could not determine current runlevel | |
invoke-rc.d: policy-rc.d denied execution of start. | |
Removing intermediate container 58504f63b406 | |
---> 2805e84a8883 | |
Step 4/5 : COPY entrypoint.sh /entrypoint.sh | |
---> 2ec533ad1a71 | |
Step 5/5 : ENTRYPOINT ["/entrypoint.sh"] | |
---> Running in de3898313408 | |
Removing intermediate container de3898313408 | |
---> c4b7998ef67a | |
Successfully built c4b7998ef67a | |
Successfully tagged sentry-cleanup-onpremise-local:latest | |
Docker images built. | |
▶ Turning things off ... | |
--no-ansi option is deprecated and will be removed in future versions. | |
Removing network onpremise_default | |
Network onpremise_default not found. | |
--no-ansi option is deprecated and will be removed in future versions. | |
Removing network sentry_onpremise_default | |
Network sentry_onpremise_default not found. | |
▶ Setting up Zookeeper ... | |
--no-ansi option is deprecated and will be removed in future versions. | |
Creating network "sentry_onpremise_default" with the default driver | |
Creating volume "sentry_onpremise_sentry-secrets" with default driver | |
Creating volume "sentry_onpremise_sentry-smtp" with default driver | |
Creating volume "sentry_onpremise_sentry-zookeeper-log" with default driver | |
Creating volume "sentry_onpremise_sentry-kafka-log" with default driver | |
Creating volume "sentry_onpremise_sentry-smtp-log" with default driver | |
Creating volume "sentry_onpremise_sentry-clickhouse-log" with default driver | |
Creating sentry_onpremise_zookeeper_run ... | |
Creating sentry_onpremise_zookeeper_run ... done | |
▶ Bootstrapping and migrating Snuba ... | |
--no-ansi option is deprecated and will be removed in future versions. | |
Creating sentry_onpremise_clickhouse_1 ... | |
Creating sentry_onpremise_zookeeper_1 ... | |
Creating sentry_onpremise_redis_1 ... | |
Creating sentry_onpremise_clickhouse_1 ... done | |
Creating sentry_onpremise_zookeeper_1 ... done | |
Creating sentry_onpremise_kafka_1 ... | |
Creating sentry_onpremise_redis_1 ... done | |
Creating sentry_onpremise_kafka_1 ... done | |
Creating sentry_onpremise_snuba-api_run ... | |
Creating sentry_onpremise_snuba-api_run ... done | |
+ '[' b = - ']' | |
+ snuba bootstrap --help | |
+ set -- snuba bootstrap --no-migrate --force | |
+ set gosu snuba snuba bootstrap --no-migrate --force | |
+ exec gosu snuba snuba bootstrap --no-migrate --force | |
%3|1615263992.404|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.5:9092 failed: Connection refused (after 11ms in state CONNECT) | |
2021-03-09 04:26:33,399 Connection to Kafka failed (attempt 0) | |
Traceback (most recent call last): | |
File "/usr/src/snuba/snuba/cli/bootstrap.py", line 55, in bootstrap | |
client.list_topics(timeout=1) | |
cimpl.KafkaException: KafkaError{code=_TRANSPORT,val=-195,str="Failed to get metadata: Local: Broker transport failure"} | |
%3|1615263993.400|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.5:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
%3|1615263994.401|FAIL|rdkafka#producer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.5:9092 failed: Connection refused (after 0ms in state CONNECT) | |
%3|1615263995.402|FAIL|rdkafka#producer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.5:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
2021-03-09 04:26:35,402 Connection to Kafka failed (attempt 1) | |
Traceback (most recent call last): | |
File "/usr/src/snuba/snuba/cli/bootstrap.py", line 55, in bootstrap | |
client.list_topics(timeout=1) | |
cimpl.KafkaException: KafkaError{code=_TRANSPORT,val=-195,str="Failed to get metadata: Local: Broker transport failure"} | |
%3|1615263996.405|FAIL|rdkafka#producer-3| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.5:9092 failed: Connection refused (after 0ms in state CONNECT) | |
2021-03-09 04:26:37,406 Connection to Kafka failed (attempt 2) | |
Traceback (most recent call last): | |
File "/usr/src/snuba/snuba/cli/bootstrap.py", line 55, in bootstrap | |
client.list_topics(timeout=1) | |
cimpl.KafkaException: KafkaError{code=_TRANSPORT,val=-195,str="Failed to get metadata: Local: Broker transport failure"} | |
2021-03-09 04:26:39,060 Topic cdc created | |
2021-03-09 04:26:39,060 Topic events created | |
2021-03-09 04:26:39,060 Topic snuba-commit-log created | |
2021-03-09 04:26:39,061 Topic event-replacements created | |
2021-03-09 04:26:39,061 Topic outcomes created | |
2021-03-09 04:26:39,061 Topic ingest-sessions created | |
--no-ansi option is deprecated and will be removed in future versions. | |
Creating sentry_onpremise_snuba-api_run ... | |
Creating sentry_onpremise_snuba-api_run ... done | |
+ '[' m = - ']' | |
+ snuba migrations --help | |
+ set -- snuba migrations migrate --force | |
+ set gosu snuba snuba migrations migrate --force | |
+ exec gosu snuba snuba migrations migrate --force | |
Finished running migrations | |
▶ Creating additional Kafka topics ... | |
--no-ansi option is deprecated and will be removed in future versions. | |
Creating sentry_onpremise_kafka_run ... | |
Creating sentry_onpremise_kafka_run ... done | |
Created topic ingest-attachments. | |
--no-ansi option is deprecated and will be removed in future versions. | |
Creating sentry_onpremise_kafka_run ... | |
Creating sentry_onpremise_kafka_run ... done | |
Created topic ingest-transactions. | |
--no-ansi option is deprecated and will be removed in future versions. | |
Creating sentry_onpremise_kafka_run ... | |
Creating sentry_onpremise_kafka_run ... done | |
Created topic ingest-events. | |
▶ Ensuring proper PostgreSQL version ... | |
▶ Setting up database ... | |
--no-ansi option is deprecated and will be removed in future versions. | |
Creating sentry_onpremise_smtp_1 ... | |
Creating sentry_onpremise_postgres_1 ... | |
Creating sentry_onpremise_symbolicator_1 ... | |
Creating sentry_onpremise_memcached_1 ... | |
Creating sentry_onpremise_snuba-transactions-consumer_1 ... | |
Creating sentry_onpremise_snuba-consumer_1 ... | |
Creating sentry_onpremise_snuba-sessions-consumer_1 ... | |
Creating sentry_onpremise_snuba-replacer_1 ... | |
Creating sentry_onpremise_snuba-subscription-consumer-events_1 ... | |
Creating sentry_onpremise_snuba-outcomes-consumer_1 ... | |
Creating sentry_onpremise_snuba-api_1 ... | |
Creating sentry_onpremise_snuba-subscription-consumer-transactions_1 ... | |
Creating sentry_onpremise_postgres_1 ... done | |
Creating sentry_onpremise_symbolicator_1 ... done | |
Creating sentry_onpremise_snuba-transactions-consumer_1 ... done | |
Creating sentry_onpremise_snuba-consumer_1 ... done | |
Creating sentry_onpremise_memcached_1 ... done | |
Creating sentry_onpremise_snuba-replacer_1 ... done | |
Creating sentry_onpremise_smtp_1 ... done | |
Creating sentry_onpremise_snuba-subscription-consumer-events_1 ... done | |
Creating sentry_onpremise_snuba-sessions-consumer_1 ... done | |
Creating sentry_onpremise_snuba-outcomes-consumer_1 ... done | |
Creating sentry_onpremise_snuba-api_1 ... done | |
Creating sentry_onpremise_snuba-subscription-consumer-transactions_1 ... done | |
Creating sentry_onpremise_web_run ... | |
Creating sentry_onpremise_web_run ... done | |
Installing additional dependencies... | |
04:27:21 [WARNING] sentry.utils.geo: Error opening GeoIP database: /geoip/GeoLite2-City.mmdb | |
04:27:21 [WARNING] sentry.utils.geo: Error opening GeoIP database in Rust: /geoip/GeoLite2-City.mmdb | |
04:27:27 [INFO] sentry.plugins.github: apps-not-configured | |
* Unknown config option found: 'slack.legacy-app' | |
Operations to perform: | |
Apply all migrations: admin, auth, contenttypes, jira_ac, nodestore, sentry, sessions, sites, social_auth | |
Running migrations: | |
Applying sentry.0001_initial... OK | |
Applying contenttypes.0001_initial... OK | |
Applying admin.0001_initial... OK | |
Applying admin.0002_logentry_remove_auto_add... OK | |
Applying contenttypes.0002_remove_content_type_name... OK | |
Applying auth.0001_initial... OK | |
Applying auth.0002_alter_permission_name_max_length... OK | |
Applying auth.0003_alter_user_email_max_length... OK | |
Applying auth.0004_alter_user_username_opts... OK | |
Applying auth.0005_alter_user_last_login_null... OK | |
Applying auth.0006_require_contenttypes_0002... OK | |
Applying auth.0007_alter_validators_add_error_messages... OK | |
Applying auth.0008_alter_user_username_max_length... OK | |
Applying jira_ac.0001_initial... OK | |
Applying nodestore.0001_initial... OK | |
Applying nodestore.0002_nodestore_no_dictfield... OK | |
Applying sentry.0002_912_to_recent... OK | |
Applying sentry.0003_auto_20191022_0122... OK | |
Applying sentry.0004_bitfieldtestmodel_blankjsonfieldtestmodel_callabledefaultmodel_jsonfieldtestmodel_jsonfieldwithdefau... OK | |
Applying sentry.0005_fix_content_types... OK | |
Applying sentry.0006_sentryapp_date_published... OK | |
Applying sentry.0007_auto_20191029_0131... OK | |
Applying sentry.0008_auto_20191030_0016... OK | |
Applying sentry.0009_auto_20191101_1608... OK | |
Applying sentry.0010_auto_20191104_1641... OK | |
Applying sentry.0011_remove_pagerdutyservice_service_id_from_state... OK | |
Applying sentry.0012_remove_pagerdutyservice_service_id... OK | |
Applying sentry.0013_auto_20191111_1829... OK | |
Applying sentry.0014_delete_sentryappwebhookerror... OK | |
Applying sentry.0015_delete_sentryappwebhookerror_db... OK | |
Applying sentry.0016_delete_alert_rule_deprecated_fields... OK | |
Applying sentry.0017_incident_aggregation... OK | |
Applying sentry.0018_discoversavedquery_version... OK | |
Applying sentry.0019_auto_20191114_2040... OK | |
Applying sentry.0020_auto_20191125_1420... OK | |
Applying sentry.0021_auto_20191203_1803... OK | |
Applying sentry.0021_auto_20191202_1716... OK | |
Applying sentry.0022_merge... OK | |
Applying sentry.0023_hide_environment_none_20191126... OK | |
Applying sentry.0024_auto_20191230_2052...Nothing to do, skipping migration. | |
OK | |
Applying sentry.0025_organizationaccessrequest_requester... OK | |
Applying sentry.0026_delete_event... OK | |
Applying sentry.0027_exporteddata... OK | |
Applying sentry.0028_user_reports... OK | |
Applying sentry.0029_discover_query_upgrade... OK | |
Applying sentry.0030_auto_20200201_0039... OK | |
Applying sentry.0031_delete_alert_rules_and_incidents... OK | |
Applying sentry.0032_delete_alert_email... OK | |
Applying sentry.0033_auto_20200210_2137... OK | |
Applying sentry.0034_auto_20200210_2311... OK | |
Applying sentry.0035_auto_20200127_1711... OK | |
Applying sentry.0036_auto_20200213_0106... OK | |
Applying sentry.0037_auto_20200213_0140... OK | |
Applying sentry.0038_auto_20200213_1904... OK | |
Applying sentry.0039_delete_incidentsuspectcommit... OK | |
Applying sentry.0040_remove_incidentsuspectcommittable... OK | |
Applying sentry.0041_incidenttrigger_date_modified... OK | |
Applying sentry.0042_auto_20200214_1607... OK | |
Applying sentry.0043_auto_20200218_1903... OK | |
Applying sentry.0044_auto_20200219_0018... OK | |
Applying sentry.0045_remove_incidentactivity_event_stats_snapshot... OK | |
Applying sentry.0046_auto_20200221_1735... OK | |
Applying sentry.0047_auto_20200224_2319... OK | |
Applying sentry.0048_auto_20200302_1825... OK | |
Applying sentry.0049_auto_20200304_0254... OK | |
Applying sentry.0050_auto_20200306_2346... OK | |
Applying sentry.0051_fix_auditlog_pickled_data... OK | |
Applying sentry.0052_organizationonboardingtask_completion_seen... OK | |
Applying sentry.0053_migrate_alert_task_onboarding... OK | |
Applying sentry.0054_create_key_transaction... OK | |
Applying sentry.0055_query_subscription_status... OK | |
Applying sentry.0056_remove_old_functions... OK | |
Applying sentry.0057_remove_unused_project_flag... OK | |
Applying sentry.0058_project_issue_alerts_targeting... OK | |
Applying sentry.0059_add_new_sentry_app_features... OK | |
Applying sentry.0060_add_file_eventattachment_index... OK | |
Applying sentry.0061_alertrule_partial_index... OK | |
Applying sentry.0062_key_transactions_unique_with_owner... OK | |
Applying sentry.0063_drop_alertrule_constraint... OK | |
Applying sentry.0064_project_has_transactions... OK | |
Applying sentry.0065_add_incident_status_method... OK | |
Applying sentry.0066_alertrule_manager... OK | |
Applying sentry.0067_migrate_rules_alert_targeting... OK | |
Applying sentry.0068_project_default_flags... OK | |
Applying sentry.0069_remove_tracked_superusers... OK | |
Applying sentry.0070_incident_snapshot_support... OK | |
Applying sentry.0071_add_default_fields_model_subclass... OK | |
Applying sentry.0072_alert_rules_query_changes... OK | |
Applying sentry.0073_migrate_alert_query_model... OK | |
Applying sentry.0074_add_metric_alert_feature... OK | |
Applying sentry.0075_metric_alerts_fix_releases... OK | |
Applying sentry.0076_alert_rules_disable_constraints... OK | |
Applying sentry.0077_alert_query_col_drop_state... OK | |
Applying sentry.0078_incident_field_updates... OK | |
Applying sentry.0079_incidents_remove_query_field_state... OK | |
Applying sentry.0080_alert_rules_drop_unused_tables_cols... OK | |
Applying sentry.0081_add_integraiton_upgrade_audit_log... OK | |
Applying sentry.0082_alert_rules_threshold_float... OK | |
Applying sentry.0083_add_max_length_webhook_url... OK | |
Applying sentry.0084_exported_data_blobs... OK | |
Applying sentry.0085_fix_error_rate_snuba_query... OK | |
Applying sentry.0086_sentry_app_installation_for_provider... OK | |
Applying sentry.0087_fix_time_series_data_type... OK | |
Applying sentry.0088_rule_level_resolve_threshold_type... OK | |
Applying sentry.0089_rule_level_fields_backfill... OK | |
Applying sentry.0090_fix_auditlog_pickled_data_take_2... OK | |
Applying sentry.0091_alertruleactivity... OK | |
Applying sentry.0092_remove_trigger_threshold_type_nullable... OK | |
Applying sentry.0093_make_identity_user_id_textfield... OK | |
Applying sentry.0094_cleanup_unreferenced_event_files... OK | |
Applying sentry.0095_ruleactivity... OK | |
Applying sentry.0096_sentry_app_component_skip_load_on_open... OK | |
Applying sentry.0097_add_sentry_app_id_to_sentry_alertruletriggeraction... OK | |
Applying sentry.0098_add-performance-onboarding... OK | |
Applying sentry.0099_fix_project_platforms... OK | |
Applying sentry.0100_file_type_on_event_attachment... OK | |
Applying sentry.0101_backfill_file_type_on_event_attachment... OK | |
Applying sentry.0102_collect_relay_analytics... OK | |
Applying sentry.0103_project_has_alert_filters... OK | |
Applying sentry.0104_collect_relay_public_key_usage... OK | |
Applying sentry.0105_remove_nullability_of_event_attachment_type... OK | |
Applying sentry.0106_service_hook_project_id_nullable... OK | |
Applying sentry.0107_remove_spaces_from_slugs... OK | |
Applying sentry.0108_update_fileblob_action... OK | |
Applying sentry.0109_sentry_app_creator... OK | |
Applying sentry.0110_sentry_app_creator_backill... OK | |
Applying sentry.0111_snuba_query_event_type... OK | |
Applying sentry.0112_groupinboxmodel... OK | |
Applying sentry.0113_add_repositoryprojectpathconfig... OK | |
Applying sentry.0114_add_unhandled_savedsearch... OK | |
Applying sentry.0115_add_checksum_to_debug_file... OK | |
Applying sentry.0116_backfill_debug_file_checksum... OK | |
Applying sentry.0117_dummy-activityupdate... OK | |
Applying sentry.0118_backfill_snuba_query_event_types... OK | |
Applying sentry.0119_fix_set_none... OK | |
Applying sentry.0120_commit_author_charfield... OK | |
GroupInbox: 100% |# | ETA: --:--:-- | |
OK | |
Applying sentry.0122_add_release_status... OK | |
Applying sentry.0123_groupinbox_addprojandorg... OK | |
Applying sentry.0124_add_release_status_model... OK | |
Applying sentry.0125_add_platformexternalissue_project_id... OK | |
Applying sentry.0126_make_platformexternalissue_group_id_flexfk... OK | |
Applying sentry.0127_backfill_platformexternalissue_project_id... OK | |
Applying sentry.0128_change_dashboards... OK | |
Applying sentry.0129_remove_dashboard_keys... OK | |
Applying sentry.0130_remove_old_widget_models... OK | |
Applying sentry.0131_drop_widget_tables... OK | |
Applying sentry.0132_groupownermodel... OK | |
Applying sentry.0133_dashboard_delete_object_status... OK | |
Applying sentry.0134_dashboard_drop_object_status_column... OK | |
Applying sentry.0135_removinguniquegroupownerconstraint... OK | |
Applying sentry.0136_issue_alert_filter_all_orgs... OK | |
Applying sentry.0137_dashboard_widget_interval... OK | |
Applying sentry.0138_widget_query_remove_interval... OK | |
Applying sentry.0139_remove_widgetquery_interval... OK | |
Applying sentry.0140_subscription_checker... OK | |
Applying sentry.0141_remove_widget_constraints... OK | |
Applying sentry.0142_add_dashboard_tombstone... OK | |
Applying sentry.0143_add_alerts_integrationfeature... OK | |
Applying sentry.0144_add_publish_request_inprogress_status... OK | |
Applying sentry.0145_rename_alert_rule_feature... OK | |
Applying sentry.0146_backfill_members_alert_write... OK | |
Applying sentry.0147_add_groupinbox_date_added_index... OK | |
Applying sentry.0148_group_id_bigint... OK | |
Applying sentry.0149_bigint... OK | |
Applying sentry.0150_remove_userreport_eventattachment_constraints... OK | |
Applying sentry.0151_add_world_map_dashboard_widget_type... OK | |
Applying sentry.0152_remove_slack_workspace_orgintegrations... OK | |
Applying sentry.0153_add_big_number_dashboard_widget_type... OK | |
Applying sentry.0154_groupedmessage_inbox_sort... OK | |
Applying sentry.0155_add_dashboard_query_orderby... OK | |
Applying sentry.0156_add_mark_reviewed_activity... OK | |
Applying sentry.0157_make_repositoryprojectpathconfig_organization_integration_nullable... OK | |
Applying sentry.0158_create_externalteam_table... OK | |
Applying sentry.0159_create_externaluser_table... OK | |
Applying sentry.0160_create_projectcodeowners_table... OK | |
Applying sentry.0161_add_saved_search_sort... OK | |
Applying sentry.0162_backfill_saved_search_sort... OK | |
Applying sentry.0163_add_organizationmember_and_external_name... OK | |
Applying sentry.0164_add_protect_on_delete_codeowners... OK | |
Applying sentry.0165_metric_alerts_fix_group_ids... OK | |
Applying sentry.0166_create_notificationsetting_table... OK | |
Applying sentry.0167_rm_organization_integration_from_projectcodeowners... OK | |
Applying sentry.0168_demo_orgs_users... OK | |
Applying sentry.0169_delete_organization_integration_from_projectcodeowners... OK | |
Applying sentry.0170_actor_introduction... OK | |
Applying sentry.0171_backfill_actors... OK | |
Applying sessions.0001_initial... OK | |
Applying sites.0001_initial... OK | |
Applying sites.0002_alter_domain_unique... OK | |
Applying social_auth.0001_initial... OK | |
/usr/local/lib/python3.6/site-packages/sentry/receivers/onboarding.py:69: DeprecationWarning: The 'warn' method is deprecated, use 'warning' instead | |
project.organization_id, | |
04:28:22 [WARNING] sentry: Cannot initiate onboarding for organization (1) due to missing owners | |
Created internal Sentry project (slug=internal, id=1) | |
Would you like to create a user account now? [Y/n]: y | |
Email: | |
Password: | |
Repeat for confirmation: | |
Added to organization: sentry | |
User created: | |
Creating missing DSNs | |
Correcting Group.num_comments counter | |
▶ Migrating file storage ... | |
Unable to find image 'alpine:latest' locally | |
latest: Pulling from library/alpine | |
ba3557a56b15: Already exists | |
Digest: sha256:a75afd8b57e7f34e4dad8d65e2c7ba2e1975c795ce1ee22fa34f8cf46f96a3be | |
Status: Downloaded newer image for alpine:latest | |
▶ Generating Relay credentials ... | |
--no-ansi option is deprecated and will be removed in future versions. | |
Creating sentry_onpremise_relay_run ... | |
Creating sentry_onpremise_relay_run ... done | |
Relay credentials written to relay/credentials.json | |
▶ Setting up GeoIP integration ... | |
Setting up IP address geolocation ... | |
Installing (empty) IP address geolocation database ... done. | |
IP address geolocation is not configured for updates. | |
See https://develop.sentry.dev/self-hosted/geolocation/ for instructions. | |
Error setting up IP address geolocation. | |
----------------------------------------------------------------- | |
You're all done! Run the following command to get Sentry running: | |
docker-compose up -d | |
----------------------------------------------------------------- | |
root@server:/opt/sentry/onpremise# docker-compose up | |
Starting sentry_onpremise_redis_1 ... done | |
Starting sentry_onpremise_symbolicator_1 ... done | |
Starting sentry_onpremise_memcached_1 ... done | |
Starting sentry_onpremise_zookeeper_1 ... done | |
Starting sentry_onpremise_clickhouse_1 ... done | |
Starting sentry_onpremise_smtp_1 ... done | |
Starting sentry_onpremise_postgres_1 ... done | |
Creating sentry_onpremise_symbolicator-cleanup_1 ... done | |
Creating sentry_onpremise_geoipupdate_1 ... done | |
Starting sentry_onpremise_kafka_1 ... done | |
Starting sentry_onpremise_snuba-subscription-consumer-events_1 ... done | |
Starting sentry_onpremise_snuba-transactions-consumer_1 ... done | |
Starting sentry_onpremise_snuba-outcomes-consumer_1 ... done | |
Creating sentry_onpremise_relay_1 ... done | |
Starting sentry_onpremise_snuba-api_1 ... done | |
Starting sentry_onpremise_snuba-replacer_1 ... done | |
Starting sentry_onpremise_snuba-subscription-consumer-transactions_1 ... done | |
Starting sentry_onpremise_snuba-consumer_1 ... done | |
Starting sentry_onpremise_snuba-sessions-consumer_1 ... done | |
Creating sentry_onpremise_snuba-cleanup_1 ... done | |
Creating sentry_onpremise_cron_1 ... done | |
Creating sentry_onpremise_subscription-consumer-transactions_1 ... done | |
Creating sentry_onpremise_post-process-forwarder_1 ... done | |
Creating sentry_onpremise_ingest-consumer_1 ... done | |
Creating sentry_onpremise_subscription-consumer-events_1 ... done | |
Creating sentry_onpremise_worker_1 ... done | |
Creating sentry_onpremise_web_1 ... done | |
Creating sentry_onpremise_sentry-cleanup_1 ... done | |
Creating sentry_onpremise_nginx_1 ... done | |
Attaching to sentry_onpremise_memcached_1, sentry_onpremise_redis_1, sentry_onpremise_symbolicator_1, sentry_onpremise_clickhouse_1, sentry_onpremise_symbolicator-cleanup_1, sentry_onpremise_zookeeper_1, sentry_onpremise_postgres_1, sentry_onpremise_smtp_1, sentry_onpremise_geoipupdate_1, sentry_onpremise_kafka_1, sentry_onpremise_snuba-subscription-consumer-events_1, sentry_onpremise_snuba-transactions-consumer_1, sentry_onpremise_snuba-outcomes-consumer_1, sentry_onpremise_snuba-consumer_1, sentry_onpremise_snuba-cleanup_1, sentry_onpremise_snuba-sessions-consumer_1, sentry_onpremise_snuba-api_1, sentry_onpremise_snuba-subscription-consumer-transactions_1, sentry_onpremise_relay_1, sentry_onpremise_snuba-replacer_1, sentry_onpremise_subscription-consumer-transactions_1, sentry_onpremise_post-process-forwarder_1, sentry_onpremise_worker_1, sentry_onpremise_cron_1, sentry_onpremise_web_1, sentry_onpremise_sentry-cleanup_1, sentry_onpremise_subscription-consumer-events_1, sentry_onpremise_ingest-consumer_1, sentry_onpremise_nginx_1 | |
clickhouse_1 | Processing configuration file '/etc/clickhouse-server/config.xml'. | |
clickhouse_1 | Merging configuration file '/etc/clickhouse-server/config.d/docker_related_config.xml'. | |
clickhouse_1 | Merging configuration file '/etc/clickhouse-server/config.d/sentry.xml'. | |
clickhouse_1 | Include not found: clickhouse_remote_servers | |
clickhouse_1 | Include not found: clickhouse_compression | |
clickhouse_1 | Logging information to /var/log/clickhouse-server/clickhouse-server.log | |
clickhouse_1 | Logging errors to /var/log/clickhouse-server/clickhouse-server.err.log | |
clickhouse_1 | Logging information to console | |
clickhouse_1 | 2021.03.09 05:41:57.413208 [ 1 ] {} <Information> : Starting ClickHouse 20.3.9.70 with revision 54433 | |
clickhouse_1 | 2021.03.09 05:41:57.426941 [ 1 ] {} <Information> Application: starting up | |
clickhouse_1 | Include not found: networks | |
clickhouse_1 | 2021.03.09 05:41:57.473065 [ 1 ] {} <Information> Application: Uncompressed cache size was lowered to 7.82 GiB because the system has low amount of memory | |
clickhouse_1 | 2021.03.09 05:41:57.473315 [ 1 ] {} <Information> Application: Loading metadata from /var/lib/clickhouse/ | |
clickhouse_1 | 2021.03.09 05:41:57.489723 [ 1 ] {} <Information> DatabaseOrdinary (system): Total 1 tables and 0 dictionaries. | |
clickhouse_1 | 2021.03.09 05:41:57.513336 [ 44 ] {} <Information> BackgroundProcessingPool: Create BackgroundProcessingPool with 16 threads | |
clickhouse_1 | 2021.03.09 05:41:57.632238 [ 1 ] {} <Information> DatabaseOrdinary (system): Starting up tables. | |
clickhouse_1 | 2021.03.09 05:41:57.753388 [ 1 ] {} <Information> DatabaseOrdinary (default): Total 13 tables and 0 dictionaries. | |
clickhouse_1 | 2021.03.09 05:41:57.843073 [ 1 ] {} <Information> DatabaseOrdinary (default): Starting up tables. | |
clickhouse_1 | 2021.03.09 05:41:57.844386 [ 1 ] {} <Information> BackgroundSchedulePool: Create BackgroundSchedulePool with 16 threads | |
clickhouse_1 | 2021.03.09 05:41:57.845596 [ 1 ] {} <Information> Application: It looks like the process has no CAP_NET_ADMIN capability, 'taskstats' performance statistics will be disabled. It could happen due to incorrect ClickHouse package installation. You could resolve the problem manually with 'sudo setcap cap_net_admin=+ep /usr/bin/clickhouse'. Note that it will not work on 'nosuid' mounted filesystems. It also doesn't work if you run clickhouse-server inside network namespace as it happens in some containers. | |
clickhouse_1 | 2021.03.09 05:41:57.845634 [ 1 ] {} <Information> Application: It looks like the process has no CAP_SYS_NICE capability, the setting 'os_thread_nice' will have no effect. It could happen due to incorrect ClickHouse package installation. You could resolve the problem manually with 'sudo setcap cap_sys_nice=+ep /usr/bin/clickhouse'. Note that it will not work on 'nosuid' mounted filesystems. | |
clickhouse_1 | 2021.03.09 05:41:57.847402 [ 1 ] {} <Error> Application: Listen [::]:8123 failed: Poco::Exception. Code: 1000, e.code() = 0, e.displayText() = DNS error: EAI: -9 (version 20.3.9.70 (official build)). If it is an IPv6 or IPv4 address and your host has disabled IPv6 or IPv4, then consider to specify not disabled IPv4 or IPv6 address to listen in <listen_host> element of configuration file. Example for disabled IPv6: <listen_host>0.0.0.0</listen_host> . Example for disabled IPv4: <listen_host>::</listen_host> | |
clickhouse_1 | 2021.03.09 05:41:57.848102 [ 1 ] {} <Error> Application: Listen [::]:9000 failed: Poco::Exception. Code: 1000, e.code() = 0, e.displayText() = DNS error: EAI: -9 (version 20.3.9.70 (official build)). If it is an IPv6 or IPv4 address and your host has disabled IPv6 or IPv4, then consider to specify not disabled IPv4 or IPv6 address to listen in <listen_host> element of configuration file. Example for disabled IPv6: <listen_host>0.0.0.0</listen_host> . Example for disabled IPv4: <listen_host>::</listen_host> | |
clickhouse_1 | 2021.03.09 05:41:57.848315 [ 1 ] {} <Error> Application: Listen [::]:9009 failed: Poco::Exception. Code: 1000, e.code() = 0, e.displayText() = DNS error: EAI: -9 (version 20.3.9.70 (official build)). If it is an IPv6 or IPv4 address and your host has disabled IPv6 or IPv4, then consider to specify not disabled IPv4 or IPv6 address to listen in <listen_host> element of configuration file. Example for disabled IPv6: <listen_host>0.0.0.0</listen_host> . Example for disabled IPv4: <listen_host>::</listen_host> | |
clickhouse_1 | 2021.03.09 05:41:57.848527 [ 1 ] {} <Error> Application: Listen [::]:9004 failed: Poco::Exception. Code: 1000, e.code() = 0, e.displayText() = DNS error: EAI: -9 (version 20.3.9.70 (official build)). If it is an IPv6 or IPv4 address and your host has disabled IPv6 or IPv4, then consider to specify not disabled IPv4 or IPv6 address to listen in <listen_host> element of configuration file. Example for disabled IPv6: <listen_host>0.0.0.0</listen_host> . Example for disabled IPv4: <listen_host>::</listen_host> | |
clickhouse_1 | 2021.03.09 05:41:57.849663 [ 1 ] {} <Information> Application: Listening for http://0.0.0.0:8123 | |
clickhouse_1 | 2021.03.09 05:41:57.849714 [ 1 ] {} <Information> Application: Listening for connections with native protocol (tcp): 0.0.0.0:9000 | |
clickhouse_1 | 2021.03.09 05:41:57.849752 [ 1 ] {} <Information> Application: Listening for replica communication (interserver): http://0.0.0.0:9009 | |
clickhouse_1 | 2021.03.09 05:41:58.519559 [ 1 ] {} <Information> Application: Listening for MySQL compatibility protocol: 0.0.0.0:9004 | |
clickhouse_1 | 2021.03.09 05:41:58.520373 [ 1 ] {} <Information> Application: Available RAM: 15.64 GiB; physical cores: 1; logical cores: 2. | |
clickhouse_1 | 2021.03.09 05:41:58.520563 [ 1 ] {} <Information> Application: Ready for connections. | |
clickhouse_1 | Include not found: clickhouse_remote_servers | |
clickhouse_1 | Include not found: clickhouse_compression | |
geoipupdate_1 | error loading configuration file /sentry/GeoIP.conf: error opening file: open /sentry/GeoIP.conf: no such file or directory | |
sentry_onpremise_geoipupdate_1 exited with code 1 | |
kafka_1 | ===> ENV Variables ... | |
kafka_1 | ALLOW_UNSIGNED=false | |
kafka_1 | COMPONENT=kafka | |
kafka_1 | CONFLUENT_DEB_VERSION=1 | |
kafka_1 | CONFLUENT_PLATFORM_LABEL= | |
kafka_1 | CONFLUENT_SUPPORT_METRICS_ENABLE=false | |
kafka_1 | CONFLUENT_VERSION=5.5.0 | |
kafka_1 | CUB_CLASSPATH=/etc/confluent/docker/docker-utils.jar | |
kafka_1 | HOME=/root | |
kafka_1 | HOSTNAME=2a49effa25f6 | |
kafka_1 | KAFKA_ADVERTISED_LISTENERS=PLAINTEXT://kafka:9092 | |
kafka_1 | KAFKA_LOG4J_LOGGERS=kafka.cluster=WARN,kafka.controller=WARN,kafka.coordinator=WARN,kafka.log=WARN,kafka.server=WARN,kafka.zookeeper=WARN,state.change.logger=WARN | |
kafka_1 | KAFKA_LOG4J_ROOT_LOGLEVEL=WARN | |
kafka_1 | KAFKA_LOG_RETENTION_HOURS=24 | |
kafka_1 | KAFKA_MAX_REQUEST_SIZE=50000000 | |
kafka_1 | KAFKA_MESSAGE_MAX_BYTES=50000000 | |
kafka_1 | KAFKA_OFFSETS_TOPIC_NUM_PARTITIONS=1 | |
kafka_1 | KAFKA_OFFSETS_TOPIC_REPLICATION_FACTOR=1 | |
kafka_1 | KAFKA_TOOLS_LOG4J_LOGLEVEL=WARN | |
kafka_1 | KAFKA_VERSION= | |
kafka_1 | KAFKA_ZOOKEEPER_CONNECT=zookeeper:2181 | |
kafka_1 | LANG=C.UTF-8 | |
kafka_1 | PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin | |
kafka_1 | PWD=/ | |
kafka_1 | PYTHON_PIP_VERSION=8.1.2 | |
kafka_1 | PYTHON_VERSION=2.7.9-1 | |
kafka_1 | SCALA_VERSION=2.12 | |
kafka_1 | SHLVL=1 | |
kafka_1 | ZULU_OPENJDK_VERSION=8=8.38.0.13 | |
kafka_1 | _=/usr/bin/env | |
kafka_1 | ===> User | |
kafka_1 | uid=0(root) gid=0(root) groups=0(root) | |
kafka_1 | ===> Configuring ... | |
postgres_1 | | |
postgres_1 | PostgreSQL Database directory appears to contain a database; Skipping initialization | |
postgres_1 | | |
postgres_1 | LOG: database system was shut down at 2021-03-09 04:29:31 UTC | |
postgres_1 | LOG: MultiXact member wraparound protections are now enabled | |
postgres_1 | LOG: database system is ready to accept connections | |
postgres_1 | LOG: autovacuum launcher started | |
redis_1 | 1:C 09 Mar 2021 05:41:56.020 # oO0OoO0OoO0Oo Redis is starting oO0OoO0OoO0Oo | |
redis_1 | 1:C 09 Mar 2021 05:41:56.020 # Redis version=5.0.12, bits=64, commit=00000000, modified=0, pid=1, just started | |
redis_1 | 1:C 09 Mar 2021 05:41:56.020 # Warning: no config file specified, using the default config. In order to specify a config file use redis-server /path/to/redis.conf | |
redis_1 | 1:M 09 Mar 2021 05:41:56.021 * Running mode=standalone, port=6379. | |
redis_1 | 1:M 09 Mar 2021 05:41:56.021 # Server initialized | |
redis_1 | 1:M 09 Mar 2021 05:41:56.022 # WARNING overcommit_memory is set to 0! Background save may fail under low memory condition. To fix this issue add 'vm.overcommit_memory = 1' to /etc/sysctl.conf and then reboot or run the command 'sysctl vm.overcommit_memory=1' for this to take effect. | |
redis_1 | 1:M 09 Mar 2021 05:41:56.022 # WARNING you have Transparent Huge Pages (THP) support enabled in your kernel. This will create latency and memory usage issues with Redis. To fix this issue run the command 'echo never > /sys/kernel/mm/transparent_hugepage/enabled' as root, and add it to your /etc/rc.local in order to retain the setting after a reboot. Redis must be restarted after THP is disabled. | |
redis_1 | 1:M 09 Mar 2021 05:41:56.030 * DB loaded from disk: 0.008 seconds | |
redis_1 | 1:M 09 Mar 2021 05:41:56.030 * Ready to accept connections | |
relay_1 | 2021-03-09T05:42:06Z [rdkafka::client] ERROR: librdkafka: FAIL [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 14ms in state CONNECT) | |
relay_1 | 2021-03-09T05:42:06Z [rdkafka::client] ERROR: librdkafka: Global error: BrokerTransportFailure (Local: Broker transport failure): kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 14ms in state CONNECT) | |
relay_1 | 2021-03-09T05:42:06Z [rdkafka::client] ERROR: librdkafka: Global error: AllBrokersDown (Local: All broker connections are down): 1/1 brokers are down | |
relay_1 | 2021-03-09T05:42:06Z [rdkafka::client] ERROR: librdkafka: FAIL [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT) | |
relay_1 | 2021-03-09T05:42:06Z [rdkafka::client] ERROR: librdkafka: Global error: BrokerTransportFailure (Local: Broker transport failure): kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT) | |
relay_1 | 2021-03-09T05:42:06Z [rdkafka::client] ERROR: librdkafka: Global error: AllBrokersDown (Local: All broker connections are down): 1/1 brokers are down | |
relay_1 | 2021-03-09T05:42:07Z [actix::actors::resolver] WARN: Can not create system dns resolver: io error | |
relay_1 | 2021-03-09T05:42:07Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
relay_1 | 2021-03-09T05:42:07Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
relay_1 | 2021-03-09T05:42:07Z [rdkafka::client] ERROR: librdkafka: FAIL [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 1ms in state CONNECT, 1 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:07Z [rdkafka::client] ERROR: librdkafka: Global error: BrokerTransportFailure (Local: Broker transport failure): kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 1ms in state CONNECT, 1 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:07Z [rdkafka::client] ERROR: librdkafka: FAIL [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:07Z [rdkafka::client] ERROR: librdkafka: Global error: BrokerTransportFailure (Local: Broker transport failure): kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:08Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
relay_1 | 2021-03-09T05:42:09Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
relay_1 | 2021-03-09T05:42:12Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
sentry-cleanup_1 | SHELL=/bin/bash | |
sentry-cleanup_1 | BASH_ENV=/container.env | |
sentry-cleanup_1 | 0 0 * * * gosu sentry sentry cleanup --days 90 > /proc/1/fd/1 2>/proc/1/fd/2 | |
smtp_1 | + sed -ri ' | |
smtp_1 | s/^#?(dc_local_interfaces)=.*/\1='\''0.0.0.0 ; ::0'\''/; | |
smtp_1 | s/^#?(dc_other_hostnames)=.*/\1='\'''\''/; | |
smtp_1 | s/^#?(dc_relay_nets)=.*/\1='\''0.0.0.0\/0'\''/; | |
smtp_1 | s/^#?(dc_eximconfig_configtype)=.*/\1='\''internet'\''/; | |
smtp_1 | ' /etc/exim4/update-exim4.conf.conf | |
smtp_1 | + update-exim4.conf -v | |
smtp_1 | using non-split configuration scheme from /etc/exim4/exim4.conf.template | |
smtp_1 | 272 LOG: MAIN | |
smtp_1 | 272 exim 4.92 daemon started: pid=272, no queue runs, listening for SMTP on port 25 (IPv6 and IPv4) | |
snuba-api_1 | + '[' a = - ']' | |
snuba-api_1 | + snuba api --help | |
snuba-cleanup_1 | SHELL=/bin/bash | |
snuba-cleanup_1 | BASH_ENV=/container.env | |
snuba-cleanup_1 | */5 * * * * gosu snuba snuba cleanup --dry-run False > /proc/1/fd/1 2>/proc/1/fd/2 | |
snuba-consumer_1 | + '[' c = - ']' | |
snuba-consumer_1 | + snuba consumer --help | |
snuba-outcomes-consumer_1 | + '[' c = - ']' | |
snuba-outcomes-consumer_1 | + snuba consumer --help | |
snuba-replacer_1 | + '[' r = - ']' | |
snuba-replacer_1 | + snuba replacer --help | |
snuba-sessions-consumer_1 | + '[' c = - ']' | |
snuba-sessions-consumer_1 | + snuba consumer --help | |
snuba-subscription-consumer-events_1 | + '[' s = - ']' | |
snuba-subscription-consumer-events_1 | + snuba subscriptions --help | |
snuba-subscription-consumer-transactions_1 | + '[' s = - ']' | |
snuba-subscription-consumer-transactions_1 | + snuba subscriptions --help | |
snuba-transactions-consumer_1 | + '[' c = - ']' | |
snuba-transactions-consumer_1 | + snuba consumer --help | |
symbolicator-cleanup_1 | SHELL=/bin/bash | |
symbolicator-cleanup_1 | BASH_ENV=/container.env | |
symbolicator-cleanup_1 | 55 23 * * * gosu symbolicator symbolicator cleanup > /proc/1/fd/1 2>/proc/1/fd/2 | |
zookeeper_1 | ===> ENV Variables ... | |
zookeeper_1 | ALLOW_UNSIGNED=false | |
zookeeper_1 | COMPONENT=zookeeper | |
zookeeper_1 | CONFLUENT_DEB_VERSION=1 | |
zookeeper_1 | CONFLUENT_PLATFORM_LABEL= | |
zookeeper_1 | CONFLUENT_SUPPORT_METRICS_ENABLE=false | |
zookeeper_1 | CONFLUENT_VERSION=5.5.0 | |
zookeeper_1 | CUB_CLASSPATH=/etc/confluent/docker/docker-utils.jar | |
zookeeper_1 | HOME=/root | |
zookeeper_1 | HOSTNAME=b8d177c4e99b | |
zookeeper_1 | KAFKA_VERSION= | |
zookeeper_1 | LANG=C.UTF-8 | |
zookeeper_1 | PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin | |
zookeeper_1 | PWD=/ | |
zookeeper_1 | PYTHON_PIP_VERSION=8.1.2 | |
zookeeper_1 | PYTHON_VERSION=2.7.9-1 | |
zookeeper_1 | SCALA_VERSION=2.12 | |
zookeeper_1 | SHLVL=1 | |
zookeeper_1 | ZOOKEEPER_CLIENT_PORT=2181 | |
zookeeper_1 | ZOOKEEPER_LOG4J_ROOT_LOGLEVEL=WARN | |
zookeeper_1 | ZOOKEEPER_TOOLS_LOG4J_LOGLEVEL=WARN | |
zookeeper_1 | ZULU_OPENJDK_VERSION=8=8.38.0.13 | |
zookeeper_1 | _=/usr/bin/env | |
zookeeper_1 | ===> User | |
zookeeper_1 | uid=0(root) gid=0(root) groups=0(root) | |
zookeeper_1 | ===> Configuring ... | |
zookeeper_1 | ===> Running preflight checks ... | |
zookeeper_1 | ===> Check if /var/lib/zookeeper/data is writable ... | |
relay_1 | 2021-03-09T05:42:15Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
zookeeper_1 | ===> Check if /var/lib/zookeeper/log is writable ... | |
snuba-subscription-consumer-events_1 | + set -- snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-events-subscriptions-consumers --topic=events --result-topic=events-subscription-results --dataset=events --commit-log-topic=snuba-commit-log --commit-log-group=snuba-consumers --delay-seconds=60 --schedule-ttl=60 | |
snuba-subscription-consumer-events_1 | + set gosu snuba snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-events-subscriptions-consumers --topic=events --result-topic=events-subscription-results --dataset=events --commit-log-topic=snuba-commit-log --commit-log-group=snuba-consumers --delay-seconds=60 --schedule-ttl=60 | |
snuba-subscription-consumer-events_1 | + exec gosu snuba snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-events-subscriptions-consumers --topic=events --result-topic=events-subscription-results --dataset=events --commit-log-topic=snuba-commit-log --commit-log-group=snuba-consumers --delay-seconds=60 --schedule-ttl=60 | |
relay_1 | 2021-03-09T05:42:20Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 0ns | |
relay_1 | 2021-03-09T05:42:20Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
relay_1 | 2021-03-09T05:42:20Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 1s | |
snuba-transactions-consumer_1 | + set -- snuba consumer --storage transactions --consumer-group transactions_group --auto-offset-reset=latest --max-batch-time-ms 750 --commit-log-topic=snuba-commit-log | |
snuba-transactions-consumer_1 | + set gosu snuba snuba consumer --storage transactions --consumer-group transactions_group --auto-offset-reset=latest --max-batch-time-ms 750 --commit-log-topic=snuba-commit-log | |
snuba-transactions-consumer_1 | + exec gosu snuba snuba consumer --storage transactions --consumer-group transactions_group --auto-offset-reset=latest --max-batch-time-ms 750 --commit-log-topic=snuba-commit-log | |
relay_1 | 2021-03-09T05:42:21Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 1.5s | |
zookeeper_1 | ===> Launching ... | |
zookeeper_1 | ===> Launching zookeeper ... | |
snuba-outcomes-consumer_1 | + set -- snuba consumer --storage outcomes_raw --auto-offset-reset=earliest --max-batch-time-ms 750 | |
snuba-outcomes-consumer_1 | + set gosu snuba snuba consumer --storage outcomes_raw --auto-offset-reset=earliest --max-batch-time-ms 750 | |
snuba-outcomes-consumer_1 | + exec gosu snuba snuba consumer --storage outcomes_raw --auto-offset-reset=earliest --max-batch-time-ms 750 | |
snuba-consumer_1 | + set -- snuba consumer --storage events --auto-offset-reset=latest --max-batch-time-ms 750 | |
snuba-consumer_1 | + set gosu snuba snuba consumer --storage events --auto-offset-reset=latest --max-batch-time-ms 750 | |
snuba-consumer_1 | + exec gosu snuba snuba consumer --storage events --auto-offset-reset=latest --max-batch-time-ms 750 | |
relay_1 | 2021-03-09T05:42:22Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 2.25s | |
relay_1 | 2021-03-09T05:42:25Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 3.375s | |
snuba-replacer_1 | + set -- snuba replacer --storage events --auto-offset-reset=latest --max-batch-size 3 | |
snuba-replacer_1 | + set gosu snuba snuba replacer --storage events --auto-offset-reset=latest --max-batch-size 3 | |
snuba-replacer_1 | + exec gosu snuba snuba replacer --storage events --auto-offset-reset=latest --max-batch-size 3 | |
snuba-sessions-consumer_1 | + set -- snuba consumer --storage sessions_raw --auto-offset-reset=latest --max-batch-time-ms 750 | |
snuba-sessions-consumer_1 | + set gosu snuba snuba consumer --storage sessions_raw --auto-offset-reset=latest --max-batch-time-ms 750 | |
snuba-sessions-consumer_1 | + exec gosu snuba snuba consumer --storage sessions_raw --auto-offset-reset=latest --max-batch-time-ms 750 | |
snuba-api_1 | + set -- snuba api | |
snuba-api_1 | + set gosu snuba snuba api | |
snuba-api_1 | + exec gosu snuba snuba api | |
snuba-subscription-consumer-transactions_1 | + set -- snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-transactions-subscriptions-consumers --topic=events --result-topic=transactions-subscription-results --dataset=transactions --commit-log-topic=snuba-commit-log --commit-log-group=transactions_group --delay-seconds=60 --schedule-ttl=60 | |
snuba-subscription-consumer-transactions_1 | + set gosu snuba snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-transactions-subscriptions-consumers --topic=events --result-topic=transactions-subscription-results --dataset=transactions --commit-log-topic=snuba-commit-log --commit-log-group=transactions_group --delay-seconds=60 --schedule-ttl=60 | |
snuba-subscription-consumer-transactions_1 | + exec gosu snuba snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-transactions-subscriptions-consumers --topic=events --result-topic=transactions-subscription-results --dataset=transactions --commit-log-topic=snuba-commit-log --commit-log-group=transactions_group --delay-seconds=60 --schedule-ttl=60 | |
kafka_1 | ===> Running preflight checks ... | |
kafka_1 | ===> Check if /var/lib/kafka/data is writable ... | |
relay_1 | 2021-03-09T05:42:28Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
relay_1 | 2021-03-09T05:42:28Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 5.0625s | |
zookeeper_1 | [2021-03-09 05:42:29,280] WARN Either no config or no quorum defined in config, running in standalone mode (org.apache.zookeeper.server.quorum.QuorumPeerMain) | |
kafka_1 | ===> Check if Zookeeper is healthy ... | |
zookeeper_1 | [2021-03-09 05:42:33,252] WARN o.e.j.s.ServletContextHandler@4d95d2a2{/,null,UNAVAILABLE} contextPath ends with /* (org.eclipse.jetty.server.handler.ContextHandler) | |
zookeeper_1 | [2021-03-09 05:42:33,282] WARN Empty contextPath (org.eclipse.jetty.server.handler.ContextHandler) | |
relay_1 | 2021-03-09T05:42:33Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 7.59375s | |
snuba-subscription-consumer-events_1 | %3|1615268555.276|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 2ms in state CONNECT) | |
snuba-subscription-consumer-events_1 | %3|1615268556.266|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT) | |
snuba-subscription-consumer-events_1 | %3|1615268556.266|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-subscription-consumer-events_1 | %3|1615268557.265|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-transactions-consumer_1 | %3|1615268557.633|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 1ms in state CONNECT) | |
snuba-transactions-consumer_1 | %3|1615268557.659|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 8ms in state CONNECT) | |
relay_1 | 2021-03-09T05:42:37Z [rdkafka::client] ERROR: librdkafka: FAIL [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 30 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:37Z [rdkafka::client] ERROR: librdkafka: Global error: BrokerTransportFailure (Local: Broker transport failure): kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 30 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:37Z [rdkafka::client] ERROR: librdkafka: FAIL [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 30 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:37Z [rdkafka::client] ERROR: librdkafka: Global error: BrokerTransportFailure (Local: Broker transport failure): kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 30 identical error(s) suppressed) | |
snuba-consumer_1 | %3|1615268558.251|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 2ms in state CONNECT) | |
snuba-consumer_1 | %3|1615268558.255|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 23ms in state CONNECT) | |
snuba-outcomes-consumer_1 | %3|1615268558.265|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 7ms in state CONNECT) | |
snuba-outcomes-consumer_1 | %3|1615268558.266|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 19ms in state CONNECT) | |
snuba-transactions-consumer_1 | %3|1615268558.631|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-transactions-consumer_1 | %3|1615268558.634|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:zookeeper.version=3.5.7-f0fdd52973d373ffd9c86b81d99842dc2c7f660e, built on 02/10/2020 11:30 GMT | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:host.name=2a49effa25f6 | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:java.version=1.8.0_212 | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:java.vendor=Azul Systems, Inc. | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:java.home=/usr/lib/jvm/zulu-8-amd64/jre | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:java.class.path=/etc/confluent/docker/docker-utils.jar | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:java.library.path=/usr/java/packages/lib/amd64:/usr/lib64:/lib64:/lib:/usr/lib | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:java.io.tmpdir=/tmp | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:java.compiler=<NA> | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:os.name=Linux | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:os.arch=amd64 | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:os.version=5.4.0-1037-gcp | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:user.name=root | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:user.home=/root | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:user.dir=/ | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:os.memory.free=237MB | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:os.memory.max=3559MB | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Client environment:os.memory.total=241MB | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Initiating client connection, connectString=zookeeper:2181 sessionTimeout=40000 watcher=io.confluent.admin.utils.ZookeeperConnectionWatcher@cc34f4d | |
snuba-consumer_1 | %3|1615268559.248|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-outcomes-consumer_1 | %3|1615268559.249|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-outcomes-consumer_1 | %3|1615268559.249|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-consumer_1 | %3|1615268559.250|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
kafka_1 | [main] INFO org.apache.zookeeper.common.X509Util - Setting -D jdk.tls.rejectClientInitiatedRenegotiation=true to disable client-initiated TLS renegotiation | |
kafka_1 | [main] INFO org.apache.zookeeper.ClientCnxnSocket - jute.maxbuffer value is 4194304 Bytes | |
kafka_1 | [main] INFO org.apache.zookeeper.ClientCnxn - zookeeper.request.timeout value is 0. feature enabled= | |
relay_1 | 2021-03-09T05:42:39Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
kafka_1 | [main-SendThread(zookeeper:2181)] INFO org.apache.zookeeper.ClientCnxn - Opening socket connection to server zookeeper/172.18.0.6:2181. Will not attempt to authenticate using SASL (unknown error) | |
snuba-replacer_1 | %3|1615268560.226|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 7ms in state CONNECT) | |
kafka_1 | [main-SendThread(zookeeper:2181)] INFO org.apache.zookeeper.ClientCnxn - Socket connection established, initiating session, client: /172.18.0.11:49766, server: zookeeper/172.18.0.6:2181 | |
kafka_1 | [main-SendThread(zookeeper:2181)] INFO org.apache.zookeeper.ClientCnxn - Session establishment complete on server zookeeper/172.18.0.6:2181, sessionid = 0x100049a0a470000, negotiated timeout = 40000 | |
snuba-api_1 | *** Starting uWSGI 2.0.18 (64bit) on [Tue Mar 9 05:42:40 2021] *** | |
snuba-api_1 | compiled with version: 8.3.0 on 23 February 2021 15:53:08 | |
snuba-api_1 | os: Linux-5.4.0-1037-gcp #40-Ubuntu SMP Fri Feb 5 11:57:53 UTC 2021 | |
snuba-api_1 | nodename: 39c09f6125f6 | |
snuba-api_1 | machine: x86_64 | |
snuba-api_1 | clock source: unix | |
snuba-api_1 | pcre jit disabled | |
snuba-api_1 | detected number of CPU cores: 2 | |
snuba-api_1 | current working directory: /usr/src/snuba | |
snuba-api_1 | detected binary path: /usr/local/bin/uwsgi | |
snuba-api_1 | your memory page size is 4096 bytes | |
snuba-api_1 | detected max file descriptor number: 1048576 | |
snuba-api_1 | lock engine: pthread robust mutexes | |
snuba-api_1 | thunder lock: enabled | |
snuba-api_1 | uwsgi socket 0 bound to TCP address 0.0.0.0:1218 fd 3 | |
snuba-api_1 | Python version: 3.8.8 (default, Feb 19 2021, 18:07:06) [GCC 8.3.0] | |
snuba-api_1 | Set PythonHome to /usr/local | |
snuba-api_1 | Python main interpreter initialized at 0x557a1d981be0 | |
snuba-sessions-consumer_1 | %3|1615268560.672|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 8ms in state CONNECT) | |
snuba-sessions-consumer_1 | %3|1615268560.673|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 2ms in state CONNECT) | |
kafka_1 | [main] INFO org.apache.zookeeper.ZooKeeper - Session: 0x100049a0a470000 closed | |
kafka_1 | [main-EventThread] INFO org.apache.zookeeper.ClientCnxn - EventThread shut down for session: 0x100049a0a470000 | |
snuba-api_1 | python threads support enabled | |
snuba-api_1 | your server socket listen backlog is limited to 100 connections | |
snuba-api_1 | your mercy for graceful operations on workers is 60 seconds | |
snuba-api_1 | mapped 145808 bytes (142 KB) for 1 cores | |
snuba-api_1 | *** Operational MODE: single process *** | |
snuba-api_1 | initialized 38 metrics | |
snuba-api_1 | spawned uWSGI master process (pid: 1) | |
snuba-api_1 | spawned uWSGI worker 1 (pid: 14, cores: 1) | |
snuba-api_1 | metrics collector thread started | |
kafka_1 | ===> Launching ... | |
kafka_1 | ===> Launching kafka ... | |
snuba-replacer_1 | %3|1615268561.208|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:41Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 11.390625s | |
snuba-sessions-consumer_1 | %3|1615268561.666|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-sessions-consumer_1 | %3|1615268561.669|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-subscription-consumer-transactions_1 | %3|1615268561.701|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 13ms in state CONNECT) | |
snuba-subscription-consumer-transactions_1 | %3|1615268562.687|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT) | |
snuba-subscription-consumer-transactions_1 | %3|1615268562.687|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
snuba-subscription-consumer-transactions_1 | %3|1615268563.687|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
kafka_1 | [2021-03-09 05:42:45,249] INFO Registered kafka:type=kafka.Log4jController MBean (kafka.utils.Log4jControllerRegistration$) | |
snuba-api_1 | WSGI app 0 (mountpoint='') ready in 6 seconds on interpreter 0x557a1d981be0 pid: 14 (default app) | |
kafka_1 | [2021-03-09 05:42:49,960] WARN The package io.confluent.support.metrics.collectors.FullCollector for collecting the full set of support metrics could not be loaded, so we are reverting to anonymous, basic metric collection. If you are a Confluent customer, please refer to the Confluent Platform documentation, section Proactive Support, on how to activate full metrics collection. (io.confluent.support.metrics.KafkaSupportConfig) | |
kafka_1 | [2021-03-09 05:42:49,967] WARN The support metrics collection feature ("Metrics") of Proactive Support is disabled. (io.confluent.support.metrics.SupportedServerStartable) | |
subscription-consumer-transactions_1 | 05:42:52 [INFO] sentry.plugins.github: apps-not-configured | |
post-process-forwarder_1 | 05:42:52 [INFO] sentry.plugins.github: apps-not-configured | |
web_1 | 05:42:52 [INFO] sentry.plugins.github: apps-not-configured | |
subscription-consumer-events_1 | 05:42:52 [INFO] sentry.plugins.github: apps-not-configured | |
worker_1 | 05:42:52 [INFO] sentry.plugins.github: apps-not-configured | |
ingest-consumer_1 | 05:42:52 [INFO] sentry.plugins.github: apps-not-configured | |
cron_1 | 05:42:52 [INFO] sentry.plugins.github: apps-not-configured | |
relay_1 | 2021-03-09T05:42:52Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 17.0859375s | |
subscription-consumer-transactions_1 | %3|1615268574.107|FAIL|rdkafka#producer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 4ms in state CONNECT) | |
post-process-forwarder_1 | %3|1615268574.184|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT) | |
post-process-forwarder_1 | %3|1615268574.186|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 2ms in state CONNECT) | |
subscription-consumer-events_1 | %3|1615268574.293|FAIL|rdkafka#producer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT) | |
worker_1 | 05:42:54 [INFO] sentry.bgtasks: bgtask.spawn (task_name='sentry.bgtasks.clean_dsymcache:clean_dsymcache') | |
web_1 | *** Starting uWSGI 2.0.18 (64bit) on [Tue Mar 9 05:42:54 2021] *** | |
web_1 | compiled with version: 8.3.0 on 08 March 2021 18:55:51 | |
web_1 | os: Linux-5.4.0-1037-gcp #40-Ubuntu SMP Fri Feb 5 11:57:53 UTC 2021 | |
web_1 | nodename: 27393629cb97 | |
web_1 | machine: x86_64 | |
web_1 | clock source: unix | |
web_1 | detected number of CPU cores: 2 | |
web_1 | current working directory: / | |
web_1 | detected binary path: /usr/local/bin/uwsgi | |
web_1 | !!! no internal routing support, rebuild with pcre support !!! | |
web_1 | your memory page size is 4096 bytes | |
web_1 | detected max file descriptor number: 1048576 | |
web_1 | lock engine: pthread robust mutexes | |
web_1 | thunder lock: enabled | |
web_1 | uWSGI http bound on 0.0.0.0:9000 fd 4 | |
web_1 | uwsgi socket 0 bound to TCP address 127.0.0.1:41627 (port auto-assigned) fd 3 | |
web_1 | Python version: 3.6.13 (default, Feb 16 2021, 20:33:02) [GCC 8.3.0] | |
web_1 | Set PythonHome to /usr/local | |
worker_1 | 05:42:54 [INFO] sentry.bgtasks: bgtask.spawn (task_name='sentry.bgtasks.clean_releasefilecache:clean_releasefilecache') | |
worker_1 | * Unknown config option found: 'slack.legacy-app' | |
web_1 | Python main interpreter initialized at 0x5640c2025ef0 | |
web_1 | python threads support enabled | |
web_1 | your server socket listen backlog is limited to 100 connections | |
web_1 | your mercy for graceful operations on workers is 60 seconds | |
web_1 | setting request body buffering size to 65536 bytes | |
web_1 | mapped 1924224 bytes (1879 KB) for 12 cores | |
web_1 | *** Operational MODE: preforking+threaded *** | |
web_1 | spawned uWSGI master process (pid: 19) | |
web_1 | spawned uWSGI worker 1 (pid: 23, cores: 4) | |
web_1 | spawned uWSGI worker 2 (pid: 24, cores: 4) | |
web_1 | spawned uWSGI worker 3 (pid: 25, cores: 4) | |
web_1 | spawned uWSGI http 1 (pid: 26) | |
cron_1 | * Unknown config option found: 'slack.legacy-app' | |
ingest-consumer_1 | %3|1615268574.643|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 3ms in state CONNECT) | |
subscription-consumer-transactions_1 | %3|1615268575.103|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT) | |
subscription-consumer-transactions_1 | %3|1615268575.108|FAIL|rdkafka#producer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
post-process-forwarder_1 | %3|1615268575.182|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
post-process-forwarder_1 | %3|1615268575.184|FAIL|rdkafka#consumer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
subscription-consumer-events_1 | %3|1615268575.293|FAIL|rdkafka#producer-2| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
subscription-consumer-events_1 | %3|1615268575.293|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT) | |
kafka_1 | [2021-03-09 05:42:55,405] INFO Starting the log cleaner (kafka.log.LogCleaner) | |
ingest-consumer_1 | %3|1615268575.639|FAIL|rdkafka#producer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
kafka_1 | [2021-03-09 05:42:55,735] INFO [kafka-log-cleaner-thread-0]: Starting (kafka.log.LogCleaner) | |
subscription-consumer-transactions_1 | %3|1615268576.104|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
subscription-consumer-events_1 | %3|1615268576.293|FAIL|rdkafka#consumer-1| [thrd:kafka:9092/bootstrap]: kafka:9092/bootstrap: Connect to ipv4#172.18.0.11:9092 failed: Connection refused (after 0ms in state CONNECT, 1 identical error(s) suppressed) | |
relay_1 | 2021-03-09T05:42:56Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
kafka_1 | [2021-03-09 05:42:57,283] INFO Awaiting socket connections on 0.0.0.0:9092. (kafka.network.Acceptor) | |
kafka_1 | [2021-03-09 05:42:57,436] INFO [SocketServer brokerId=1001] Created data-plane acceptor and processors for endpoint : EndPoint(0.0.0.0,9092,ListenerName(PLAINTEXT),PLAINTEXT) (kafka.network.SocketServer) | |
kafka_1 | [2021-03-09 05:42:57,491] INFO [SocketServer brokerId=1001] Started 1 acceptor threads for data-plane (kafka.network.SocketServer) | |
kafka_1 | [2021-03-09 05:42:57,881] INFO Creating /brokers/ids/1001 (is it secure? false) (kafka.zk.KafkaZkClient) | |
kafka_1 | [2021-03-09 05:42:57,930] INFO Stat of the created znode at /brokers/ids/1001 is: 109,109,1615268577915,1615268577915,1,0,0,72062653681827841,180,0,109 | |
kafka_1 | (kafka.zk.KafkaZkClient) | |
kafka_1 | [2021-03-09 05:42:57,931] INFO Registered broker 1001 at path /brokers/ids/1001 with addresses: ArrayBuffer(EndPoint(kafka,9092,ListenerName(PLAINTEXT),PLAINTEXT)), czxid (broker epoch): 109 (kafka.zk.KafkaZkClient) | |
worker_1 | | |
worker_1 | -------------- celery@7afd6977fee3 v4.4.7 (cliffs) | |
worker_1 | --- ***** ----- | |
worker_1 | -- ******* ---- Linux-5.4.0-1037-gcp-x86_64-with-debian-10.8 2021-03-09 05:42:57 | |
worker_1 | - *** --- * --- | |
worker_1 | - ** ---------- [config] | |
worker_1 | - ** ---------- .> app: sentry:0x7f4ebf9c8b38 | |
worker_1 | - ** ---------- .> transport: redis://redis:6379/0 | |
worker_1 | - ** ---------- .> results: disabled:// | |
worker_1 | - *** --- * --- .> concurrency: 2 (prefork) | |
worker_1 | -- ******* ---- .> task events: OFF (enable -E to monitor tasks in this worker) | |
worker_1 | --- ***** ----- | |
worker_1 | -------------- [queues] | |
worker_1 | .> activity.notify exchange=default(direct) key=activity.notify | |
worker_1 | .> alerts exchange=default(direct) key=alerts | |
worker_1 | .> app_platform exchange=default(direct) key=app_platform | |
worker_1 | .> assemble exchange=default(direct) key=assemble | |
worker_1 | .> auth exchange=default(direct) key=auth | |
worker_1 | .> buffers.process_pending exchange=default(direct) key=buffers.process_pending | |
worker_1 | .> cleanup exchange=default(direct) key=cleanup | |
worker_1 | .> commits exchange=default(direct) key=commits | |
worker_1 | .> counters-0 exchange=counters(direct) key=default | |
worker_1 | .> data_export exchange=default(direct) key=data_export | |
worker_1 | .> default exchange=default(direct) key=default | |
worker_1 | .> digests.delivery exchange=default(direct) key=digests.delivery | |
worker_1 | .> digests.scheduling exchange=default(direct) key=digests.scheduling | |
worker_1 | .> email exchange=default(direct) key=email | |
worker_1 | .> events.preprocess_event exchange=default(direct) key=events.preprocess_event | |
worker_1 | .> events.process_event exchange=default(direct) key=events.process_event | |
worker_1 | .> events.reprocess_events exchange=default(direct) key=events.reprocess_events | |
worker_1 | .> events.reprocessing.preprocess_event exchange=default(direct) key=events.reprocessing.preprocess_event | |
worker_1 | .> events.reprocessing.process_event exchange=default(direct) key=events.reprocessing.process_event | |
worker_1 | .> events.reprocessing.symbolicate_event exchange=default(direct) key=events.reprocessing.symbolicate_event | |
worker_1 | .> events.save_event exchange=default(direct) key=events.save_event | |
worker_1 | .> events.symbolicate_event exchange=default(direct) key=events.symbolicate_event | |
worker_1 | .> files.delete exchange=default(direct) key=files.delete | |
worker_1 | .> group_owners.process_suspect_commits exchange=default(direct) key=group_owners.process_suspect_commits | |
worker_1 | .> incident_snapshots exchange=default(direct) key=incident_snapshots | |
worker_1 | .> incidents exchange=default(direct) key=incidents | |
worker_1 | .> integrations exchange=default(direct) key=integrations | |
worker_1 | .> merge exchange=default(direct) key=merge | |
worker_1 | .> options exchange=default(direct) key=options | |
worker_1 | .> relay_config exchange=default(direct) key=relay_config | |
worker_1 | .> reports.deliver exchange=default(direct) key=reports.deliver | |
worker_1 | .> reports.prepare exchange=default(direct) key=reports.prepare | |
worker_1 | .> search exchange=default(direct) key=search | |
worker_1 | .> sleep exchange=default(direct) key=sleep | |
worker_1 | .> stats exchange=default(direct) key=stats | |
worker_1 | .> subscriptions exchange=default(direct) key=subscriptions | |
worker_1 | .> triggers-0 exchange=triggers(direct) key=default | |
worker_1 | .> unmerge exchange=default(direct) key=unmerge | |
worker_1 | .> update exchange=default(direct) key=update | |
worker_1 | | |
kafka_1 | [2021-03-09 05:42:59,164] INFO [/config/changes-event-process-thread]: Starting (kafka.common.ZkNodeChangeNotificationListener$ChangeEventProcessThread) | |
kafka_1 | [2021-03-09 05:42:59,329] INFO [SocketServer brokerId=1001] Started data-plane processors for 1 acceptors (kafka.network.SocketServer) | |
kafka_1 | [2021-03-09 05:42:59,714] INFO Creating topic transactions-subscription-results with configuration {} and initial partition assignment Map(0 -> ArrayBuffer(1001)) (kafka.zk.AdminZkClient) | |
kafka_1 | [2021-03-09 05:42:59,716] INFO Creating topic events-subscription-results with configuration {} and initial partition assignment Map(0 -> ArrayBuffer(1001)) (kafka.zk.AdminZkClient) | |
ingest-consumer_1 | 05:42:59 [WARNING] batching-kafka-consumer: Topic 'ingest-events' or its partitions are not ready, retrying... | |
subscription-consumer-events_1 | 05:42:59 [WARNING] batching-kafka-consumer: Topic 'events-subscription-results' or its partitions are not ready, retrying... | |
ingest-consumer_1 | 05:42:59 [WARNING] batching-kafka-consumer: Topic 'ingest-events' or its partitions are not ready, retrying... | |
subscription-consumer-transactions_1 | 05:42:59 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-events_1 | 05:42:59 [WARNING] batching-kafka-consumer: Topic 'events-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-transactions_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
ingest-consumer_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'ingest-events' or its partitions are not ready, retrying... | |
subscription-consumer-transactions_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
ingest-consumer_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'ingest-events' or its partitions are not ready, retrying... | |
subscription-consumer-events_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'events-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-transactions_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-events_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'events-subscription-results' or its partitions are not ready, retrying... | |
snuba-replacer_1 | 2021-03-09 05:43:00,314 Caught ConsumerError('KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"}'), shutting down... | |
post-process-forwarder_1 | * Unknown config option found: 'slack.legacy-app' | |
snuba-consumer_1 | 2021-03-09 05:43:00,391 Caught ConsumerError('KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"}'), shutting down... | |
snuba-replacer_1 | Traceback (most recent call last): | |
snuba-replacer_1 | File "/usr/local/bin/snuba", line 33, in <module> | |
snuba-replacer_1 | sys.exit(load_entry_point('snuba', 'console_scripts', 'snuba')()) | |
snuba-replacer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 829, in __call__ | |
snuba-replacer_1 | return self.main(*args, **kwargs) | |
snuba-replacer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 782, in main | |
snuba-replacer_1 | rv = self.invoke(ctx) | |
snuba-replacer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1259, in invoke | |
snuba-replacer_1 | return _process_result(sub_ctx.command.invoke(sub_ctx)) | |
snuba-replacer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1066, in invoke | |
snuba-replacer_1 | return ctx.invoke(self.callback, **ctx.params) | |
snuba-replacer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke | |
snuba-replacer_1 | return callback(*args, **kwargs) | |
snuba-replacer_1 | File "/usr/src/snuba/snuba/cli/replacer.py", line 133, in replacer | |
snuba-replacer_1 | replacer.run() | |
snuba-replacer_1 | File "/usr/src/snuba/snuba/utils/streams/processing/processor.py", line 112, in run | |
snuba-replacer_1 | self._run_once() | |
snuba-replacer_1 | File "/usr/src/snuba/snuba/utils/streams/processing/processor.py", line 142, in _run_once | |
snuba-replacer_1 | self.__message = self.__consumer.poll(timeout=1.0) | |
snuba-replacer_1 | File "/usr/src/snuba/snuba/utils/streams/backends/kafka.py", line 404, in poll | |
snuba-replacer_1 | raise ConsumerError(str(error)) | |
snuba-replacer_1 | snuba.utils.streams.backends.abstract.ConsumerError: KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"} | |
snuba-consumer_1 | Traceback (most recent call last): | |
snuba-consumer_1 | File "/usr/local/bin/snuba", line 33, in <module> | |
snuba-consumer_1 | sys.exit(load_entry_point('snuba', 'console_scripts', 'snuba')()) | |
snuba-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 829, in __call__ | |
snuba-consumer_1 | return self.main(*args, **kwargs) | |
snuba-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 782, in main | |
snuba-consumer_1 | rv = self.invoke(ctx) | |
snuba-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1259, in invoke | |
snuba-consumer_1 | return _process_result(sub_ctx.command.invoke(sub_ctx)) | |
snuba-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1066, in invoke | |
snuba-consumer_1 | return ctx.invoke(self.callback, **ctx.params) | |
snuba-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke | |
snuba-consumer_1 | return callback(*args, **kwargs) | |
snuba-consumer_1 | File "/usr/src/snuba/snuba/cli/consumer.py", line 161, in consumer | |
snuba-consumer_1 | consumer.run() | |
snuba-consumer_1 | File "/usr/src/snuba/snuba/utils/streams/processing/processor.py", line 112, in run | |
snuba-consumer_1 | self._run_once() | |
snuba-consumer_1 | File "/usr/src/snuba/snuba/utils/streams/processing/processor.py", line 142, in _run_once | |
snuba-consumer_1 | self.__message = self.__consumer.poll(timeout=1.0) | |
snuba-consumer_1 | File "/usr/src/snuba/snuba/utils/streams/backends/kafka.py", line 767, in poll | |
snuba-consumer_1 | return super().poll(timeout) | |
snuba-consumer_1 | File "/usr/src/snuba/snuba/utils/streams/backends/kafka.py", line 404, in poll | |
snuba-consumer_1 | raise ConsumerError(str(error)) | |
snuba-consumer_1 | snuba.utils.streams.backends.abstract.ConsumerError: KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"} | |
post-process-forwarder_1 | Traceback (most recent call last): | |
post-process-forwarder_1 | File "/usr/local/bin/sentry", line 8, in <module> | |
post-process-forwarder_1 | sys.exit(main()) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/sentry/runner/__init__.py", line 164, in main | |
post-process-forwarder_1 | cli(prog_name=get_prog(), obj={}, max_content_width=100) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 829, in __call__ | |
post-process-forwarder_1 | return self.main(*args, **kwargs) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 782, in main | |
post-process-forwarder_1 | rv = self.invoke(ctx) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1259, in invoke | |
post-process-forwarder_1 | return _process_result(sub_ctx.command.invoke(sub_ctx)) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1259, in invoke | |
post-process-forwarder_1 | return _process_result(sub_ctx.command.invoke(sub_ctx)) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1066, in invoke | |
post-process-forwarder_1 | return ctx.invoke(self.callback, **ctx.params) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 610, in invoke | |
post-process-forwarder_1 | return callback(*args, **kwargs) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/decorators.py", line 21, in new_func | |
post-process-forwarder_1 | return f(get_current_context(), *args, **kwargs) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/sentry/runner/decorators.py", line 66, in inner | |
post-process-forwarder_1 | return ctx.invoke(f, *args, **kwargs) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 610, in invoke | |
post-process-forwarder_1 | return callback(*args, **kwargs) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/decorators.py", line 21, in new_func | |
post-process-forwarder_1 | return f(get_current_context(), *args, **kwargs) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/sentry/runner/decorators.py", line 28, in inner | |
post-process-forwarder_1 | return ctx.invoke(f, *args, **kwargs) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 610, in invoke | |
post-process-forwarder_1 | return callback(*args, **kwargs) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/sentry/runner/commands/run.py", line 333, in post_process_forwarder | |
post-process-forwarder_1 | initial_offset_reset=options["initial_offset_reset"], | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/sentry/utils/services.py", line 102, in <lambda> | |
post-process-forwarder_1 | context[key] = (lambda f: lambda *a, **k: getattr(self, f)(*a, **k))(key) | |
post-process-forwarder_1 | File "/usr/local/lib/python3.6/site-packages/sentry/eventstream/kafka/backend.py", line 204, in run_post_process_forwarder | |
post-process-forwarder_1 | raise Exception(error) | |
post-process-forwarder_1 | Exception: KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"} | |
snuba-outcomes-consumer_1 | 2021-03-09 05:43:00,443 Caught ConsumerError('KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"}'), shutting down... | |
snuba-outcomes-consumer_1 | Traceback (most recent call last): | |
snuba-outcomes-consumer_1 | File "/usr/local/bin/snuba", line 33, in <module> | |
snuba-outcomes-consumer_1 | sys.exit(load_entry_point('snuba', 'console_scripts', 'snuba')()) | |
snuba-outcomes-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 829, in __call__ | |
snuba-outcomes-consumer_1 | return self.main(*args, **kwargs) | |
snuba-outcomes-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 782, in main | |
snuba-outcomes-consumer_1 | rv = self.invoke(ctx) | |
snuba-outcomes-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1259, in invoke | |
snuba-outcomes-consumer_1 | return _process_result(sub_ctx.command.invoke(sub_ctx)) | |
snuba-outcomes-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1066, in invoke | |
snuba-outcomes-consumer_1 | return ctx.invoke(self.callback, **ctx.params) | |
snuba-outcomes-consumer_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke | |
snuba-outcomes-consumer_1 | return callback(*args, **kwargs) | |
snuba-outcomes-consumer_1 | File "/usr/src/snuba/snuba/cli/consumer.py", line 161, in consumer | |
snuba-outcomes-consumer_1 | consumer.run() | |
snuba-outcomes-consumer_1 | File "/usr/src/snuba/snuba/utils/streams/processing/processor.py", line 112, in run | |
snuba-outcomes-consumer_1 | self._run_once() | |
snuba-outcomes-consumer_1 | File "/usr/src/snuba/snuba/utils/streams/processing/processor.py", line 142, in _run_once | |
snuba-outcomes-consumer_1 | self.__message = self.__consumer.poll(timeout=1.0) | |
snuba-outcomes-consumer_1 | File "/usr/src/snuba/snuba/utils/streams/backends/kafka.py", line 404, in poll | |
snuba-outcomes-consumer_1 | raise ConsumerError(str(error)) | |
snuba-outcomes-consumer_1 | snuba.utils.streams.backends.abstract.ConsumerError: KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"} | |
subscription-consumer-transactions_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-events_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'events-subscription-results' or its partitions are not ready, retrying... | |
ingest-consumer_1 | * Unknown config option found: 'slack.legacy-app' | |
ingest-consumer_1 | /usr/local/lib/python3.6/site-packages/sentry/utils/batching_kafka_consumer.py:51: DeprecationWarning: The 'warn' method is deprecated, use 'warning' instead | |
ingest-consumer_1 | logger.warn("Topic '%s' or its partitions are not ready, retrying...", topic) | |
snuba-subscription-consumer-events_1 | Traceback (most recent call last): | |
snuba-subscription-consumer-events_1 | File "/usr/local/bin/snuba", line 33, in <module> | |
snuba-subscription-consumer-events_1 | sys.exit(load_entry_point('snuba', 'console_scripts', 'snuba')()) | |
snuba-subscription-consumer-events_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 829, in __call__ | |
snuba-subscription-consumer-events_1 | return self.main(*args, **kwargs) | |
snuba-subscription-consumer-events_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 782, in main | |
snuba-subscription-consumer-events_1 | rv = self.invoke(ctx) | |
snuba-subscription-consumer-events_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1259, in invoke | |
snuba-subscription-consumer-events_1 | return _process_result(sub_ctx.command.invoke(sub_ctx)) | |
snuba-subscription-consumer-events_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 1066, in invoke | |
snuba-subscription-consumer-events_1 | return ctx.invoke(self.callback, **ctx.params) | |
snuba-subscription-consumer-events_1 | File "/usr/local/lib/python3.8/site-packages/click/core.py", line 610, in invoke | |
snuba-subscription-consumer-events_1 | return callback(*args, **kwargs) | |
snuba-subscription-consumer-events_1 | File "/usr/src/snuba/snuba/cli/subscriptions.py", line 138, in subscriptions | |
snuba-subscription-consumer-events_1 | SynchronizedConsumer( | |
snuba-subscription-consumer-events_1 | File "/usr/src/snuba/snuba/utils/streams/synchronized.py", line 106, in __init__ | |
snuba-subscription-consumer-events_1 | self.__commit_log_worker.result() | |
snuba-subscription-consumer-events_1 | File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 432, in result | |
snuba-subscription-consumer-events_1 | return self.__get_result() | |
snuba-subscription-consumer-events_1 | File "/usr/local/lib/python3.8/concurrent/futures/_base.py", line 388, in __get_result | |
snuba-subscription-consumer-events_1 | raise self._exception | |
snuba-subscription-consumer-events_1 | File "/usr/src/snuba/snuba/utils/concurrent.py", line 33, in run | |
snuba-subscription-consumer-events_1 | result = function() | |
snuba-subscription-consumer-events_1 | File "/usr/src/snuba/snuba/utils/streams/synchronized.py", line 130, in __run_commit_log_worker | |
snuba-subscription-consumer-events_1 | message = self.__commit_log_consumer.poll(0.1) | |
snuba-subscription-consumer-events_1 | File "/usr/src/snuba/snuba/utils/streams/backends/kafka.py", line 404, in poll | |
snuba-subscription-consumer-events_1 | raise ConsumerError(str(error)) | |
snuba-subscription-consumer-events_1 | snuba.utils.streams.backends.abstract.ConsumerError: KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"} | |
subscription-consumer-events_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'events-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-transactions_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
ingest-consumer_1 | Traceback (most recent call last): | |
ingest-consumer_1 | File "/usr/local/bin/sentry", line 8, in <module> | |
ingest-consumer_1 | sys.exit(main()) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/sentry/runner/__init__.py", line 164, in main | |
ingest-consumer_1 | cli(prog_name=get_prog(), obj={}, max_content_width=100) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 829, in __call__ | |
ingest-consumer_1 | return self.main(*args, **kwargs) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 782, in main | |
ingest-consumer_1 | rv = self.invoke(ctx) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1259, in invoke | |
ingest-consumer_1 | return _process_result(sub_ctx.command.invoke(sub_ctx)) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1259, in invoke | |
ingest-consumer_1 | return _process_result(sub_ctx.command.invoke(sub_ctx)) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 1066, in invoke | |
ingest-consumer_1 | return ctx.invoke(self.callback, **ctx.params) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 610, in invoke | |
ingest-consumer_1 | return callback(*args, **kwargs) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/decorators.py", line 21, in new_func | |
ingest-consumer_1 | return f(get_current_context(), *args, **kwargs) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/sentry/runner/decorators.py", line 66, in inner | |
ingest-consumer_1 | return ctx.invoke(f, *args, **kwargs) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 610, in invoke | |
ingest-consumer_1 | return callback(*args, **kwargs) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/decorators.py", line 21, in new_func | |
ingest-consumer_1 | return f(get_current_context(), *args, **kwargs) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/sentry/runner/decorators.py", line 28, in inner | |
ingest-consumer_1 | return ctx.invoke(f, *args, **kwargs) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/click/core.py", line 610, in invoke | |
ingest-consumer_1 | return callback(*args, **kwargs) | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/sentry/runner/commands/run.py", line 486, in ingest_consumer | |
ingest-consumer_1 | get_ingest_consumer(consumer_types=consumer_types, **options).run() | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/sentry/utils/batching_kafka_consumer.py", line 257, in run | |
ingest-consumer_1 | self._run_once() | |
ingest-consumer_1 | File "/usr/local/lib/python3.6/site-packages/sentry/utils/batching_kafka_consumer.py", line 275, in _run_once | |
ingest-consumer_1 | raise Exception(msg.error()) | |
ingest-consumer_1 | Exception: KafkaError{code=COORDINATOR_LOAD_IN_PROGRESS,val=14,str="JoinGroup failed: Broker: Coordinator load in progress"} | |
subscription-consumer-transactions_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-events_1 | 05:43:00 [WARNING] batching-kafka-consumer: Topic 'events-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-transactions_1 | 05:43:01 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
subscription-consumer-transactions_1 | 05:43:01 [WARNING] batching-kafka-consumer: Topic 'transactions-subscription-results' or its partitions are not ready, retrying... | |
snuba-transactions-consumer_1 | 2021-03-09 05:43:05,039 New partitions assigned: {Partition(topic=Topic(name='events'), index=0): 0} | |
snuba-sessions-consumer_1 | 2021-03-09 05:43:05,041 New partitions assigned: {Partition(topic=Topic(name='ingest-sessions'), index=0): 0} | |
subscription-consumer-events_1 | 05:43:07 [INFO] sentry.snuba.query_subscription_consumer: query-subscription-consumer.on_assign (offsets='{0: None}' partitions='[TopicPartition{topic=events-subscription-results,partition=0,offset=-1001,error=None}]') | |
subscription-consumer-transactions_1 | 05:43:07 [INFO] sentry.snuba.query_subscription_consumer: query-subscription-consumer.on_assign (offsets='{0: None}' partitions='[TopicPartition{topic=transactions-subscription-results,partition=0,offset=-1001,error=None}]') | |
snuba-subscription-consumer-transactions_1 | 2021-03-09 05:43:08,239 New partitions assigned: {Partition(topic=Topic(name='events'), index=0): 0} | |
relay_1 | 2021-03-09T05:43:09Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 25.62890625s | |
snuba-outcomes-consumer_1 | + set -- snuba consumer --storage outcomes_raw --auto-offset-reset=earliest --max-batch-time-ms 750 | |
snuba-outcomes-consumer_1 | + set gosu snuba snuba consumer --storage outcomes_raw --auto-offset-reset=earliest --max-batch-time-ms 750 | |
snuba-outcomes-consumer_1 | + exec gosu snuba snuba consumer --storage outcomes_raw --auto-offset-reset=earliest --max-batch-time-ms 750 | |
snuba-replacer_1 | + set -- snuba replacer --storage events --auto-offset-reset=latest --max-batch-size 3 | |
snuba-replacer_1 | + set gosu snuba snuba replacer --storage events --auto-offset-reset=latest --max-batch-size 3 | |
snuba-replacer_1 | + exec gosu snuba snuba replacer --storage events --auto-offset-reset=latest --max-batch-size 3 | |
snuba-subscription-consumer-events_1 | + set -- snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-events-subscriptions-consumers --topic=events --result-topic=events-subscription-results --dataset=events --commit-log-topic=snuba-commit-log --commit-log-group=snuba-consumers --delay-seconds=60 --schedule-ttl=60 | |
snuba-subscription-consumer-events_1 | + set gosu snuba snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-events-subscriptions-consumers --topic=events --result-topic=events-subscription-results --dataset=events --commit-log-topic=snuba-commit-log --commit-log-group=snuba-consumers --delay-seconds=60 --schedule-ttl=60 | |
snuba-subscription-consumer-events_1 | + exec gosu snuba snuba subscriptions --auto-offset-reset=latest --consumer-group=snuba-events-subscriptions-consumers --topic=events --result-topic=events-subscription-results --dataset=events --commit-log-topic=snuba-commit-log --commit-log-group=snuba-consumers --delay-seconds=60 --schedule-ttl=60 | |
snuba-consumer_1 | + set -- snuba consumer --storage events --auto-offset-reset=latest --max-batch-time-ms 750 | |
snuba-consumer_1 | + set gosu snuba snuba consumer --storage events --auto-offset-reset=latest --max-batch-time-ms 750 | |
snuba-consumer_1 | + exec gosu snuba snuba consumer --storage events --auto-offset-reset=latest --max-batch-time-ms 750 | |
relay_1 | 2021-03-09T05:43:22Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
snuba-sessions-consumer_1 | 2021-03-09 05:43:23,049 Partitions revoked: [Partition(topic=Topic(name='ingest-sessions'), index=0)] | |
snuba-outcomes-consumer_1 | 2021-03-09 05:43:23,081 New partitions assigned: {Partition(topic=Topic(name='outcomes'), index=0): 0} | |
snuba-consumer_1 | 2021-03-09 05:43:23,083 New partitions assigned: {Partition(topic=Topic(name='events'), index=0): 0} | |
snuba-sessions-consumer_1 | 2021-03-09 05:43:23,479 New partitions assigned: {Partition(topic=Topic(name='ingest-sessions'), index=0): 0} | |
snuba-replacer_1 | 2021-03-09 05:43:23,990 New partitions assigned: {Partition(topic=Topic(name='event-replacements'), index=0): 0} | |
post-process-forwarder_1 | 05:43:24 [INFO] sentry.plugins.github: apps-not-configured | |
ingest-consumer_1 | 05:43:24 [INFO] sentry.plugins.github: apps-not-configured | |
web_1 | 05:43:27 [INFO] sentry.plugins.github: apps-not-configured | |
web_1 | 05:43:27 [INFO] sentry.plugins.github: apps-not-configured | |
snuba-subscription-consumer-events_1 | 2021-03-09 05:43:27,499 New partitions assigned: {Partition(topic=Topic(name='events'), index=0): 0} | |
web_1 | 05:43:27 [INFO] sentry.plugins.github: apps-not-configured | |
web_1 | * Unknown config option found: 'slack.legacy-app' | |
web_1 | * Unknown config option found: 'slack.legacy-app' | |
web_1 | * Unknown config option found: 'slack.legacy-app' | |
web_1 | WSGI app 0 (mountpoint='') ready in 34 seconds on interpreter 0x5640c2025ef0 pid: 25 (default app) | |
web_1 | WSGI app 0 (mountpoint='') ready in 34 seconds on interpreter 0x5640c2025ef0 pid: 24 (default app) | |
web_1 | WSGI app 0 (mountpoint='') ready in 34 seconds on interpreter 0x5640c2025ef0 pid: 23 (default app) | |
post-process-forwarder_1 | 05:43:28 [INFO] sentry.eventstream.kafka.backend: Received partition assignment: [TopicPartition{topic=events,partition=0,offset=-1001,error=None}] | |
ingest-consumer_1 | 05:43:28 [INFO] batching-kafka-consumer: New partitions assigned: [TopicPartition{topic=ingest-attachments,partition=0,offset=-1001,error=None}, TopicPartition{topic=ingest-events,partition=0,offset=-1001,error=None}, TopicPartition{topic=ingest-transactions,partition=0,offset=-1001,error=None}] | |
relay_1 | 2021-03-09T05:43:35Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 38.443359375s | |
relay_1 | 2021-03-09T05:44:00Z [relay_server::actors::upstream] ERROR: authentication encountered error: could not send request to upstream | |
relay_1 | caused by: could not send request using reqwest | |
relay_1 | caused by: error sending request for url (http://web:9000/api/0/relays/register/challenge/): error trying to connect: dns error: no record found for name: web.google.internal. type: AAAA class: IN | |
relay_1 | 2021-03-09T05:44:13Z [relay_server::actors::upstream] WARN: Network outage, scheduling another check in 57.665039062s |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment