forked from codecov/worker
-
Notifications
You must be signed in to change notification settings - Fork 0
/
celery_config.py
180 lines (154 loc) · 6.83 KB
/
celery_config.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
# http://docs.celeryq.org/en/latest/configuration.html#configuration
import gc
import logging
import logging.config
import os
from datetime import timedelta
from celery import signals
from celery.beat import BeatLazyFunc
from celery.schedules import crontab
from celery.signals import worker_process_init
from codecovopentelem import (
CoverageSpanFilter,
UnableToStartProcessorException,
get_codecov_opentelemetry_instances,
)
from shared.celery_config import (
BaseCeleryConfig,
brolly_stats_rollup_task_name,
gh_app_webhook_check_task_name,
health_check_task_name,
profiling_finding_task_name,
)
from shared.config import get_config
from shared.helpers.cache import RedisBackend
from celery_task_router import route_task
from helpers.cache import cache
from helpers.clock import get_utc_now_as_iso_format
from helpers.environment import is_enterprise
from helpers.health_check import get_health_check_interval_seconds
from helpers.version import get_current_version
from services.redis import get_redis_connection
log = logging.getLogger(__name__)
@signals.worker_before_create_process.connect
def prefork_gc_freeze(**kwargs) -> None:
# This comes from https://github.com/getsentry/sentry/pull/63001
# More info https://www.youtube.com/watch?v=Hgw_RlCaIds
# The idea is to save memory in the worker subprocesses
# By freezing all the stuff we can just read from
gc.freeze()
@signals.setup_logging.connect
def initialize_logging(loglevel=logging.INFO, **kwargs):
celery_logger = logging.getLogger("celery")
celery_logger.setLevel(loglevel)
log.info("Initialized celery logging")
return celery_logger
@signals.worker_process_init.connect
def initialize_cache(**kwargs):
log.info("Initialized cache")
redis_cache_backend = RedisBackend(get_redis_connection())
cache.configure(redis_cache_backend)
@worker_process_init.connect(weak=False)
def init_celery_tracing(*args, **kwargs):
if (
os.getenv("OPENTELEMETRY_ENDPOINT")
and os.getenv("OPENTELEMETRY_TOKEN")
and os.getenv("OPENTELEMETRY_CODECOV_RATE")
and not is_enterprise()
):
from opentelemetry import trace
from opentelemetry.instrumentation.celery import CeleryInstrumentor
from opentelemetry.sdk.trace import TracerProvider
from opentelemetry.sdk.trace.export import BatchSpanProcessor
log.info("Configuring opentelemetry exporter")
provider = TracerProvider()
trace.set_tracer_provider(provider)
export_rate = float(os.getenv("OPENTELEMETRY_CODECOV_RATE"))
current_version = get_current_version()
current_env = "production"
try:
generator, exporter = get_codecov_opentelemetry_instances(
repository_token=os.getenv("OPENTELEMETRY_TOKEN"),
version_identifier=current_version,
sample_rate=export_rate,
filters={
CoverageSpanFilter.regex_name_filter: None,
CoverageSpanFilter.span_kind_filter: [
trace.SpanKind.SERVER,
trace.SpanKind.CONSUMER,
],
},
code=f"{current_version}:{current_env}",
untracked_export_rate=export_rate,
codecov_endpoint=os.getenv("OPENTELEMETRY_ENDPOINT"),
environment=current_env,
)
provider.add_span_processor(generator)
provider.add_span_processor(BatchSpanProcessor(exporter))
CeleryInstrumentor().instrument()
except UnableToStartProcessorException:
log.warning("Unable to instrument opentelemetry on worker")
hourly_check_task_name = "app.cron.hourly_check.HourlyCheckTask"
daily_plan_manager_task_name = "app.cron.daily.PlanManagerTask"
# Backfill GH Apps
backfill_existing_gh_app_installations_name = "app.tasks.backfill_existing_gh_app_installations.BackfillExistingGHAppInstallationsTask"
backfill_existing_individual_gh_app_installation_name = "app.tasks.backfill_existing_individual_gh_app_installation.BackfillExistingIndividualGHAppInstallationTask"
backfill_owners_without_gh_app_installations_name = "app.tasks.backfill_owners_without_gh_app_installations.BackfillOwnersWithoutGHAppInstallationsTask"
backfill_owners_without_gh_app_installation_individual_name = "app.tasks.backfill_owners_without_gh_app_installation_individual.BackfillOwnersWithoutGHAppInstallationIndividualTask"
trial_expiration_task_name = "app.tasks.plan.TrialExpirationTask"
trial_expiration_cron_task_name = "app.cron.plan.TrialExpirationCronTask"
update_branches_task_name = "app.cron.branches.UpdateBranchesTask"
update_branches_task_name = "app.cron.test_instances.BackfillTestInstancesTask"
def _beat_schedule():
beat_schedule = {
"hourly_check": {
"task": hourly_check_task_name,
"schedule": crontab(minute="0"),
"kwargs": {
"cron_task_generation_time_iso": BeatLazyFunc(get_utc_now_as_iso_format)
},
},
"github_app_webhooks_task": {
"task": gh_app_webhook_check_task_name,
"schedule": crontab(minute="0", hour="0,6,12,18"),
"kwargs": {
"cron_task_generation_time_iso": BeatLazyFunc(get_utc_now_as_iso_format)
},
},
"trial_expiration_cron": {
"task": trial_expiration_cron_task_name,
# 4 UTC is 12am EDT
"schedule": crontab(minute="0", hour="4"),
"kwargs": {
"cron_task_generation_time_iso": BeatLazyFunc(get_utc_now_as_iso_format)
},
},
}
if get_config("setup", "find_uncollected_profilings", "enabled", default=True):
beat_schedule["find_uncollected_profilings"] = {
"task": profiling_finding_task_name,
"schedule": crontab(minute="0,15,30,45"),
"kwargs": {
"cron_task_generation_time_iso": BeatLazyFunc(get_utc_now_as_iso_format)
},
}
if get_config("setup", "health_check", "enabled", default=False):
beat_schedule["health_check_task"] = {
"task": health_check_task_name,
"schedule": timedelta(seconds=get_health_check_interval_seconds()),
"kwargs": {
"cron_task_generation_time_iso": BeatLazyFunc(get_utc_now_as_iso_format)
},
}
if get_config("setup", "telemetry", "enabled", default=True):
beat_schedule["brolly_stats_rollup"] = {
"task": brolly_stats_rollup_task_name,
"schedule": crontab(minute="0", hour="2"),
"kwargs": {
"cron_task_generation_time_iso": BeatLazyFunc(get_utc_now_as_iso_format)
},
}
return beat_schedule
class CeleryWorkerConfig(BaseCeleryConfig):
beat_schedule = _beat_schedule()
task_routes = route_task