-
-
Notifications
You must be signed in to change notification settings - Fork 181
/
Copy pathconnections.py
395 lines (341 loc) · 15.3 KB
/
connections.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
import asyncio
import functools
import logging
from dataclasses import dataclass
from datetime import datetime, timedelta
from operator import attrgetter
from typing import TYPE_CHECKING, Any, Callable, List, Optional, Tuple, TypeVar, Union
from urllib.parse import parse_qs, urlparse
from uuid import uuid4
from redis.asyncio import ConnectionPool, Redis
from redis.asyncio.cluster import ClusterPipeline, PipelineCommand, RedisCluster # type: ignore
from redis.asyncio.sentinel import Sentinel
from redis.exceptions import RedisError, WatchError
from redis.typing import EncodableT, KeyT
from .constants import default_queue_name, expires_extra_ms, job_key_prefix, result_key_prefix
from .jobs import Deserializer, Job, JobDef, JobResult, Serializer, deserialize_job, serialize_job
from .utils import timestamp_ms, to_ms, to_unix_ms
logger = logging.getLogger('arq.connections')
logging.basicConfig(level=logging.DEBUG)
_KeyT = TypeVar('_KeyT', bound=KeyT)
@dataclass
class RedisSettings:
"""
No-Op class used to hold redis connection redis_settings.
Used by :func:`arq.connections.create_pool` and :class:`arq.worker.Worker`.
"""
host: Union[str, List[Tuple[str, int]]] = 'test-cluster.aqtke6.clustercfg.use2.cache.amazonaws.com'
port: int = 6379
unix_socket_path: Optional[str] = None
database: int = 0
username: Optional[str] = None
password: Optional[str] = None
ssl: bool = False
ssl_keyfile: Optional[str] = None
ssl_certfile: Optional[str] = None
ssl_cert_reqs: str = 'required'
ssl_ca_certs: Optional[str] = None
ssl_ca_data: Optional[str] = None
ssl_check_hostname: bool = False
conn_timeout: int = 1
conn_retries: int = 5
conn_retry_delay: int = 1
cluster_mode: bool = True
sentinel: bool = False
sentinel_master: str = 'mymaster'
@classmethod
def from_dsn(cls, dsn: str) -> 'RedisSettings':
conf = urlparse(dsn)
assert conf.scheme in {'redis', 'rediss', 'unix'}, 'invalid DSN scheme'
query_db = parse_qs(conf.query).get('db')
if query_db:
# e.g. redis://localhost:6379?db=1
database = int(query_db[0])
else:
database = int(conf.path.lstrip('/')) if conf.path else 0
return RedisSettings(
host=conf.hostname or 'localhost',
port=conf.port or 6379,
ssl=conf.scheme == 'rediss',
username=conf.username,
password=conf.password,
database=database,
unix_socket_path=conf.path if conf.scheme == 'unix' else None,
)
def __repr__(self) -> str:
return 'RedisSettings({})'.format(', '.join(f'{k}={v!r}' for k, v in self.__dict__.items()))
if TYPE_CHECKING:
BaseRedis = Redis[bytes]
else:
BaseRedis = Redis
class ArqRedis(BaseRedis):
"""
Thin subclass of ``redis.asyncio.Redis`` which adds :func:`arq.connections.enqueue_job`.
:param redis_settings: an instance of ``arq.connections.RedisSettings``.
:param job_serializer: a function that serializes Python objects to bytes, defaults to pickle.dumps
:param job_deserializer: a function that deserializes bytes into Python objects, defaults to pickle.loads
:param default_queue_name: the default queue name to use, defaults to ``arq.queue``.
:param expires_extra_ms: the default length of time from when a job is expected to start
after which the job expires, defaults to 1 day in ms.
:param kwargs: keyword arguments directly passed to ``redis.asyncio.Redis``.
"""
def __init__(
self,
pool_or_conn: Optional[ConnectionPool] = None,
job_serializer: Optional[Serializer] = None,
job_deserializer: Optional[Deserializer] = None,
default_queue_name: str = default_queue_name,
expires_extra_ms: int = expires_extra_ms,
**kwargs: Any,
) -> None:
self.job_serializer = job_serializer
self.job_deserializer = job_deserializer
self.default_queue_name = default_queue_name
if pool_or_conn:
kwargs['connection_pool'] = pool_or_conn
self.expires_extra_ms = expires_extra_ms
super().__init__(**kwargs)
async def enqueue_job(
self,
function: str,
*args: Any,
_job_id: Optional[str] = None,
_queue_name: Optional[str] = None,
_defer_until: Optional[datetime] = None,
_defer_by: Union[None, int, float, timedelta] = None,
_expires: Union[None, int, float, timedelta] = None,
_job_try: Optional[int] = None,
**kwargs: Any,
) -> Optional[Job]:
"""
Enqueue a job.
:param function: Name of the function to call
:param args: args to pass to the function
:param _job_id: ID of the job, can be used to enforce job uniqueness
:param _queue_name: queue of the job, can be used to create job in different queue
:param _defer_until: datetime at which to run the job
:param _defer_by: duration to wait before running the job
:param _expires: do not start or retry a job after this duration;
defaults to 24 hours plus deferring time, if any
:param _job_try: useful when re-enqueueing jobs within a job
:param kwargs: any keyword arguments to pass to the function
:return: :class:`arq.jobs.Job` instance or ``None`` if a job with this ID already exists
"""
if _queue_name is None:
_queue_name = self.default_queue_name
job_id = _job_id or uuid4().hex
job_key = job_key_prefix + job_id
assert not (_defer_until and _defer_by), "use either 'defer_until' or 'defer_by' or neither, not both"
defer_by_ms = to_ms(_defer_by)
expires_ms = to_ms(_expires)
async with self.pipeline(transaction=True) as pipe:
await pipe.watch(job_key)
if await pipe.exists(job_key, result_key_prefix + job_id):
await pipe.reset()
return None
enqueue_time_ms = timestamp_ms()
if _defer_until is not None:
score = to_unix_ms(_defer_until)
elif defer_by_ms:
score = enqueue_time_ms + defer_by_ms
else:
score = enqueue_time_ms
expires_ms = expires_ms or score - enqueue_time_ms + self.expires_extra_ms
job = serialize_job(function, args, kwargs, _job_try, enqueue_time_ms, serializer=self.job_serializer)
pipe.multi()
pipe.psetex(job_key, expires_ms, job) # type: ignore[no-untyped-call]
pipe.zadd(_queue_name, {job_id: score}) # type: ignore[unused-coroutine]
try:
await pipe.execute()
except WatchError:
# job got enqueued since we checked 'job_exists'
return None
the_job = Job(job_id, redis=self, _queue_name=_queue_name, _deserializer=self.job_deserializer)
logger.debug(the_job)
return the_job
async def _get_job_result(self, key: bytes) -> JobResult:
job_id = key[len(result_key_prefix) :].decode()
job = Job(job_id, self, _deserializer=self.job_deserializer)
r = await job.result_info()
if r is None:
raise KeyError(f'job "{key.decode()}" not found')
r.job_id = job_id
return r
async def all_job_results(self) -> List[JobResult]:
"""
Get results for all jobs in redis.
"""
keys = await self.keys(result_key_prefix + '*')
results = await asyncio.gather(*[self._get_job_result(k) for k in keys])
return sorted(results, key=attrgetter('enqueue_time'))
async def _get_job_def(self, job_id: bytes, score: int) -> JobDef:
key = job_key_prefix + job_id.decode()
v = await self.get(key)
assert v is not None, f'job "{key}" not found'
jd = deserialize_job(v, deserializer=self.job_deserializer)
jd.score = score
return jd
async def queued_jobs(self, *, queue_name: Optional[str] = None) -> List[JobDef]:
"""
Get information about queued, mostly useful when testing.
"""
if queue_name is None:
queue_name = self.default_queue_name
jobs = await self.zrange(queue_name, withscores=True, start=0, end=-1)
return await asyncio.gather(*[self._get_job_def(job_id, int(score)) for job_id, score in jobs])
class ArqRedisCluster(RedisCluster): # type: ignore
"""
Thin subclass of ``from redis.asyncio.cluster.RedisCluster`` which patches methods of RedisClusterPipeline
to support redis cluster`.
:param redis_settings: an instance of ``arq.connections.RedisSettings``.
:param job_serializer: a function that serializes Python objects to bytes, defaults to pickle.dumps
:param job_deserializer: a function that deserializes bytes into Python objects, defaults to pickle.loads
:param default_queue_name: the default queue name to use, defaults to ``arq.queue``.
:param expires_extra_ms: the default length of time from when a job is expected to start
after which the job expires, defaults to 1 day in ms.
:param kwargs: keyword arguments directly passed to ``from redis.asyncio.cluster.RedisCluster``.
"""
def __init__(
self,
job_serializer: Optional[Serializer] = None,
job_deserializer: Optional[Deserializer] = None,
default_queue_name: str = default_queue_name,
expires_extra_ms: int = expires_extra_ms,
**kwargs: Any,
) -> None:
self.job_serializer = job_serializer
self.job_deserializer = job_deserializer
self.default_queue_name = default_queue_name
self.expires_extra_ms = expires_extra_ms
super().__init__(**kwargs)
enqueue_job = ArqRedis.enqueue_job
_get_job_result = ArqRedis._get_job_result
all_job_results = ArqRedis.all_job_results
_get_job_def = ArqRedis._get_job_def
queued_jobs = ArqRedis.queued_jobs
def pipeline(self, transaction: Any | None = None, shard_hint: Any | None = None) -> ClusterPipeline:
return ArqRedisClusterPipeline(self)
class ArqRedisClusterPipeline(ClusterPipeline): # type: ignore
def __init__(self, client: RedisCluster) -> None:
self.watching = False
super().__init__(client)
async def watch(self, *names: KeyT) -> None:
self.watching = True
def multi(self) -> None:
self.watching = False
def execute_command(self, *args: Union[KeyT, EncodableT], **kwargs: Any) -> 'ClusterPipeline':
cmd = PipelineCommand(len(self._command_stack), *args, **kwargs)
if self.watching:
return self.immediate_execute_command(cmd)
self._command_stack.append(cmd)
return self
async def immediate_execute_command(self, cmd: PipelineCommand) -> Any:
try:
return await self._client.execute_command(*cmd.args, **cmd.kwargs)
except Exception as e:
cmd.result = e
def _split_command_across_slots(self, command: str, *keys: KeyT) -> 'ClusterPipeline':
for slot_keys in self._client._partition_keys_by_slot(keys).values():
if self.watching:
return self.execute_command(command, *slot_keys)
return self
async def create_pool(
settings_: RedisSettings = None,
*,
retry: int = 0,
job_serializer: Optional[Serializer] = None,
job_deserializer: Optional[Deserializer] = None,
default_queue_name: str = default_queue_name,
expires_extra_ms: int = expires_extra_ms,
) -> ArqRedis:
"""
Create a new redis pool, retrying up to ``conn_retries`` times if the connection fails.
Returns a :class:`arq.connections.ArqRedis` instance or :class: `arq.connections.ArqRedisCluster` depending on
whether `cluster_mode` flag is enabled in `RedisSettings`, thus allowing job enqueuing.
"""
settings: RedisSettings = RedisSettings() if settings_ is None else settings_
assert not (
type(settings.host) is str and settings.sentinel
), "str provided for 'host' but 'sentinel' is true; list of sentinels expected"
if settings.sentinel:
def pool_factory(*args: Any, **kwargs: Any) -> ArqRedis:
client = Sentinel( # type: ignore[misc]
*args,
sentinels=settings.host,
ssl=settings.ssl,
**kwargs,
)
return client.master_for(settings.sentinel_master, redis_class=ArqRedis)
elif settings.cluster_mode:
pool_factory = functools.partial(
ArqRedisCluster,
host=settings.host,
port=settings.port,
socket_connect_timeout=settings.conn_timeout,
ssl=settings.ssl,
ssl_keyfile=settings.ssl_keyfile,
ssl_certfile=settings.ssl_certfile,
ssl_cert_reqs=settings.ssl_cert_reqs,
ssl_ca_certs=settings.ssl_ca_certs,
ssl_ca_data=settings.ssl_ca_data,
ssl_check_hostname=settings.ssl_check_hostname,
)
else:
pool_factory = functools.partial(
ArqRedis,
db=settings.database,
username=settings.username,
host=settings.host,
port=settings.port,
unix_socket_path=settings.unix_socket_path,
socket_connect_timeout=settings.conn_timeout,
ssl=settings.ssl,
ssl_keyfile=settings.ssl_keyfile,
ssl_certfile=settings.ssl_certfile,
ssl_cert_reqs=settings.ssl_cert_reqs,
ssl_ca_certs=settings.ssl_ca_certs,
ssl_ca_data=settings.ssl_ca_data,
ssl_check_hostname=settings.ssl_check_hostname,
)
while True:
try:
pool = await pool_factory(password=settings.password, encoding='utf8')
pool.job_serializer = job_serializer
pool.job_deserializer = job_deserializer
pool.default_queue_name = default_queue_name
pool.expires_extra_ms = expires_extra_ms
except (ConnectionError, OSError, RedisError, asyncio.TimeoutError) as e:
if retry < settings.conn_retries:
logger.warning(
'redis connection error %s:%s %s %s, %d retries remaining...',
settings.host,
settings.port,
e.__class__.__name__,
e,
settings.conn_retries - retry,
)
await asyncio.sleep(settings.conn_retry_delay)
retry = retry + 1
else:
raise
else:
if retry > 0:
logger.info('redis connection successful')
return pool
# TODO
async def log_redis_info(redis: 'Redis[bytes]', log_func: Callable[[str], Any]) -> None:
async with redis.pipeline() as pipe:
pipe.info(section='Server') # type: ignore[unused-coroutine]
pipe.info(section='Memory') # type: ignore[unused-coroutine]
pipe.info(section='Clients') # type: ignore[unused-coroutine]
pipe.dbsize() # type: ignore[unused-coroutine]
info_server, info_memory, info_clients, key_count = await pipe.execute()
redis_version = info_server.get('redis_version', '?')
mem_usage = info_memory.get('used_memory_human', '?')
clients_connected = info_clients.get('connected_clients', '?')
log_func(
f'redis_version={redis_version} '
f'mem_usage={mem_usage} '
f'clients_connected={clients_connected} '
f'db_keys={88}'
)