1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
|
# R overlay -- dependency resolution, dependency resolver
# -*- coding: utf-8 -*-
# Copyright (C) 2012 André Erdmann <dywi@mailerd.de>
# Distributed under the terms of the GNU General Public License;
# either version 2 of the License, or (at your option) any later version.
"""dependency resolver"""
__all__ = [ 'DependencyResolver', ]
import logging
import threading
try:
import queue
except ImportError:
# python 2
import Queue as queue
from roverlay import config
from roverlay.depres import communication, deptype, events
import roverlay.depres.simpledeprule.reader
import roverlay.depres.simpledeprule.dynpool
# if false: do not use the "negative" result caching which stores
# unresolvable deps in a set for should-be faster lookups
USING_DEPRES_CACHE = True
# if True: verify that channels are unique for a resolver instance
SAFE_CHANNEL_IDS = True
class DependencyResolver ( object ):
"""Main object for dependency resolution."""
def __init__ ( self, err_queue ):
"""Initializes a DependencyResolver."""
self.logger = logging.getLogger ( self.__class__.__name__ )
self.logger_unresolvable = self.logger.getChild ( "UNRESOLVABLE" )
self.logger_resolved = self.logger.getChild ( "RESOLVED" )
self.listenermask = events.ALL
self.logmask = events.get_reverse_eventmask (
'RESOLVED', 'UNRESOLVABLE'
)
self._jobs = config.get ( "DEPRES.jobcount", 0 )
# used to lock the run methods,
self._runlock = threading.Lock()
self._need_reload = False
if self._jobs > 1:
# the dep res main thread
self._mainthread = None
self._thread_close = False
self.err_queue = err_queue
# the list of registered listener modules
self.listeners = list ()
# fifo queue for dep resolution
self._depqueue = queue.Queue()
# the queue of failed dep resolutions
# they can either be reinserted into the depqueue
# or marked as unresolvable
self._depqueue_failed = queue.Queue()
# the 'negative' result cache, stores unresolvable deps
# has to be (selectively?) cleared when
# new dep rule found / new rulepool etc.
if USING_DEPRES_CACHE:
self._dep_unresolvable = set ()
# map: channel identifier -> queue of done deps (resolved/unresolvable)
# this serves two purposes:
# (a) channels can do a blocking call on this queue
# (b) the keys of this dict are the list of known channels
self._depqueue_done = dict ()
# list of rule pools that have been created from reading files
self.static_rule_pools = list()
# list of rule pools that are recreatable (at runtime)
self.dynamic_rule_pools = list()
if SAFE_CHANNEL_IDS:
# this lock is used in register_channel
self._chanlock = threading.Lock()
# this stores all channel ids ever registered to this resolver
self.all_channel_ids = set()
# --- end of __init__ (...) ---
def get_threadcount ( self ):
return self._jobs
# --- end of get_threadcount (...) ---
def _sort ( self ):
"""Sorts the rule pools of this resolver."""
for pool in self.static_rule_pools: pool.sort()
poolsort = lambda pool : ( pool.priority, pool.rule_weight )
self.static_rule_pools.sort ( key=poolsort )
self.dynamic_rule_pools.sort ( key=poolsort )
# --- end of sort (...) ---
def _reset_unresolvable ( self ):
if USING_DEPRES_CACHE:
self._dep_unresolvable.clear()
# --- end of _reset_unresolvable (...) ---
def _new_rulepools_added ( self ):
"""Called after adding new rool pools."""
self._reset_unresolvable()
self._sort()
# --- end of _new_rulepools_added (...) ---
def get_reader ( self ):
return roverlay.depres.simpledeprule.reader.SimpleDependencyRuleReader (
pool_add=self.static_rule_pools.append,
when_done=self._new_rulepools_added
)
# --- end of get_reader (...) ---
def make_selfdep_pool ( self, rule_kw_function, reload_now=False ):
"""Creates an dynamic selfdep pool and adds it to this resolver.
arguments:
* rule_kw_function -- function that returns an rule creation keyword
generator (basically, it has to return a list
of dicts)
"""
pool = roverlay.depres.simpledeprule.dynpool.get ( rule_kw_function )
self.dynamic_rule_pools.append ( pool )
if reload_now:
pool.reload()
self._new_rulepools_added()
# --- end of make_selfdep_pool (...) ---
def add_rulepool ( self, rulepool, pool_type=None ):
"""Adds a (static) rule pool to this resolver.
Calls self.sort() afterwards.
arguments:
* rulepool -- rule pool type, 1 for dynamic, else static
* pool_type -- ignored.
"""
if pool_type == 1:
self.dynamic_rule_pools.append ( rulepool )
else:
self.static_rule_pools.append ( rulepool )
self._new_rulepools_added()
# --- end of add_rulepool (...) ---
def need_reload ( self ):
self._need_reload = True
# --- end of need_reload (...) ---
def reload_pools ( self, only_if_required=False ):
if not only_if_required or self._need_reload:
one = False
for pool in self.dynamic_rule_pools:
one = True
pool.reload()
if one:
self._new_rulepools_added()
self._need_reload = False
# --- end of reload_pools (...) ---
def _report_event ( self, event, dep_env=None, pkg_env=None, msg=None ):
"""Reports an event to the log and listeners.
arguments:
* event -- name of the event (RESOLVED etc., use capslock!)
* dep_env -- dependency env
* pkg_env -- package env, reserved for future usage
returns: None (implicit)
"""
# COULDFIX: needs some cleanup
# for example, it's already known whether event==(UN)RESOLVED
event_type = events.DEPRES_EVENTS [event]
if self.logmask & event_type:
# log this event
if event_type == events.DEPRES_EVENTS ['RESOLVED']:
self.logger_resolved.info (
"0x{:x}, {!r} as {!r}".format (
dep_env.deptype_mask,
dep_env.dep_str, dep_env.resolved_by.dep
)
)
elif event_type == events.DEPRES_EVENTS ['UNRESOLVABLE']:
self.logger_unresolvable.info (
"0x{:x}, {!r}".format ( dep_env.deptype_mask, dep_env.dep_str ) )
else:
# "generic" event, expects that kw msg is set
self.logger.debug ( "event {}: {}".format ( event, msg ) )
# --- if
if self.listenermask & event_type:
# notify listeners
for lis in self.listeners:
lis.notify ( event_type, dep_env=dep_env, pkg_env=pkg_env )
# --- end of _report_event (...) ---
def set_logmask ( self, mask ):
"""Sets the logmask for this DependencyResolver which can be used to
filter events that would normally go into the log file.
Useful if a listener module reports such events in an extra file.
arguments:
* mask -- new logmask that defines which events are logged
returns: None (implicit)
"""
self.logmask = events.ALL if mask < 0 or mask > events.ALL else mask
# --- end of set_logmask (...) ---
def set_listenermask ( self, mask ):
"""Set the mask for the listener modules. This is totally independent
from the per-listener mask setting and can be used to filter certain
events.
arguments:
* mask -- new listenermask that defines which events are passed
returns: None (implicit)
"""
self.listenermask = events.ALL if mask < 0 or mask > events.ALL else mask
# --- end of set_listenermask (...) ---
def add_listener ( self, listener ):
"""Adds a listener, which listens to events such as
"dependency is unresolvable".
Possible use cases include redirecting such events into a file
for further parsing.
arguments:
* listener --
returns: None (implicit)
"""
self.listeners.append ( listener )
# --- end of add_listener (...) ---
def register_channel ( self, channel ):
"""Registers a communication channel with this resolver.
This channel can then be used to _talk_, e.g. queue dependencies for
resolution and collect the results later.
arguments:
* channel -- channel to be registered
automatically sets channel's resolver to self if it is None
raises: Exception if channel is already registered with this resolver
returns: channel
"""
if SAFE_CHANNEL_IDS:
try:
self._chanlock.acquire()
if channel.ident in self.all_channel_ids:
raise Exception ( "channel id reused!" )
else:
self.all_channel_ids.add ( channel.ident )
# register channel and allocate a queue in depqueue_done
self._depqueue_done [channel.ident] = queue.Queue()
channel.set_resolver (
self, channel_queue=self._depqueue_done [channel.ident]
)
finally:
self._chanlock.release()
else:
if channel.ident in self._depqueue_done:
raise Exception ( "channel is already registered." )
# register channel and allocate a queue in depqueue_done
self._depqueue_done [channel.ident] = queue.Queue()
channel.set_resolver (
self, channel_queue=self._depqueue_done [channel.ident]
)
return channel
# --- end of register_channel (...) ---
def channel_closed ( self, channel_id ):
"""Tells the dependency resolver that a channel has been closed.
It will then unregister the channel; this operation does not fail if the
channel is not registered with this resolver.
arguments:
* channel_id -- identifier of the closed channel
returns: None (implicit)
"""
# not removing channel_id's DepEnvs from the queues
# 'cause this costs time
try:
del self._depqueue_done [channel_id]
except KeyError as expected:
# ok
pass
# --- end of channel_closed (...) ---
def _queue_previously_failed ( self ):
"""Inserts all previously failed dependency lookups into the queue
again.
returns: None (implicit)
"""
while not self._depqueue_failed.empty():
# it has to be guaranteed that no items are removed from
# _depqueue_failed while calling this method,
# else queue.Empty will be raised
self._depqueue.put ( self._depqueue_failed.get_nowait() )
# --- end of _queue_previously_failed (...) ---
def start ( self ):
if self._jobs < 2:
if not self._depqueue.empty():
self._run_resolver()
if not self.err_queue.really_empty():
self.err_queue.unblock_queues()
else:
# new resolver threads run async and
# can be started with an empty depqueue
if self._runlock.acquire ( False ):
# else resolver is running
self._mainthread = threading.Thread (
target=self._thread_run_resolver
)
self._mainthread.start()
# _thread_run_resolver has to release the lock when done
# --- end of start (...) ---
def _process_unresolvable_queue ( self ):
# iterate over _depqueue_failed and report unresolved
while not self._depqueue_failed.empty() and self.err_queue.empty:
try:
channel_id, dep_env = self._depqueue_failed.get_nowait()
except queue.Empty:
# race cond empty() <-> get_nowait()
return
dep_env.set_unresolvable()
self._report_event ( 'UNRESOLVABLE', dep_env )
try:
if channel_id in self._depqueue_done:
self._depqueue_done [channel_id].put_nowait ( dep_env )
except KeyError:
# channel has been closed before calling put, ignore this
pass
# --- end of _process_unresolvable_queue (...) ---
def _process_dep ( self, queue_item ):
channel_id, dep_env = queue_item
# drop dep if channel closed
if not channel_id in self._depqueue_done: return
self.logger.debug (
"Trying to resolve {!r}.".format ( dep_env.dep_str )
)
resolved = None
# resolved can be None, so use a tri-state int for checking
# 0 -> unresolved, but resolvable
# 1 -> unresolved and (currently, new rules may change this)
# not resolvable
# 2 -> resolved
is_resolved = 0
if USING_DEPRES_CACHE and dep_env.dep_str_low in self._dep_unresolvable:
# cannot resolve
is_resolved = 1
else:
for rulepool in (
p for p in self.dynamic_rule_pools if p.accepts ( dep_env )
):
result = rulepool.matches ( dep_env )
if result:
resolved = result
is_resolved = 2
break
# TRY_OTHER searching is disabled for dynamic rule pools,
# (a) no dyn pool uses it, (b) probably not useful
# search for a match in the rule pools that accept the dep type
for rulepool in (
p for p in self.static_rule_pools if p.accepts ( dep_env )
):
result = rulepool.matches ( dep_env )
if result:
resolved = result
is_resolved = 2
break
if is_resolved == 0 and dep_env.deptype_mask & deptype.try_other:
## TRY_OTHER bit is set
# search for a match in the rule pools
# that (normally) don't accept the dep type
for rulepool in (
p for p in self.static_rule_pools
if p.accepts_other ( dep_env )
):
result = rulepool.matches ( dep_env )
if result:
resolved = result
is_resolved = 2
break
# --
# -- done with resolving
if is_resolved != 2:
# could not resolve dep_env
self._depqueue_failed.put ( queue_item )
if USING_DEPRES_CACHE:
# does not work when adding new rules is possible
self._dep_unresolvable.add ( dep_env.dep_str_low )
else:
# successfully resolved
dep_env.set_resolved ( resolved, append=False )
self._report_event ( 'RESOLVED', dep_env )
try:
self._depqueue_done [channel_id].put ( dep_env )
except KeyError:
# channel gone while resolving
pass
"""
## only useful if new rules can be created
# new rule found, requeue all previously
# failed dependency searches
if have_new_rule:
self._queue_previously_failed
if USING_DEPRES_CACHE:
self._dep_unresolvable.clear() #?
"""
# --- end of _process_dep (...) ---
def _run_resolver ( self ):
# single-threaded variant of run
# still checking err_queue 'cause other modules
# could be run with threads
if self._depqueue.empty(): return
try:
self._runlock.acquire()
while not self._depqueue.empty() and self.err_queue.empty:
to_resolve = self._depqueue.get_nowait()
self._process_dep ( queue_item=to_resolve )
self._depqueue.task_done()
self._process_unresolvable_queue()
except ( Exception, KeyboardInterrupt ) as e:
# single-threaded exception catcher:
# * push exception to inform other threads (if any)
# * unblock queues (automatically when calling push)
# * reraise
self.err_queue.push ( id ( self ), e )
raise e
finally:
self._runlock.release()
# --- end of _run_resolver (...) ---
def _thread_run_resolver ( self ):
"""master thread"""
try:
self.logger.debug (
"Running in concurrent mode with {} worker threads.".format (
self._jobs
)
)
send_queues = tuple (
queue.Queue ( maxsize=1 ) for k in range ( self._jobs )
)
rec_queues = tuple (
queue.Queue ( maxsize=1 ) for k in range ( self._jobs )
)
threads = tuple (
threading.Thread (
target=self._thread_resolve,
# this thread's send queue is the worker thread's receive queue
# and vice versa
kwargs={ 'recq' : send_queues [n], 'sendq' : rec_queues [n] }
) for n in range ( self._jobs )
)
try:
for t in threads: t.start()
# *loop forever*
# wait for the resolver threads to process the dep queue,
# mark remaining deps as unresolvable and
# tell the threads to continue
while self.err_queue.really_empty() and not self._thread_close:
for q in rec_queues:
if q.get() != 0:
self._thread_close = True
break
else:
self._process_unresolvable_queue()
# tell the threads to continue
for q in send_queues: q.put_nowait ( 0 )
except ( Exception, KeyboardInterrupt ) as e:
self.err_queue.push ( context=id ( self ), error=e )
self._thread_close = True
# on-error code (self.err_queue not empty or close requested)
try:
for q in send_queues: q.put_nowait ( 2 )
except:
pass
for t in threads: t.join()
finally:
self._runlock.release()
# --- end of _thread_run_resolver (...) ---
def _thread_resolve ( self, sendq=0, recq=0 ):
"""worker thread"""
try:
while not self._thread_close and self.err_queue.empty:
try:
# process remaining deps
while not self._thread_close and self.err_queue.empty:
self._process_dep ( self._depqueue.get_nowait() )
except queue.Empty:
pass
# dep queue has been processed,
# let the master thread process all unresolvable deps
# only 0 means continue, anything else stops this thread
sendq.put_nowait ( 0 )
if recq.get() != 0: break
except ( Exception, KeyboardInterrupt ) as e:
self.err_queue.push ( id ( self ), e )
# this is on-error code (err_queue is not empty or close requested)
self._thread_close = True
try:
sendq.put_nowait ( 2 )
except queue.Full:
pass
# --- end of _thread_resolve (...) ---
def enqueue ( self, dep_env, channel_id ):
"""Adds a DepEnv to the queue of deps to resolve.
arguments:
* dep_env -- to add
* channel_id -- identifier of the channel associated with the dep_env
returns: None (implicit)
"""
self._depqueue.put ( ( channel_id, dep_env ) )
# --- end of enqueue (...) ---
def close ( self ):
if self._jobs > 1:
self._thread_close = True
if self._mainthread:
self._mainthread.join()
for lis in self.listeners: lis.close()
del self.listeners
if SAFE_CHANNEL_IDS:
self.logger.debug (
"{} channels were in use.".format ( len ( self.all_channel_ids ) )
)
# --- end of close (...) ---
|