1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
|
# cache.py - apt cache abstraction
#
# Copyright (c) 2005-2009 Canonical
#
# Author: Michael Vogt <michael.vogt@ubuntu.com>
#
# This program is free software; you can redistribute it and/or
# modify it under the terms of the GNU General Public License as
# published by the Free Software Foundation; either version 2 of the
# License, or (at your option) any later version.
#
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU General Public License for more details.
#
# You should have received a copy of the GNU General Public License
# along with this program; if not, write to the Free Software
# Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
# USA
import os
import apt_pkg
from apt import Package
import apt.progress
class FetchCancelledException(IOError):
"""Exception that is thrown when the user cancels a fetch operation."""
class FetchFailedException(IOError):
"""Exception that is thrown when fetching fails."""
class LockFailedException(IOError):
"""Exception that is thrown when locking fails."""
class Cache(object):
"""Dictionary-like package cache.
This class has all the packages that are available in it's
dictionary
"""
def __init__(self, progress=None, rootdir=None, memonly=False):
self._callbacks = {}
if memonly:
# force apt to build its caches in memory
apt_pkg.Config.Set("Dir::Cache::pkgcache", "")
if rootdir:
if os.path.exists(rootdir+"/etc/apt/apt.conf"):
apt_pkg.ReadConfigFile(apt_pkg.Config,
rootdir + "/etc/apt/apt.conf")
if os.path.isdir(rootdir+"/etc/apt/apt.conf.d"):
apt_pkg.ReadConfigDir(apt_pkg.Config,
rootdir + "/etc/apt/apt.conf.d")
apt_pkg.Config.Set("Dir", rootdir)
apt_pkg.Config.Set("Dir::State::status",
rootdir + "/var/lib/dpkg/status")
# create required dirs/files when run with special rootdir
# automatically
self._check_and_create_required_dirs(rootdir)
# Call InitSystem so the change to Dir::State::Status is actually
# recognized (LP: #320665)
apt_pkg.InitSystem()
self.open(progress)
def _check_and_create_required_dirs(self, rootdir):
"""
check if the required apt directories/files are there and if
not create them
"""
files = ["/var/lib/dpkg/status",
"/etc/apt/sources.list",
]
dirs = ["/var/lib/dpkg",
"/etc/apt/",
"/var/cache/apt/archives/partial",
"/var/lib/apt/lists/partial",
]
for d in dirs:
if not os.path.exists(rootdir+d):
print "creating: ",rootdir+d
os.makedirs(rootdir+d)
for f in files:
if not os.path.exists(rootdir+f):
open(rootdir+f,"w")
def _runCallbacks(self, name):
""" internal helper to run a callback """
if name in self._callbacks:
for callback in self._callbacks[name]:
callback()
def open(self, progress):
""" Open the package cache, after that it can be used like
a dictionary
"""
if progress is None:
progress = apt.progress.OpProgress()
self._runCallbacks("cache_pre_open")
self._cache = apt_pkg.GetCache(progress)
self._depcache = apt_pkg.GetDepCache(self._cache)
self._records = apt_pkg.GetPkgRecords(self._cache)
self._list = apt_pkg.GetPkgSourceList()
self._list.ReadMainList()
self._dict = {}
progress.Op = "Building data structures"
i=last=0
size=len(self._cache.Packages)
for pkg in self._cache.Packages:
if progress is not None and last+100 < i:
progress.update(i/float(size)*100)
last=i
# drop stuff with no versions (cruft)
if len(pkg.VersionList) > 0:
self._dict[pkg.Name] = Package(self, pkg)
i += 1
progress.done()
self._runCallbacks("cache_post_open")
def __getitem__(self, key):
""" look like a dictionary (get key) """
return self._dict[key]
def __iter__(self):
for pkgname in self._dict.keys():
yield self._dict[pkgname]
raise StopIteration
def has_key(self, key):
return (key in self._dict)
def __contains__(self, key):
return (key in self._dict)
def __len__(self):
return len(self._dict)
def keys(self):
return self._dict.keys()
def getChanges(self):
""" Get the marked changes """
changes = []
for name in self._dict.keys():
p = self._dict[name]
if p.markedUpgrade or p.markedInstall or p.markedDelete or \
p.markedDowngrade or p.markedReinstall:
changes.append(p)
return changes
def upgrade(self, distUpgrade=False):
""" Upgrade the all package, DistUpgrade will also install
new dependencies
"""
self.cachePreChange()
self._depcache.Upgrade(distUpgrade)
self.cachePostChange()
@property
def requiredDownload(self):
"""Get the size of the packages that are required to download."""
pm = apt_pkg.GetPackageManager(self._depcache)
fetcher = apt_pkg.GetAcquire()
pm.GetArchives(fetcher, self._list, self._records)
return fetcher.FetchNeeded
@property
def additionalRequiredSpace(self):
"""Get the size of the additional required space on the fs."""
return self._depcache.UsrSize
@property
def reqReinstallPkgs(self):
"""Return the packages not downloadable packages in reqreinst state."""
reqreinst = set()
for pkg in self:
if (not pkg.candidate.downloadable and
(pkg._pkg.InstState == apt_pkg.InstStateReInstReq or
pkg._pkg.InstState == apt_pkg.InstStateHoldReInstReq)):
reqreinst.add(pkg.name)
return reqreinst
def _runFetcher(self, fetcher):
# do the actual fetching
res = fetcher.Run()
# now check the result (this is the code from apt-get.cc)
failed = False
transient = False
errMsg = ""
for item in fetcher.Items:
if item.Status == item.StatDone:
continue
if item.StatIdle:
transient = True
continue
errMsg += "Failed to fetch %s %s\n" % (item.DescURI,
item.ErrorText)
failed = True
# we raise a exception if the download failed or it was cancelt
if res == fetcher.ResultCancelled:
raise FetchCancelledException(errMsg)
elif failed:
raise FetchFailedException(errMsg)
return res
def _fetchArchives(self, fetcher, pm):
""" fetch the needed archives """
# get lock
lockfile = apt_pkg.Config.FindDir("Dir::Cache::Archives") + "lock"
lock = apt_pkg.GetLock(lockfile)
if lock < 0:
raise LockFailedException("Failed to lock %s" % lockfile)
try:
# this may as well throw a SystemError exception
if not pm.GetArchives(fetcher, self._list, self._records):
return False
# now run the fetcher, throw exception if something fails to be
# fetched
return self._runFetcher(fetcher)
finally:
os.close(lock)
def isVirtualPackage(self, pkgname):
"""Return whether the package is a virtual package."""
pkg = self._cache[pkgname]
return bool(pkg.ProvidesList and not pkg.VersionList)
def getProvidingPackages(self, virtual):
"""
Return a list of packages which provide the virtual package of the
specified name
"""
providers = []
try:
vp = self._cache[virtual]
if len(vp.VersionList) != 0:
return providers
except KeyError:
return providers
for pkg in self:
v = self._depcache.GetCandidateVer(pkg._pkg)
if v is None:
continue
for p in v.ProvidesList:
if virtual == p[0]:
# we found a pkg that provides this virtual pkg
providers.append(pkg)
return providers
def update(self, fetchProgress=None):
" run the equivalent of apt-get update "
lockfile = apt_pkg.Config.FindDir("Dir::State::Lists") + "lock"
lock = apt_pkg.GetLock(lockfile)
if lock < 0:
raise LockFailedException("Failed to lock %s" % lockfile)
try:
if fetchProgress is None:
fetchProgress = apt.progress.FetchProgress()
return self._cache.Update(fetchProgress, self._list)
finally:
os.close(lock)
def installArchives(self, pm, installProgress):
installProgress.startUpdate()
res = installProgress.run(pm)
installProgress.finishUpdate()
return res
def commit(self, fetchProgress=None, installProgress=None):
""" Apply the marked changes to the cache """
# FIXME:
# use the new acquire/pkgmanager interface here,
# raise exceptions when a download or install fails
# and send proper error strings to the application.
# Current a failed download will just display "error"
# which is less than optimal!
if fetchProgress is None:
fetchProgress = apt.progress.FetchProgress()
if installProgress is None:
installProgress = apt.progress.InstallProgress()
pm = apt_pkg.GetPackageManager(self._depcache)
fetcher = apt_pkg.GetAcquire(fetchProgress)
while True:
# fetch archives first
res = self._fetchArchives(fetcher, pm)
# then install
res = self.installArchives(pm, installProgress)
if res == pm.ResultCompleted:
break
if res == pm.ResultFailed:
raise SystemError("installArchives() failed")
# reload the fetcher for media swaping
fetcher.Shutdown()
return (res == pm.ResultCompleted)
def clear(self):
""" Unmark all changes """
self._depcache.Init()
# cache changes
def cachePostChange(self):
" called internally if the cache has changed, emit a signal then "
self._runCallbacks("cache_post_change")
def cachePreChange(self):
""" called internally if the cache is about to change, emit
a signal then """
self._runCallbacks("cache_pre_change")
def connect(self, name, callback):
""" connect to a signal, currently only used for
cache_{post,pre}_{changed,open} """
if not name in self._callbacks:
self._callbacks[name] = []
self._callbacks[name].append(callback)
@property
def broken_count(self):
"""Return the number of packages with broken dependencies."""
return self._depcache.broken_count
@property
def delete_count(self):
"""Return the number of packages marked for deletion."""
return self._depcache.del_count
@property
def install_count(self):
"""Return the number of packages marked for installation."""
return self._depcache.inst_count
@property
def keep_count(self):
"""Return the number of packages marked as keep."""
return self._depcache.keep_count
# ----------------------------- experimental interface
class Filter(object):
""" Filter base class """
def apply(self, pkg):
""" Filter function, return True if the package matchs a
filter criteria and False otherwise
"""
return True
class MarkedChangesFilter(Filter):
""" Filter that returns all marked changes """
def apply(self, pkg):
if pkg.markedInstall or pkg.markedDelete or pkg.markedUpgrade:
return True
else:
return False
class FilteredCache(object):
""" A package cache that is filtered.
Can work on a existing cache or create a new one
"""
def __init__(self, cache=None, progress=None):
if cache is None:
self.cache = Cache(progress)
else:
self.cache = cache
self.cache.connect("cache_post_change", self.filterCachePostChange)
self.cache.connect("cache_post_open", self.filterCachePostChange)
self._filtered = {}
self._filters = []
def __len__(self):
return len(self._filtered)
def __getitem__(self, key):
return self.cache._dict[key]
def __iter__(self):
for pkgname in self._filtered:
yield self.cache[pkgname]
def keys(self):
return self._filtered.keys()
def has_key(self, key):
return (key in self._filtered)
def __contains__(self, key):
return (key in self._filtered)
def _reapplyFilter(self):
" internal helper to refilter "
self._filtered = {}
for pkg in self.cache._dict.keys():
for f in self._filters:
if f.apply(self.cache._dict[pkg]):
self._filtered[pkg] = 1
break
def setFilter(self, filter):
"""Set the current active filter."""
self._filters = []
self._filters.append(filter)
#self._reapplyFilter()
# force a cache-change event that will result in a refiltering
self.cache.cachePostChange()
def filterCachePostChange(self):
"""Called internally if the cache changes, emit a signal then."""
#print "filterCachePostChange()"
self._reapplyFilter()
# def connect(self, name, callback):
# self.cache.connect(name, callback)
def __getattr__(self, key):
"""we try to look exactly like a real cache."""
#print "getattr: %s " % key
return getattr(self.cache, key)
def cache_pre_changed():
print "cache pre changed"
def cache_post_changed():
print "cache post changed"
# internal test code
if __name__ == "__main__":
print "Cache self test"
apt_pkg.init()
c = Cache(apt.progress.OpTextProgress())
c.connect("cache_pre_change", cache_pre_changed)
c.connect("cache_post_change", cache_post_changed)
print ("aptitude" in c)
p = c["aptitude"]
print p.name
print len(c)
for pkg in c.keys():
x= c[pkg].name
c.upgrade()
changes = c.getChanges()
print len(changes)
for p in changes:
#print p.name
x = p.name
# see if fetching works
for d in ["/tmp/pytest", "/tmp/pytest/partial"]:
if not os.path.exists(d):
os.mkdir(d)
apt_pkg.Config.Set("Dir::Cache::Archives", "/tmp/pytest")
pm = apt_pkg.GetPackageManager(c._depcache)
fetcher = apt_pkg.GetAcquire(apt.progress.TextFetchProgress())
c._fetchArchives(fetcher, pm)
#sys.exit(1)
print "Testing filtered cache (argument is old cache)"
f = FilteredCache(c)
f.cache.connect("cache_pre_change", cache_pre_changed)
f.cache.connect("cache_post_change", cache_post_changed)
f.cache.upgrade()
f.setFilter(MarkedChangesFilter())
print len(f)
for pkg in f.keys():
#print c[pkg].name
x = f[pkg].name
print len(f)
print "Testing filtered cache (no argument)"
f = FilteredCache(progress=apt.progress.OpTextProgress())
f.cache.connect("cache_pre_change", cache_pre_changed)
f.cache.connect("cache_post_change", cache_post_changed)
f.cache.upgrade()
f.setFilter(MarkedChangesFilter())
print len(f)
for pkg in f.keys():
#print c[pkg].name
x = f[pkg].name
print len(f)
|