Update to 1.3.3 release of bitbake
git-svn-id: https://svn.o-hand.com/repos/poky/trunk@269 311d38ba-8fff-0310-9ca6-ca027cbcb966
This commit is contained in:
parent
b236090192
commit
ce241d7e6d
|
@ -1,4 +1,4 @@
|
||||||
Changes in BitBake 1.3.?:
|
Changes in BitBake 1.3.3:
|
||||||
- Create a new Fetcher module to ease the
|
- Create a new Fetcher module to ease the
|
||||||
development of new Fetchers.
|
development of new Fetchers.
|
||||||
Issue #438 fixed by rpurdie@openedhand.com
|
Issue #438 fixed by rpurdie@openedhand.com
|
||||||
|
@ -7,13 +7,19 @@ Changes in BitBake 1.3.?:
|
||||||
Issue #555 fixed by chris@openedhand.com
|
Issue #555 fixed by chris@openedhand.com
|
||||||
- Expand PREFERRED_PROVIDER properly
|
- Expand PREFERRED_PROVIDER properly
|
||||||
Issue #436 fixed by rprudie@openedhand.com
|
Issue #436 fixed by rprudie@openedhand.com
|
||||||
- Typo fix for Issue #531 Philipp Zabel for the
|
- Typo fix for Issue #531 by Philipp Zabel for the
|
||||||
BitBake Shell
|
BitBake Shell
|
||||||
- Introduce a new special variable SRCDATE as
|
- Introduce a new special variable SRCDATE as
|
||||||
a generic naming to take over CVSDATE.
|
a generic naming to replace CVSDATE.
|
||||||
- Introduce a new keyword 'required' In contrast
|
- Introduce a new keyword 'required'. In contrast
|
||||||
to include parsing will fail if a to be included
|
to 'include' parsing will fail if a to be included
|
||||||
can not be found.
|
file can not be found.
|
||||||
|
- Remove hardcoding of the STAMP directory. Patch
|
||||||
|
courtsey pHilipp Zabel
|
||||||
|
- Track the RDEPENDS of each package (rpurdie@openedhand.com)
|
||||||
|
- Introduce BUILD_ALL_DEPS to build all RDEPENDS. E.g
|
||||||
|
this is used by the OpenEmbedded Meta Packages.
|
||||||
|
(rpurdie@openedhand.com).
|
||||||
|
|
||||||
Changes in BitBake 1.3.2:
|
Changes in BitBake 1.3.2:
|
||||||
- reintegration of make.py into BitBake
|
- reintegration of make.py into BitBake
|
||||||
|
|
|
@ -31,7 +31,7 @@ import itertools, optparse
|
||||||
parsespin = itertools.cycle( r'|/-\\' )
|
parsespin = itertools.cycle( r'|/-\\' )
|
||||||
bbdebug = 0
|
bbdebug = 0
|
||||||
|
|
||||||
__version__ = "1.3.2"
|
__version__ = "1.3.3"
|
||||||
|
|
||||||
#============================================================================#
|
#============================================================================#
|
||||||
# BBParsingStatus
|
# BBParsingStatus
|
||||||
|
@ -461,8 +461,7 @@ class BBCooker:
|
||||||
pn = bb.data.getVar('PN', the_data, 1)
|
pn = bb.data.getVar('PN', the_data, 1)
|
||||||
pv = bb.data.getVar('PV', the_data, 1)
|
pv = bb.data.getVar('PV', the_data, 1)
|
||||||
pr = bb.data.getVar('PR', the_data, 1)
|
pr = bb.data.getVar('PR', the_data, 1)
|
||||||
tmpdir = bb.data.getVar('TMPDIR', the_data, 1)
|
stamp = '%s.do_populate_staging' % bb.data.getVar('STAMP', the_data, 1)
|
||||||
stamp = '%s/stamps/%s-%s-%s.do_populate_staging' % (tmpdir, pn, pv, pr)
|
|
||||||
if os.path.exists(stamp):
|
if os.path.exists(stamp):
|
||||||
(newvers, fn) = preferred_versions[pn]
|
(newvers, fn) = preferred_versions[pn]
|
||||||
if not fn in eligible:
|
if not fn in eligible:
|
||||||
|
|
|
@ -23,7 +23,7 @@ this program; if not, write to the Free Software Foundation, Inc., 59 Temple
|
||||||
Place, Suite 330, Boston, MA 02111-1307 USA.
|
Place, Suite 330, Boston, MA 02111-1307 USA.
|
||||||
"""
|
"""
|
||||||
|
|
||||||
__version__ = "1.3.2.1"
|
__version__ = "1.3.3.0"
|
||||||
|
|
||||||
__all__ = [
|
__all__ = [
|
||||||
|
|
||||||
|
|
|
@ -1,656 +0,0 @@
|
||||||
#!/usr/bin/env python
|
|
||||||
# ex:ts=4:sw=4:sts=4:et
|
|
||||||
# -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*-
|
|
||||||
"""
|
|
||||||
BitBake 'Fetch' implementations
|
|
||||||
|
|
||||||
Classes for obtaining upstream sources for the
|
|
||||||
BitBake build tools.
|
|
||||||
|
|
||||||
Copyright (C) 2003, 2004 Chris Larson
|
|
||||||
|
|
||||||
This program is free software; you can redistribute it and/or modify it under
|
|
||||||
the terms of the GNU General Public License as published by the Free Software
|
|
||||||
Foundation; either version 2 of the License, or (at your option) any later
|
|
||||||
version.
|
|
||||||
|
|
||||||
This program is distributed in the hope that it will be useful, but WITHOUT
|
|
||||||
ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS
|
|
||||||
FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details.
|
|
||||||
|
|
||||||
You should have received a copy of the GNU General Public License along with
|
|
||||||
this program; if not, write to the Free Software Foundation, Inc., 59 Temple
|
|
||||||
Place, Suite 330, Boston, MA 02111-1307 USA.
|
|
||||||
|
|
||||||
Based on functions from the base bb module, Copyright 2003 Holger Schurig
|
|
||||||
"""
|
|
||||||
|
|
||||||
import os, re
|
|
||||||
import bb
|
|
||||||
from bb import data
|
|
||||||
|
|
||||||
class FetchError(Exception):
|
|
||||||
"""Exception raised when a download fails"""
|
|
||||||
|
|
||||||
class NoMethodError(Exception):
|
|
||||||
"""Exception raised when there is no method to obtain a supplied url or set of urls"""
|
|
||||||
|
|
||||||
class MissingParameterError(Exception):
|
|
||||||
"""Exception raised when a fetch method is missing a critical parameter in the url"""
|
|
||||||
|
|
||||||
class MD5SumError(Exception):
|
|
||||||
"""Exception raised when a MD5SUM of a file does not match the expected one"""
|
|
||||||
|
|
||||||
def uri_replace(uri, uri_find, uri_replace, d):
|
|
||||||
# bb.note("uri_replace: operating on %s" % uri)
|
|
||||||
if not uri or not uri_find or not uri_replace:
|
|
||||||
bb.debug(1, "uri_replace: passed an undefined value, not replacing")
|
|
||||||
uri_decoded = list(bb.decodeurl(uri))
|
|
||||||
uri_find_decoded = list(bb.decodeurl(uri_find))
|
|
||||||
uri_replace_decoded = list(bb.decodeurl(uri_replace))
|
|
||||||
result_decoded = ['','','','','',{}]
|
|
||||||
for i in uri_find_decoded:
|
|
||||||
loc = uri_find_decoded.index(i)
|
|
||||||
result_decoded[loc] = uri_decoded[loc]
|
|
||||||
import types
|
|
||||||
if type(i) == types.StringType:
|
|
||||||
import re
|
|
||||||
if (re.match(i, uri_decoded[loc])):
|
|
||||||
result_decoded[loc] = re.sub(i, uri_replace_decoded[loc], uri_decoded[loc])
|
|
||||||
if uri_find_decoded.index(i) == 2:
|
|
||||||
if d:
|
|
||||||
localfn = bb.fetch.localpath(uri, d)
|
|
||||||
if localfn:
|
|
||||||
result_decoded[loc] = os.path.dirname(result_decoded[loc]) + "/" + os.path.basename(bb.fetch.localpath(uri, d))
|
|
||||||
# bb.note("uri_replace: matching %s against %s and replacing with %s" % (i, uri_decoded[loc], uri_replace_decoded[loc]))
|
|
||||||
else:
|
|
||||||
# bb.note("uri_replace: no match")
|
|
||||||
return uri
|
|
||||||
# else:
|
|
||||||
# for j in i.keys():
|
|
||||||
# FIXME: apply replacements against options
|
|
||||||
return bb.encodeurl(result_decoded)
|
|
||||||
|
|
||||||
methods = []
|
|
||||||
|
|
||||||
def init(urls = [], d = None):
|
|
||||||
if d == None:
|
|
||||||
bb.debug(2,"BUG init called with None as data object!!!")
|
|
||||||
return
|
|
||||||
|
|
||||||
for m in methods:
|
|
||||||
m.urls = []
|
|
||||||
|
|
||||||
for u in urls:
|
|
||||||
for m in methods:
|
|
||||||
m.data = d
|
|
||||||
if m.supports(u, d):
|
|
||||||
m.urls.append(u)
|
|
||||||
|
|
||||||
def go(d):
|
|
||||||
"""Fetch all urls"""
|
|
||||||
for m in methods:
|
|
||||||
if m.urls:
|
|
||||||
m.go(d)
|
|
||||||
|
|
||||||
def localpaths(d):
|
|
||||||
"""Return a list of the local filenames, assuming successful fetch"""
|
|
||||||
local = []
|
|
||||||
for m in methods:
|
|
||||||
for u in m.urls:
|
|
||||||
local.append(m.localpath(u, d))
|
|
||||||
return local
|
|
||||||
|
|
||||||
def localpath(url, d):
|
|
||||||
for m in methods:
|
|
||||||
if m.supports(url, d):
|
|
||||||
return m.localpath(url, d)
|
|
||||||
return url
|
|
||||||
|
|
||||||
class Fetch(object):
|
|
||||||
"""Base class for 'fetch'ing data"""
|
|
||||||
|
|
||||||
def __init__(self, urls = []):
|
|
||||||
self.urls = []
|
|
||||||
for url in urls:
|
|
||||||
if self.supports(bb.decodeurl(url), d) is 1:
|
|
||||||
self.urls.append(url)
|
|
||||||
|
|
||||||
def supports(url, d):
|
|
||||||
"""Check to see if this fetch class supports a given url.
|
|
||||||
Expects supplied url in list form, as outputted by bb.decodeurl().
|
|
||||||
"""
|
|
||||||
return 0
|
|
||||||
supports = staticmethod(supports)
|
|
||||||
|
|
||||||
def localpath(url, d):
|
|
||||||
"""Return the local filename of a given url assuming a successful fetch.
|
|
||||||
"""
|
|
||||||
return url
|
|
||||||
localpath = staticmethod(localpath)
|
|
||||||
|
|
||||||
def setUrls(self, urls):
|
|
||||||
self.__urls = urls
|
|
||||||
|
|
||||||
def getUrls(self):
|
|
||||||
return self.__urls
|
|
||||||
|
|
||||||
urls = property(getUrls, setUrls, None, "Urls property")
|
|
||||||
|
|
||||||
def setData(self, data):
|
|
||||||
self.__data = data
|
|
||||||
|
|
||||||
def getData(self):
|
|
||||||
return self.__data
|
|
||||||
|
|
||||||
data = property(getData, setData, None, "Data property")
|
|
||||||
|
|
||||||
def go(self, urls = []):
|
|
||||||
"""Fetch urls"""
|
|
||||||
raise NoMethodError("Missing implementation for url")
|
|
||||||
|
|
||||||
class Wget(Fetch):
|
|
||||||
"""Class to fetch urls via 'wget'"""
|
|
||||||
def supports(url, d):
|
|
||||||
"""Check to see if a given url can be fetched using wget.
|
|
||||||
Expects supplied url in list form, as outputted by bb.decodeurl().
|
|
||||||
"""
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(url, d))
|
|
||||||
return type in ['http','https','ftp']
|
|
||||||
supports = staticmethod(supports)
|
|
||||||
|
|
||||||
def localpath(url, d):
|
|
||||||
# strip off parameters
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(url, d))
|
|
||||||
if "localpath" in parm:
|
|
||||||
# if user overrides local path, use it.
|
|
||||||
return parm["localpath"]
|
|
||||||
url = bb.encodeurl([type, host, path, user, pswd, {}])
|
|
||||||
|
|
||||||
return os.path.join(data.getVar("DL_DIR", d), os.path.basename(url))
|
|
||||||
localpath = staticmethod(localpath)
|
|
||||||
|
|
||||||
def go(self, d, urls = []):
|
|
||||||
"""Fetch urls"""
|
|
||||||
|
|
||||||
def md5_sum(basename, data):
|
|
||||||
"""
|
|
||||||
Fast and incomplete OVERRIDE implementation for MD5SUM handling
|
|
||||||
MD5SUM_basename = "SUM" and fallback to MD5SUM_basename
|
|
||||||
"""
|
|
||||||
var = "MD5SUM_%s" % basename
|
|
||||||
return getVar(var, data) or get("MD5SUM",data)
|
|
||||||
|
|
||||||
|
|
||||||
def fetch_uri(uri, basename, dl, md5, parm, d):
|
|
||||||
if os.path.exists(dl):
|
|
||||||
# file exists, but we didnt complete it.. trying again..
|
|
||||||
fetchcmd = data.getVar("RESUMECOMMAND", d, 1)
|
|
||||||
else:
|
|
||||||
fetchcmd = data.getVar("FETCHCOMMAND", d, 1)
|
|
||||||
|
|
||||||
bb.note("fetch " + uri)
|
|
||||||
fetchcmd = fetchcmd.replace("${URI}", uri)
|
|
||||||
fetchcmd = fetchcmd.replace("${FILE}", basename)
|
|
||||||
bb.debug(2, "executing " + fetchcmd)
|
|
||||||
ret = os.system(fetchcmd)
|
|
||||||
if ret != 0:
|
|
||||||
return False
|
|
||||||
|
|
||||||
# check if sourceforge did send us to the mirror page
|
|
||||||
dl_dir = data.getVar("DL_DIR", d, True)
|
|
||||||
if not os.path.exists(dl):
|
|
||||||
os.system("rm %s*" % dl) # FIXME shell quote it
|
|
||||||
bb.debug(2,"sourceforge.net send us to the mirror on %s" % basename)
|
|
||||||
return False
|
|
||||||
|
|
||||||
# supposedly complete.. write out md5sum
|
|
||||||
if bb.which(data.getVar('PATH', d), 'md5sum'):
|
|
||||||
try:
|
|
||||||
md5pipe = os.popen('md5sum ' + dl)
|
|
||||||
md5data = (md5pipe.readline().split() or [ "" ])[0]
|
|
||||||
md5pipe.close()
|
|
||||||
except OSError:
|
|
||||||
md5data = ""
|
|
||||||
md5out = file(md5, 'w')
|
|
||||||
md5out.write(md5data)
|
|
||||||
md5out.close()
|
|
||||||
else:
|
|
||||||
md5out = file(md5, 'w')
|
|
||||||
md5out.write("")
|
|
||||||
md5out.close()
|
|
||||||
return True
|
|
||||||
|
|
||||||
if not urls:
|
|
||||||
urls = self.urls
|
|
||||||
|
|
||||||
localdata = data.createCopy(d)
|
|
||||||
data.setVar('OVERRIDES', "wget:" + data.getVar('OVERRIDES', localdata), localdata)
|
|
||||||
data.update_data(localdata)
|
|
||||||
|
|
||||||
for uri in urls:
|
|
||||||
completed = 0
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(uri, localdata))
|
|
||||||
basename = os.path.basename(path)
|
|
||||||
dl = self.localpath(uri, d)
|
|
||||||
dl = data.expand(dl, localdata)
|
|
||||||
md5 = dl + '.md5'
|
|
||||||
|
|
||||||
if os.path.exists(md5):
|
|
||||||
# complete, nothing to see here..
|
|
||||||
continue
|
|
||||||
|
|
||||||
premirrors = [ i.split() for i in (data.getVar('PREMIRRORS', localdata, 1) or "").split('\n') if i ]
|
|
||||||
for (find, replace) in premirrors:
|
|
||||||
newuri = uri_replace(uri, find, replace, d)
|
|
||||||
if newuri != uri:
|
|
||||||
if fetch_uri(newuri, basename, dl, md5, parm, localdata):
|
|
||||||
completed = 1
|
|
||||||
break
|
|
||||||
|
|
||||||
if completed:
|
|
||||||
continue
|
|
||||||
|
|
||||||
if fetch_uri(uri, basename, dl, md5, parm, localdata):
|
|
||||||
continue
|
|
||||||
|
|
||||||
# try mirrors
|
|
||||||
mirrors = [ i.split() for i in (data.getVar('MIRRORS', localdata, 1) or "").split('\n') if i ]
|
|
||||||
for (find, replace) in mirrors:
|
|
||||||
newuri = uri_replace(uri, find, replace, d)
|
|
||||||
if newuri != uri:
|
|
||||||
if fetch_uri(newuri, basename, dl, md5, parm, localdata):
|
|
||||||
completed = 1
|
|
||||||
break
|
|
||||||
|
|
||||||
if not completed:
|
|
||||||
raise FetchError(uri)
|
|
||||||
|
|
||||||
del localdata
|
|
||||||
|
|
||||||
|
|
||||||
methods.append(Wget())
|
|
||||||
|
|
||||||
class Cvs(Fetch):
|
|
||||||
"""Class to fetch a module or modules from cvs repositories"""
|
|
||||||
def supports(url, d):
|
|
||||||
"""Check to see if a given url can be fetched with cvs.
|
|
||||||
Expects supplied url in list form, as outputted by bb.decodeurl().
|
|
||||||
"""
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(url, d))
|
|
||||||
return type in ['cvs', 'pserver']
|
|
||||||
supports = staticmethod(supports)
|
|
||||||
|
|
||||||
def localpath(url, d):
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(url, d))
|
|
||||||
if "localpath" in parm:
|
|
||||||
# if user overrides local path, use it.
|
|
||||||
return parm["localpath"]
|
|
||||||
|
|
||||||
if not "module" in parm:
|
|
||||||
raise MissingParameterError("cvs method needs a 'module' parameter")
|
|
||||||
else:
|
|
||||||
module = parm["module"]
|
|
||||||
if 'tag' in parm:
|
|
||||||
tag = parm['tag']
|
|
||||||
else:
|
|
||||||
tag = ""
|
|
||||||
if 'date' in parm:
|
|
||||||
date = parm['date']
|
|
||||||
else:
|
|
||||||
if not tag:
|
|
||||||
date = data.getVar("CVSDATE", d, 1) or data.getVar("DATE", d, 1)
|
|
||||||
else:
|
|
||||||
date = ""
|
|
||||||
|
|
||||||
return os.path.join(data.getVar("DL_DIR", d, 1),data.expand('%s_%s_%s_%s.tar.gz' % ( module.replace('/', '.'), host, tag, date), d))
|
|
||||||
localpath = staticmethod(localpath)
|
|
||||||
|
|
||||||
def go(self, d, urls = []):
|
|
||||||
"""Fetch urls"""
|
|
||||||
if not urls:
|
|
||||||
urls = self.urls
|
|
||||||
|
|
||||||
localdata = data.createCopy(d)
|
|
||||||
data.setVar('OVERRIDES', "cvs:%s" % data.getVar('OVERRIDES', localdata), localdata)
|
|
||||||
data.update_data(localdata)
|
|
||||||
|
|
||||||
for loc in urls:
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(loc, localdata))
|
|
||||||
if not "module" in parm:
|
|
||||||
raise MissingParameterError("cvs method needs a 'module' parameter")
|
|
||||||
else:
|
|
||||||
module = parm["module"]
|
|
||||||
|
|
||||||
dlfile = self.localpath(loc, localdata)
|
|
||||||
dldir = data.getVar('DL_DIR', localdata, 1)
|
|
||||||
# if local path contains the cvs
|
|
||||||
# module, consider the dir above it to be the
|
|
||||||
# download directory
|
|
||||||
# pos = dlfile.find(module)
|
|
||||||
# if pos:
|
|
||||||
# dldir = dlfile[:pos]
|
|
||||||
# else:
|
|
||||||
# dldir = os.path.dirname(dlfile)
|
|
||||||
|
|
||||||
# setup cvs options
|
|
||||||
options = []
|
|
||||||
if 'tag' in parm:
|
|
||||||
tag = parm['tag']
|
|
||||||
else:
|
|
||||||
tag = ""
|
|
||||||
|
|
||||||
if 'date' in parm:
|
|
||||||
date = parm['date']
|
|
||||||
else:
|
|
||||||
if not tag:
|
|
||||||
date = data.getVar("CVSDATE", d, 1) or data.getVar("DATE", d, 1)
|
|
||||||
else:
|
|
||||||
date = ""
|
|
||||||
|
|
||||||
if "method" in parm:
|
|
||||||
method = parm["method"]
|
|
||||||
else:
|
|
||||||
method = "pserver"
|
|
||||||
|
|
||||||
if "localdir" in parm:
|
|
||||||
localdir = parm["localdir"]
|
|
||||||
else:
|
|
||||||
localdir = module
|
|
||||||
|
|
||||||
cvs_rsh = None
|
|
||||||
if method == "ext":
|
|
||||||
if "rsh" in parm:
|
|
||||||
cvs_rsh = parm["rsh"]
|
|
||||||
|
|
||||||
tarfn = data.expand('%s_%s_%s_%s.tar.gz' % (module.replace('/', '.'), host, tag, date), localdata)
|
|
||||||
data.setVar('TARFILES', dlfile, localdata)
|
|
||||||
data.setVar('TARFN', tarfn, localdata)
|
|
||||||
|
|
||||||
dl = os.path.join(dldir, tarfn)
|
|
||||||
if os.access(dl, os.R_OK):
|
|
||||||
bb.debug(1, "%s already exists, skipping cvs checkout." % tarfn)
|
|
||||||
continue
|
|
||||||
|
|
||||||
pn = data.getVar('PN', d, 1)
|
|
||||||
cvs_tarball_stash = None
|
|
||||||
if pn:
|
|
||||||
cvs_tarball_stash = data.getVar('CVS_TARBALL_STASH_%s' % pn, d, 1)
|
|
||||||
if cvs_tarball_stash == None:
|
|
||||||
cvs_tarball_stash = data.getVar('CVS_TARBALL_STASH', d, 1)
|
|
||||||
if cvs_tarball_stash:
|
|
||||||
fetchcmd = data.getVar("FETCHCOMMAND_wget", d, 1)
|
|
||||||
uri = cvs_tarball_stash + tarfn
|
|
||||||
bb.note("fetch " + uri)
|
|
||||||
fetchcmd = fetchcmd.replace("${URI}", uri)
|
|
||||||
ret = os.system(fetchcmd)
|
|
||||||
if ret == 0:
|
|
||||||
bb.note("Fetched %s from tarball stash, skipping checkout" % tarfn)
|
|
||||||
continue
|
|
||||||
|
|
||||||
if date:
|
|
||||||
options.append("-D %s" % date)
|
|
||||||
if tag:
|
|
||||||
options.append("-r %s" % tag)
|
|
||||||
|
|
||||||
olddir = os.path.abspath(os.getcwd())
|
|
||||||
os.chdir(data.expand(dldir, localdata))
|
|
||||||
|
|
||||||
# setup cvsroot
|
|
||||||
if method == "dir":
|
|
||||||
cvsroot = path
|
|
||||||
else:
|
|
||||||
cvsroot = ":" + method + ":" + user
|
|
||||||
if pswd:
|
|
||||||
cvsroot += ":" + pswd
|
|
||||||
cvsroot += "@" + host + ":" + path
|
|
||||||
|
|
||||||
data.setVar('CVSROOT', cvsroot, localdata)
|
|
||||||
data.setVar('CVSCOOPTS', " ".join(options), localdata)
|
|
||||||
data.setVar('CVSMODULE', module, localdata)
|
|
||||||
cvscmd = data.getVar('FETCHCOMMAND', localdata, 1)
|
|
||||||
cvsupdatecmd = data.getVar('UPDATECOMMAND', localdata, 1)
|
|
||||||
|
|
||||||
if cvs_rsh:
|
|
||||||
cvscmd = "CVS_RSH=\"%s\" %s" % (cvs_rsh, cvscmd)
|
|
||||||
cvsupdatecmd = "CVS_RSH=\"%s\" %s" % (cvs_rsh, cvsupdatecmd)
|
|
||||||
|
|
||||||
# create module directory
|
|
||||||
bb.debug(2, "Fetch: checking for module directory")
|
|
||||||
pkg=data.expand('${PN}', d)
|
|
||||||
pkgdir=os.path.join(data.expand('${CVSDIR}', localdata), pkg)
|
|
||||||
moddir=os.path.join(pkgdir,localdir)
|
|
||||||
if os.access(os.path.join(moddir,'CVS'), os.R_OK):
|
|
||||||
bb.note("Update " + loc)
|
|
||||||
# update sources there
|
|
||||||
os.chdir(moddir)
|
|
||||||
myret = os.system(cvsupdatecmd)
|
|
||||||
else:
|
|
||||||
bb.note("Fetch " + loc)
|
|
||||||
# check out sources there
|
|
||||||
bb.mkdirhier(pkgdir)
|
|
||||||
os.chdir(pkgdir)
|
|
||||||
bb.debug(1, "Running %s" % cvscmd)
|
|
||||||
myret = os.system(cvscmd)
|
|
||||||
|
|
||||||
if myret != 0:
|
|
||||||
try:
|
|
||||||
os.rmdir(moddir)
|
|
||||||
except OSError:
|
|
||||||
pass
|
|
||||||
raise FetchError(module)
|
|
||||||
|
|
||||||
os.chdir(moddir)
|
|
||||||
os.chdir('..')
|
|
||||||
# tar them up to a defined filename
|
|
||||||
myret = os.system("tar -czf %s %s" % (os.path.join(dldir,tarfn), os.path.basename(moddir)))
|
|
||||||
if myret != 0:
|
|
||||||
try:
|
|
||||||
os.unlink(tarfn)
|
|
||||||
except OSError:
|
|
||||||
pass
|
|
||||||
os.chdir(olddir)
|
|
||||||
del localdata
|
|
||||||
|
|
||||||
methods.append(Cvs())
|
|
||||||
|
|
||||||
class Bk(Fetch):
|
|
||||||
def supports(url, d):
|
|
||||||
"""Check to see if a given url can be fetched via bitkeeper.
|
|
||||||
Expects supplied url in list form, as outputted by bb.decodeurl().
|
|
||||||
"""
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(url, d))
|
|
||||||
return type in ['bk']
|
|
||||||
supports = staticmethod(supports)
|
|
||||||
|
|
||||||
methods.append(Bk())
|
|
||||||
|
|
||||||
class Local(Fetch):
|
|
||||||
def supports(url, d):
|
|
||||||
"""Check to see if a given url can be fetched in the local filesystem.
|
|
||||||
Expects supplied url in list form, as outputted by bb.decodeurl().
|
|
||||||
"""
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(url, d))
|
|
||||||
return type in ['file','patch']
|
|
||||||
supports = staticmethod(supports)
|
|
||||||
|
|
||||||
def localpath(url, d):
|
|
||||||
"""Return the local filename of a given url assuming a successful fetch.
|
|
||||||
"""
|
|
||||||
path = url.split("://")[1]
|
|
||||||
newpath = path
|
|
||||||
if path[0] != "/":
|
|
||||||
filespath = data.getVar('FILESPATH', d, 1)
|
|
||||||
if filespath:
|
|
||||||
newpath = bb.which(filespath, path)
|
|
||||||
if not newpath:
|
|
||||||
filesdir = data.getVar('FILESDIR', d, 1)
|
|
||||||
if filesdir:
|
|
||||||
newpath = os.path.join(filesdir, path)
|
|
||||||
return newpath
|
|
||||||
localpath = staticmethod(localpath)
|
|
||||||
|
|
||||||
def go(self, urls = []):
|
|
||||||
"""Fetch urls (no-op for Local method)"""
|
|
||||||
# no need to fetch local files, we'll deal with them in place.
|
|
||||||
return 1
|
|
||||||
|
|
||||||
methods.append(Local())
|
|
||||||
|
|
||||||
class Svn(Fetch):
|
|
||||||
"""Class to fetch a module or modules from svn repositories"""
|
|
||||||
def supports(url, d):
|
|
||||||
"""Check to see if a given url can be fetched with svn.
|
|
||||||
Expects supplied url in list form, as outputted by bb.decodeurl().
|
|
||||||
"""
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(url, d))
|
|
||||||
return type in ['svn']
|
|
||||||
supports = staticmethod(supports)
|
|
||||||
|
|
||||||
def localpath(url, d):
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(url, d))
|
|
||||||
if "localpath" in parm:
|
|
||||||
# if user overrides local path, use it.
|
|
||||||
return parm["localpath"]
|
|
||||||
|
|
||||||
if not "module" in parm:
|
|
||||||
raise MissingParameterError("svn method needs a 'module' parameter")
|
|
||||||
else:
|
|
||||||
module = parm["module"]
|
|
||||||
if 'rev' in parm:
|
|
||||||
revision = parm['rev']
|
|
||||||
else:
|
|
||||||
revision = ""
|
|
||||||
|
|
||||||
date = data.getVar("CVSDATE", d, 1) or data.getVar("DATE", d, 1)
|
|
||||||
|
|
||||||
return os.path.join(data.getVar("DL_DIR", d, 1),data.expand('%s_%s_%s_%s_%s.tar.gz' % ( module.replace('/', '.'), host, path.replace('/','.'), revision, date), d))
|
|
||||||
localpath = staticmethod(localpath)
|
|
||||||
|
|
||||||
def go(self, d, urls = []):
|
|
||||||
"""Fetch urls"""
|
|
||||||
if not urls:
|
|
||||||
urls = self.urls
|
|
||||||
|
|
||||||
localdata = data.createCopy(d)
|
|
||||||
data.setVar('OVERRIDES', "svn:%s" % data.getVar('OVERRIDES', localdata), localdata)
|
|
||||||
data.update_data(localdata)
|
|
||||||
|
|
||||||
for loc in urls:
|
|
||||||
(type, host, path, user, pswd, parm) = bb.decodeurl(data.expand(loc, localdata))
|
|
||||||
if not "module" in parm:
|
|
||||||
raise MissingParameterError("svn method needs a 'module' parameter")
|
|
||||||
else:
|
|
||||||
module = parm["module"]
|
|
||||||
|
|
||||||
dlfile = self.localpath(loc, localdata)
|
|
||||||
dldir = data.getVar('DL_DIR', localdata, 1)
|
|
||||||
# if local path contains the svn
|
|
||||||
# module, consider the dir above it to be the
|
|
||||||
# download directory
|
|
||||||
# pos = dlfile.find(module)
|
|
||||||
# if pos:
|
|
||||||
# dldir = dlfile[:pos]
|
|
||||||
# else:
|
|
||||||
# dldir = os.path.dirname(dlfile)
|
|
||||||
|
|
||||||
# setup svn options
|
|
||||||
options = []
|
|
||||||
if 'rev' in parm:
|
|
||||||
revision = parm['rev']
|
|
||||||
else:
|
|
||||||
revision = ""
|
|
||||||
|
|
||||||
date = data.getVar("CVSDATE", d, 1) or data.getVar("DATE", d, 1)
|
|
||||||
|
|
||||||
if "method" in parm:
|
|
||||||
method = parm["method"]
|
|
||||||
else:
|
|
||||||
method = "pserver"
|
|
||||||
|
|
||||||
if "proto" in parm:
|
|
||||||
proto = parm["proto"]
|
|
||||||
else:
|
|
||||||
proto = "svn"
|
|
||||||
|
|
||||||
svn_rsh = None
|
|
||||||
if method == "ext":
|
|
||||||
if "rsh" in parm:
|
|
||||||
svn_rsh = parm["rsh"]
|
|
||||||
|
|
||||||
tarfn = data.expand('%s_%s_%s_%s_%s.tar.gz' % (module.replace('/', '.'), host, path.replace('/', '.'), revision, date), localdata)
|
|
||||||
data.setVar('TARFILES', dlfile, localdata)
|
|
||||||
data.setVar('TARFN', tarfn, localdata)
|
|
||||||
|
|
||||||
dl = os.path.join(dldir, tarfn)
|
|
||||||
if os.access(dl, os.R_OK):
|
|
||||||
bb.debug(1, "%s already exists, skipping svn checkout." % tarfn)
|
|
||||||
continue
|
|
||||||
|
|
||||||
svn_tarball_stash = data.getVar('CVS_TARBALL_STASH', d, 1)
|
|
||||||
if svn_tarball_stash:
|
|
||||||
fetchcmd = data.getVar("FETCHCOMMAND_wget", d, 1)
|
|
||||||
uri = svn_tarball_stash + tarfn
|
|
||||||
bb.note("fetch " + uri)
|
|
||||||
fetchcmd = fetchcmd.replace("${URI}", uri)
|
|
||||||
ret = os.system(fetchcmd)
|
|
||||||
if ret == 0:
|
|
||||||
bb.note("Fetched %s from tarball stash, skipping checkout" % tarfn)
|
|
||||||
continue
|
|
||||||
|
|
||||||
olddir = os.path.abspath(os.getcwd())
|
|
||||||
os.chdir(data.expand(dldir, localdata))
|
|
||||||
|
|
||||||
# setup svnroot
|
|
||||||
# svnroot = ":" + method + ":" + user
|
|
||||||
# if pswd:
|
|
||||||
# svnroot += ":" + pswd
|
|
||||||
svnroot = host + path
|
|
||||||
|
|
||||||
data.setVar('SVNROOT', svnroot, localdata)
|
|
||||||
data.setVar('SVNCOOPTS', " ".join(options), localdata)
|
|
||||||
data.setVar('SVNMODULE', module, localdata)
|
|
||||||
svncmd = data.getVar('FETCHCOMMAND', localdata, 1)
|
|
||||||
svncmd = "svn co -r {%s} %s://%s/%s" % (date, proto, svnroot, module)
|
|
||||||
|
|
||||||
if revision:
|
|
||||||
svncmd = "svn co -r %s %s://%s/%s" % (revision, proto, svnroot, module)
|
|
||||||
if svn_rsh:
|
|
||||||
svncmd = "svn_RSH=\"%s\" %s" % (svn_rsh, svncmd)
|
|
||||||
|
|
||||||
# create temp directory
|
|
||||||
bb.debug(2, "Fetch: creating temporary directory")
|
|
||||||
bb.mkdirhier(data.expand('${WORKDIR}', localdata))
|
|
||||||
data.setVar('TMPBASE', data.expand('${WORKDIR}/oesvn.XXXXXX', localdata), localdata)
|
|
||||||
tmppipe = os.popen(data.getVar('MKTEMPDIRCMD', localdata, 1) or "false")
|
|
||||||
tmpfile = tmppipe.readline().strip()
|
|
||||||
if not tmpfile:
|
|
||||||
bb.error("Fetch: unable to create temporary directory.. make sure 'mktemp' is in the PATH.")
|
|
||||||
raise FetchError(module)
|
|
||||||
|
|
||||||
# check out sources there
|
|
||||||
os.chdir(tmpfile)
|
|
||||||
bb.note("Fetch " + loc)
|
|
||||||
bb.debug(1, "Running %s" % svncmd)
|
|
||||||
myret = os.system(svncmd)
|
|
||||||
if myret != 0:
|
|
||||||
try:
|
|
||||||
os.rmdir(tmpfile)
|
|
||||||
except OSError:
|
|
||||||
pass
|
|
||||||
raise FetchError(module)
|
|
||||||
|
|
||||||
os.chdir(os.path.join(tmpfile, os.path.dirname(module)))
|
|
||||||
# tar them up to a defined filename
|
|
||||||
myret = os.system("tar -czf %s %s" % (os.path.join(dldir,tarfn), os.path.basename(module)))
|
|
||||||
if myret != 0:
|
|
||||||
try:
|
|
||||||
os.unlink(tarfn)
|
|
||||||
except OSError:
|
|
||||||
pass
|
|
||||||
# cleanup
|
|
||||||
os.system('rm -rf %s' % tmpfile)
|
|
||||||
os.chdir(olddir)
|
|
||||||
del localdata
|
|
||||||
|
|
||||||
methods.append(Svn())
|
|
Loading…
Reference in New Issue