mirror of
git://git.yoctoproject.org/layerindex-web.git
synced 2025-07-19 20:59:01 +02:00
layerindex: Reorder code in order to use in the Recipe reporting system
For use the same code base for Recipe reporting system moves non-WEB code outside layerindex, create lib directory that contains common code for Layerindex and RRS. Create scripts directory that contains common infraestructure to populate db and tool scripts, modularize update.py script move Layerindex update to own folder. Update references for utils and recipeparse modules to point lib directory. Add missing GitPython into requirements.txt. Signed-off-by: Aníbal Limón <anibal.limon@linux.intel.com>
This commit is contained in:
parent
219bac4229
commit
60fd912b44
2
README
2
README
|
@ -113,7 +113,7 @@ Usage
|
|||
|
||||
On a regular basis you need to run the update script:
|
||||
|
||||
path/to/layerindex/update.py
|
||||
path/to/scripts/update.py
|
||||
|
||||
This will fetch all of the layer repositories, analyse their contents
|
||||
and update the database with the results. Note that if you set up more
|
||||
|
|
|
@ -4,6 +4,10 @@
|
|||
#
|
||||
# Licensed under the MIT license, see COPYING.MIT for details
|
||||
|
||||
import sys
|
||||
import os
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '../lib')))
|
||||
|
||||
from django.shortcuts import get_object_or_404, render
|
||||
from django.http import HttpResponse, HttpResponseRedirect, HttpResponseForbidden, Http404
|
||||
from django.core.urlresolvers import reverse, reverse_lazy, resolve
|
||||
|
@ -218,7 +222,8 @@ def bulk_change_patch_view(request, pk):
|
|||
# FIXME this couples the web server and machine running the update script together,
|
||||
# but given that it's a separate script the way is open to decouple them in future
|
||||
try:
|
||||
ret = utils.runcmd('python bulkchange.py %d %s' % (int(pk), settings.TEMP_BASE_DIR), os.path.dirname(__file__))
|
||||
bulkchange = os.path.join(BASE_DIR, 'scripts', 'bulkchange.py')
|
||||
ret = utils.runcmd('python %s %d %s' % (bulkchange, int(pk), settings.TEMP_BASE_DIR), os.path.dirname(__file__))
|
||||
if ret:
|
||||
fn = ret.splitlines()[-1]
|
||||
if os.path.exists(fn):
|
||||
|
|
|
@ -177,3 +177,11 @@ def detect_file_type(path, subdir_start):
|
|||
|
||||
return (None, None, None)
|
||||
|
||||
def split_recipe_fn(path):
|
||||
splitfn = os.path.basename(path).split('.bb')[0].split('_', 2)
|
||||
pn = splitfn[0]
|
||||
if len(splitfn) > 1:
|
||||
pv = splitfn[1]
|
||||
else:
|
||||
pv = "1.0"
|
||||
return (pn, pv)
|
|
@ -17,3 +17,4 @@ python-nvd3==0.12.2
|
|||
regex==2014.06.28
|
||||
six==1.7.3
|
||||
wsgiref==0.1.2
|
||||
GitPython==0.3.2.1
|
||||
|
|
3
layerindex/bulkchange.py → scripts/bulkchange.py
Normal file → Executable file
3
layerindex/bulkchange.py → scripts/bulkchange.py
Normal file → Executable file
|
@ -9,6 +9,9 @@
|
|||
import sys
|
||||
import os
|
||||
import os.path
|
||||
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '../lib')))
|
||||
|
||||
import tempfile
|
||||
import tarfile
|
||||
import textwrap
|
312
layerindex/update.py → scripts/layerindex_update/__init__.py
Executable file → Normal file
312
layerindex/update.py → scripts/layerindex_update/__init__.py
Executable file → Normal file
|
@ -1,241 +1,55 @@
|
|||
#!/usr/bin/env python
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
# Fetch layer repositories and update layer index database
|
||||
#
|
||||
# Copyright (C) 2013 Intel Corporation
|
||||
# Copyright (C) 2013 - 2015 Intel Corporation
|
||||
# Author: Paul Eggleton <paul.eggleton@linux.intel.com>
|
||||
# Contributor: Aníbal Limón <anibal.limon@linux.intel.com>
|
||||
#
|
||||
# Licensed under the MIT license, see COPYING.MIT for details
|
||||
|
||||
|
||||
import sys
|
||||
import os.path
|
||||
import optparse
|
||||
import logging
|
||||
from datetime import datetime
|
||||
import re
|
||||
import tempfile
|
||||
import shutil
|
||||
from datetime import datetime
|
||||
|
||||
from distutils.version import LooseVersion
|
||||
|
||||
import settings
|
||||
from django.db import transaction
|
||||
|
||||
import utils
|
||||
import recipeparse
|
||||
|
||||
import warnings
|
||||
warnings.filterwarnings("ignore", category=DeprecationWarning)
|
||||
|
||||
logger = utils.logger_create('LayerIndexUpdate')
|
||||
from layerindex.models import LayerItem, LayerBranch, Recipe, \
|
||||
RecipeFileDependency, Machine, BBAppend, BBClass
|
||||
|
||||
# Ensure PythonGit is installed (buildhistory_analysis needs it)
|
||||
try:
|
||||
import git
|
||||
if LooseVersion(git.__version__) < '0.3.1':
|
||||
logger.error("Version of GitPython is too old, please install" \
|
||||
" GitPython (python-git) 0.3.1 or later in order to use this script")
|
||||
sys.exit(1)
|
||||
except ImportError:
|
||||
logger.error("Please install PythonGit 0.3.1 or later in order to use this script")
|
||||
sys.exit(1)
|
||||
|
||||
class LayerindexUpdater:
|
||||
def __init__(self, _options, _fetchdir, _layerquery, _fetchedrepos,
|
||||
_failedrepos, _logger):
|
||||
global options
|
||||
global fetchdir
|
||||
global layerquery
|
||||
global fetchedrepos
|
||||
global failedrepos
|
||||
global logger
|
||||
|
||||
def check_machine_conf(path, subdir_start):
|
||||
subpath = path[len(subdir_start):]
|
||||
res = conf_re.match(subpath)
|
||||
if res:
|
||||
return res.group(1)
|
||||
return None
|
||||
|
||||
def split_recipe_fn(path):
|
||||
splitfn = os.path.basename(path).split('.bb')[0].split('_', 2)
|
||||
pn = splitfn[0]
|
||||
if len(splitfn) > 1:
|
||||
pv = splitfn[1]
|
||||
else:
|
||||
pv = "1.0"
|
||||
return (pn, pv)
|
||||
|
||||
def update_recipe_file(data, path, recipe, layerdir_start, repodir):
|
||||
fn = str(os.path.join(path, recipe.filename))
|
||||
try:
|
||||
logger.debug('Updating recipe %s' % fn)
|
||||
envdata = bb.cache.Cache.loadDataFull(fn, [], data)
|
||||
envdata.setVar('SRCPV', 'X')
|
||||
recipe.pn = envdata.getVar("PN", True)
|
||||
recipe.pv = envdata.getVar("PV", True)
|
||||
recipe.summary = envdata.getVar("SUMMARY", True)
|
||||
recipe.description = envdata.getVar("DESCRIPTION", True)
|
||||
recipe.section = envdata.getVar("SECTION", True)
|
||||
recipe.license = envdata.getVar("LICENSE", True)
|
||||
recipe.homepage = envdata.getVar("HOMEPAGE", True)
|
||||
recipe.bugtracker = envdata.getVar("BUGTRACKER", True) or ""
|
||||
recipe.provides = envdata.getVar("PROVIDES", True) or ""
|
||||
recipe.bbclassextend = envdata.getVar("BBCLASSEXTEND", True) or ""
|
||||
recipe.save()
|
||||
|
||||
# Get file dependencies within this layer
|
||||
deps = envdata.getVar('__depends', True)
|
||||
filedeps = []
|
||||
for depstr, date in deps:
|
||||
found = False
|
||||
if depstr.startswith(layerdir_start) and not depstr.endswith('/conf/layer.conf'):
|
||||
filedeps.append(os.path.relpath(depstr, repodir))
|
||||
from layerindex.models import RecipeFileDependency
|
||||
RecipeFileDependency.objects.filter(recipe=recipe).delete()
|
||||
for filedep in filedeps:
|
||||
recipedep = RecipeFileDependency()
|
||||
recipedep.layerbranch = recipe.layerbranch
|
||||
recipedep.recipe = recipe
|
||||
recipedep.path = filedep
|
||||
recipedep.save()
|
||||
except KeyboardInterrupt:
|
||||
raise
|
||||
except BaseException as e:
|
||||
if not recipe.pn:
|
||||
recipe.pn = recipe.filename[:-3].split('_')[0]
|
||||
logger.error("Unable to read %s: %s", fn, str(e))
|
||||
|
||||
def update_machine_conf_file(path, machine):
|
||||
logger.debug('Updating machine %s' % path)
|
||||
desc = ""
|
||||
with open(path, 'r') as f:
|
||||
for line in f:
|
||||
if line.startswith('#@NAME:'):
|
||||
desc = line[7:].strip()
|
||||
if line.startswith('#@DESCRIPTION:'):
|
||||
desc = line[14:].strip()
|
||||
desc = re.sub(r'Machine configuration for( running)*( an)*( the)*', '', desc)
|
||||
break
|
||||
machine.description = desc
|
||||
|
||||
def main():
|
||||
if LooseVersion(git.__version__) < '0.3.1':
|
||||
logger.error("Version of GitPython is too old, please install GitPython (python-git) 0.3.1 or later in order to use this script")
|
||||
sys.exit(1)
|
||||
|
||||
|
||||
parser = optparse.OptionParser(
|
||||
usage = """
|
||||
%prog [options]""")
|
||||
|
||||
parser.add_option("-b", "--branch",
|
||||
help = "Specify branch to update",
|
||||
action="store", dest="branch", default='master')
|
||||
parser.add_option("-l", "--layer",
|
||||
help = "Specify layers to update (use commas to separate multiple). Default is all published layers.",
|
||||
action="store", dest="layers")
|
||||
parser.add_option("-r", "--reload",
|
||||
help = "Reload recipe data instead of updating since last update",
|
||||
action="store_true", dest="reload")
|
||||
parser.add_option("", "--fullreload",
|
||||
help = "Discard existing recipe data and fetch it from scratch",
|
||||
action="store_true", dest="fullreload")
|
||||
parser.add_option("-n", "--dry-run",
|
||||
help = "Don't write any data back to the database",
|
||||
action="store_true", dest="dryrun")
|
||||
parser.add_option("-x", "--nofetch",
|
||||
help = "Don't fetch repositories",
|
||||
action="store_true", dest="nofetch")
|
||||
parser.add_option("", "--nocheckout",
|
||||
help = "Don't check out branches",
|
||||
action="store_true", dest="nocheckout")
|
||||
parser.add_option("-d", "--debug",
|
||||
help = "Enable debug output",
|
||||
action="store_const", const=logging.DEBUG, dest="loglevel", default=logging.INFO)
|
||||
parser.add_option("-q", "--quiet",
|
||||
help = "Hide all output except error messages",
|
||||
action="store_const", const=logging.ERROR, dest="loglevel")
|
||||
|
||||
options, args = parser.parse_args(sys.argv)
|
||||
if len(args) > 1:
|
||||
logger.error('unexpected argument "%s"' % args[1])
|
||||
parser.print_help()
|
||||
sys.exit(1)
|
||||
|
||||
if options.fullreload:
|
||||
options.reload = True
|
||||
|
||||
utils.setup_django()
|
||||
import settings
|
||||
from layerindex.models import LayerItem, LayerBranch, Recipe, RecipeFileDependency, Machine, BBAppend, BBClass
|
||||
from django.db import transaction
|
||||
|
||||
logger.setLevel(options.loglevel)
|
||||
|
||||
branch = utils.get_branch(options.branch)
|
||||
if not branch:
|
||||
logger.error("Specified branch %s is not valid" % options.branch)
|
||||
sys.exit(1)
|
||||
|
||||
fetchdir = settings.LAYER_FETCH_DIR
|
||||
if not fetchdir:
|
||||
logger.error("Please set LAYER_FETCH_DIR in settings.py")
|
||||
sys.exit(1)
|
||||
|
||||
if options.layers:
|
||||
layerquery = LayerItem.objects.filter(classic=False).filter(name__in=options.layers.split(','))
|
||||
if layerquery.count() == 0:
|
||||
logger.error('No layers matching specified query "%s"' % options.layers)
|
||||
sys.exit(1)
|
||||
else:
|
||||
layerquery = LayerItem.objects.filter(classic=False).filter(status='P')
|
||||
if layerquery.count() == 0:
|
||||
logger.info("No published layers to update")
|
||||
sys.exit(1)
|
||||
|
||||
if not os.path.exists(fetchdir):
|
||||
os.makedirs(fetchdir)
|
||||
fetchedrepos = []
|
||||
failedrepos = []
|
||||
|
||||
lockfn = os.path.join(fetchdir, "layerindex.lock")
|
||||
lockfile = utils.lock_file(lockfn)
|
||||
if not lockfile:
|
||||
logger.error("Layer index lock timeout expired")
|
||||
sys.exit(1)
|
||||
try:
|
||||
bitbakepath = os.path.join(fetchdir, 'bitbake')
|
||||
|
||||
if not options.nofetch:
|
||||
# Fetch latest metadata from repositories
|
||||
for layer in layerquery:
|
||||
# Handle multiple layers in a single repo
|
||||
urldir = layer.get_fetch_dir()
|
||||
repodir = os.path.join(fetchdir, urldir)
|
||||
if not (layer.vcs_url in fetchedrepos or layer.vcs_url in failedrepos):
|
||||
logger.info("Fetching remote repository %s" % layer.vcs_url)
|
||||
out = None
|
||||
try:
|
||||
if not os.path.exists(repodir):
|
||||
out = utils.runcmd("git clone %s %s" % (layer.vcs_url, urldir), fetchdir, logger=logger)
|
||||
else:
|
||||
out = utils.runcmd("git fetch", repodir, logger=logger)
|
||||
except Exception as e:
|
||||
logger.error("Fetch of layer %s failed: %s" % (layer.name, str(e)))
|
||||
failedrepos.append(layer.vcs_url)
|
||||
continue
|
||||
fetchedrepos.append(layer.vcs_url)
|
||||
|
||||
if not fetchedrepos:
|
||||
logger.error("No repositories could be fetched, exiting")
|
||||
sys.exit(1)
|
||||
|
||||
logger.info("Fetching bitbake from remote repository %s" % settings.BITBAKE_REPO_URL)
|
||||
if not os.path.exists(bitbakepath):
|
||||
out = utils.runcmd("git clone %s %s" % (settings.BITBAKE_REPO_URL, 'bitbake'), fetchdir, logger=logger)
|
||||
else:
|
||||
out = utils.runcmd("git fetch", bitbakepath, logger=logger)
|
||||
|
||||
try:
|
||||
(tinfoil, tempdir) = recipeparse.init_parser(settings, branch, bitbakepath, nocheckout=options.nocheckout, logger=logger)
|
||||
except recipeparse.RecipeParseError as e:
|
||||
logger.error(str(e))
|
||||
sys.exit(1)
|
||||
|
||||
# Clear the default value of SUMMARY so that we can use DESCRIPTION instead if it hasn't been set
|
||||
tinfoil.config_data.setVar('SUMMARY', '')
|
||||
# Clear the default value of DESCRIPTION so that we can see where it's not set
|
||||
tinfoil.config_data.setVar('DESCRIPTION', '')
|
||||
# Clear the default value of HOMEPAGE ('unknown')
|
||||
tinfoil.config_data.setVar('HOMEPAGE', '')
|
||||
# Set a blank value for LICENSE so that it doesn't cause the parser to die (e.g. with meta-ti -
|
||||
# why won't they just fix that?!)
|
||||
tinfoil.config_data.setVar('LICENSE', '')
|
||||
options = _options
|
||||
fetchdir = _fetchdir
|
||||
layerquery = _layerquery
|
||||
fetchedrepos = _fetchedrepos
|
||||
failedrepos = _failedrepos
|
||||
logger = _logger
|
||||
|
||||
def run(self, tinfoil):
|
||||
# Process and extract data from each layer
|
||||
for layer in layerquery:
|
||||
transaction.enter_transaction_management()
|
||||
|
@ -540,7 +354,7 @@ def main():
|
|||
|
||||
for added in layerrecipes_add:
|
||||
# This is good enough without actually parsing the file
|
||||
(pn, pv) = split_recipe_fn(added)
|
||||
(pn, pv) = recipeparse.split_recipe_fn(added)
|
||||
oldid = -1
|
||||
for deleted in layerrecipes_delete:
|
||||
if deleted['pn'] == pn:
|
||||
|
@ -593,12 +407,62 @@ def main():
|
|||
finally:
|
||||
transaction.leave_transaction_management()
|
||||
|
||||
finally:
|
||||
utils.unlock_file(lockfile)
|
||||
def check_machine_conf(path, subdir_start):
|
||||
subpath = path[len(subdir_start):]
|
||||
res = conf_re.match(subpath)
|
||||
if res:
|
||||
return res.group(1)
|
||||
return None
|
||||
|
||||
shutil.rmtree(tempdir)
|
||||
sys.exit(0)
|
||||
def update_recipe_file(data, path, recipe, layerdir_start, repodir):
|
||||
fn = str(os.path.join(path, recipe.filename))
|
||||
try:
|
||||
logger.debug('Updating recipe %s' % fn)
|
||||
envdata = bb.cache.Cache.loadDataFull(fn, [], data)
|
||||
envdata.setVar('SRCPV', 'X')
|
||||
recipe.pn = envdata.getVar("PN", True)
|
||||
recipe.pv = envdata.getVar("PV", True)
|
||||
recipe.summary = envdata.getVar("SUMMARY", True)
|
||||
recipe.description = envdata.getVar("DESCRIPTION", True)
|
||||
recipe.section = envdata.getVar("SECTION", True)
|
||||
recipe.license = envdata.getVar("LICENSE", True)
|
||||
recipe.homepage = envdata.getVar("HOMEPAGE", True)
|
||||
recipe.bugtracker = envdata.getVar("BUGTRACKER", True) or ""
|
||||
recipe.provides = envdata.getVar("PROVIDES", True) or ""
|
||||
recipe.bbclassextend = envdata.getVar("BBCLASSEXTEND", True) or ""
|
||||
recipe.save()
|
||||
|
||||
# Get file dependencies within this layer
|
||||
deps = envdata.getVar('__depends', True)
|
||||
filedeps = []
|
||||
for depstr, date in deps:
|
||||
found = False
|
||||
if depstr.startswith(layerdir_start) and not depstr.endswith('/conf/layer.conf'):
|
||||
filedeps.append(os.path.relpath(depstr, repodir))
|
||||
from layerindex.models import RecipeFileDependency
|
||||
RecipeFileDependency.objects.filter(recipe=recipe).delete()
|
||||
for filedep in filedeps:
|
||||
recipedep = RecipeFileDependency()
|
||||
recipedep.layerbranch = recipe.layerbranch
|
||||
recipedep.recipe = recipe
|
||||
recipedep.path = filedep
|
||||
recipedep.save()
|
||||
except KeyboardInterrupt:
|
||||
raise
|
||||
except BaseException as e:
|
||||
if not recipe.pn:
|
||||
recipe.pn = recipe.filename[:-3].split('_')[0]
|
||||
logger.error("Unable to read %s: %s", fn, str(e))
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
||||
def update_machine_conf_file(path, machine):
|
||||
logger.debug('Updating machine %s' % path)
|
||||
desc = ""
|
||||
with open(path, 'r') as f:
|
||||
for line in f:
|
||||
if line.startswith('#@NAME:'):
|
||||
desc = line[7:].strip()
|
||||
if line.startswith('#@DESCRIPTION:'):
|
||||
desc = line[14:].strip()
|
||||
desc = re.sub(r'Machine configuration for( running)*( an)*( the)*', '', desc)
|
||||
break
|
||||
machine.description = desc
|
0
layerindex/recipedesc.py → scripts/recipedesc.py
Normal file → Executable file
0
layerindex/recipedesc.py → scripts/recipedesc.py
Normal file → Executable file
|
@ -11,7 +11,7 @@
|
|||
import sys
|
||||
import os.path
|
||||
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '..')))
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '../../lib')))
|
||||
|
||||
import optparse
|
||||
import logging
|
|
@ -11,7 +11,7 @@
|
|||
import sys
|
||||
import os.path
|
||||
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '..')))
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '../../lib')))
|
||||
|
||||
import optparse
|
||||
import re
|
|
@ -11,7 +11,7 @@
|
|||
import sys
|
||||
import os.path
|
||||
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '..')))
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '../../lib')))
|
||||
|
||||
import optparse
|
||||
import re
|
|
@ -11,7 +11,7 @@
|
|||
import sys
|
||||
import os.path
|
||||
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '..')))
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '../../lib')))
|
||||
|
||||
import optparse
|
||||
import re
|
|
@ -10,7 +10,7 @@
|
|||
import sys
|
||||
import os.path
|
||||
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '..')))
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '../../lib')))
|
||||
|
||||
import optparse
|
||||
import re
|
193
scripts/update.py
Executable file
193
scripts/update.py
Executable file
|
@ -0,0 +1,193 @@
|
|||
#!/usr/bin/env python
|
||||
# -*- coding: utf-8 -*-
|
||||
|
||||
# Fetch layer repositories and update layer index database
|
||||
#
|
||||
# Copyright (C) 2013 - 2015 Intel Corporation
|
||||
# Author: Paul Eggleton <paul.eggleton@linux.intel.com>
|
||||
# Contributor: Aníbal Limón <anibal.limon@linux.intel.com>
|
||||
#
|
||||
# Licensed under the MIT license, see COPYING.MIT for details
|
||||
|
||||
import sys
|
||||
import os.path
|
||||
import shutil
|
||||
|
||||
sys.path.insert(0, os.path.realpath(os.path.join(os.path.dirname(__file__), '../lib')))
|
||||
|
||||
import optparse
|
||||
import logging
|
||||
import utils
|
||||
|
||||
utils.setup_django()
|
||||
import settings
|
||||
|
||||
import warnings
|
||||
warnings.filterwarnings("ignore", category = DeprecationWarning)
|
||||
|
||||
logger = utils.logger_create('LayerindexUpdate')
|
||||
|
||||
from layerindex_update import LayerindexUpdater
|
||||
|
||||
def main():
|
||||
parser = optparse.OptionParser(
|
||||
usage = """
|
||||
%prog [options]""")
|
||||
|
||||
parser.add_option("-b", "--branch",
|
||||
help = "Specify branch to update",
|
||||
action="store", dest="branch", default='master')
|
||||
parser.add_option("-l", "--layer",
|
||||
help = "Specify layers to update (use commas to separate multiple). Default is all published layers.",
|
||||
action="store", dest="layers")
|
||||
parser.add_option("-r", "--reload",
|
||||
help = "Reload recipe data instead of updating since last update",
|
||||
action="store_true", dest="reload")
|
||||
parser.add_option("", "--fullreload",
|
||||
help = "Discard existing recipe data and fetch it from scratch",
|
||||
action="store_true", dest="fullreload")
|
||||
parser.add_option("-n", "--dry-run",
|
||||
help = "Don't write any data back to the database",
|
||||
action="store_true", dest="dryrun")
|
||||
parser.add_option("-x", "--nofetch",
|
||||
help = "Don't fetch repositories",
|
||||
action="store_true", dest="nofetch")
|
||||
parser.add_option("", "--nocheckout",
|
||||
help = "Don't check out branches",
|
||||
action="store_true", dest="nocheckout")
|
||||
parser.add_option("-d", "--debug",
|
||||
help = "Enable debug output",
|
||||
action="store_const", const=logging.DEBUG, dest="loglevel", default=logging.INFO)
|
||||
parser.add_option("-q", "--quiet",
|
||||
help = "Hide all output except error messages",
|
||||
action="store_const", const=logging.ERROR, dest="loglevel")
|
||||
|
||||
options, args = parser.parse_args(sys.argv)
|
||||
if len(args) > 1:
|
||||
logger.error('unexpected argument "%s"' % args[1])
|
||||
parser.print_help()
|
||||
sys.exit(1)
|
||||
|
||||
if options.fullreload:
|
||||
options.reload = True
|
||||
|
||||
logger.setLevel(options.loglevel)
|
||||
|
||||
branch = utils.get_branch(options.branch)
|
||||
if not branch:
|
||||
logger.error("Specified branch %s is not valid" % options.branch)
|
||||
sys.exit(1)
|
||||
|
||||
fetchdir = settings.LAYER_FETCH_DIR
|
||||
if not fetchdir:
|
||||
logger.error("Please set LAYER_FETCH_DIR in settings.py")
|
||||
sys.exit(1)
|
||||
if not os.path.exists(fetchdir):
|
||||
os.makedirs(fetchdir)
|
||||
|
||||
lockfn = os.path.join(fetchdir, "layerindex.lock")
|
||||
lockfile = utils.lock_file(lockfn)
|
||||
if not lockfile:
|
||||
logger.error("Layer index lock timeout expired")
|
||||
sys.exit(1)
|
||||
|
||||
bitbakepath = update_repo(fetchdir, 'bitbake', settings.BITBAKE_REPO_URL, logger)
|
||||
(layerquery, fetchedrepos, failedrepos) = update_layers(options, fetchdir, logger)
|
||||
(tinfoil, tempdir) = get_tinfoil(branch, bitbakepath, options, logger)
|
||||
|
||||
layerindex_updater = LayerindexUpdater(options, fetchdir, layerquery, fetchedrepos,
|
||||
failedrepos, logger)
|
||||
layerindex_updater.run(tinfoil)
|
||||
|
||||
shutil.rmtree(tempdir)
|
||||
utils.unlock_file(lockfile)
|
||||
|
||||
def update_repo(fetchdir, repo_name, repo_url, logger):
|
||||
path = os.path.join(fetchdir, repo_name)
|
||||
|
||||
logger.info("Fetching %s from remote repository %s"
|
||||
% (repo_name, repo_url))
|
||||
if not os.path.exists(path):
|
||||
out = utils.runcmd("git clone %s %s" %
|
||||
(repo_url, repo_name), fetchdir,
|
||||
logger = logger)
|
||||
else:
|
||||
out = utils.runcmd("git fetch", path, logger = logger)
|
||||
|
||||
return path
|
||||
|
||||
def update_layers(options, fetchdir, logger):
|
||||
from layerindex.models import LayerItem
|
||||
|
||||
fetchedrepos = []
|
||||
failedrepos = []
|
||||
|
||||
if options.layers:
|
||||
layerquery = LayerItem.objects.filter(classic =
|
||||
False).filter(name__in = options.layers.split(','))
|
||||
else:
|
||||
layerquery = LayerItem.objects.filter(classic =
|
||||
False).filter(status = 'P') # All published layers
|
||||
|
||||
if layerquery.count() == 0:
|
||||
logger.info("No published layers to update")
|
||||
sys.exit(1)
|
||||
|
||||
# Fetch latest metadata from repositories
|
||||
# Handle multiple layers in a single repo
|
||||
for layer in layerquery:
|
||||
urldir = layer.get_fetch_dir()
|
||||
repodir = os.path.join(fetchdir, urldir)
|
||||
|
||||
if not (layer.vcs_url in fetchedrepos or layer.vcs_url in
|
||||
failedrepos):
|
||||
logger.info("Fetching remote repository %s" %
|
||||
layer.vcs_url)
|
||||
|
||||
out = None
|
||||
try:
|
||||
if not os.path.exists(repodir):
|
||||
out = utils.runcmd("git clone %s %s" %
|
||||
(layer.vcs_url, urldir), fetchdir,
|
||||
logger = logger)
|
||||
else:
|
||||
out = utils.runcmd("git fetch", repodir, logger =
|
||||
logger)
|
||||
except Exception as e:
|
||||
logger.error("Fetch of layer %s failed: %s" %
|
||||
(layer.name, str(e)))
|
||||
failedrepos.append(layer.vcs_url)
|
||||
continue
|
||||
|
||||
fetchedrepos.append(layer.vcs_url)
|
||||
|
||||
if not fetchedrepos:
|
||||
logger.error("No repositories could be fetched, exiting")
|
||||
sys.exit(1)
|
||||
|
||||
return (layerquery, fetchedrepos, failedrepos)
|
||||
|
||||
def get_tinfoil(branch, bitbakepath, options, logger):
|
||||
import recipeparse
|
||||
try:
|
||||
(tinfoil, tempdir) = recipeparse.init_parser(settings, branch, bitbakepath,
|
||||
nocheckout = options.nocheckout, logger = logger)
|
||||
except recipeparse.RecipeParseError as e:
|
||||
logger.error(str(e))
|
||||
sys.exit(1)
|
||||
|
||||
# Clear the default value of SUMMARY so that we can use DESCRIPTION instead
|
||||
# if it hasn't been set
|
||||
tinfoil.config_data.setVar('SUMMARY', '')
|
||||
# Clear the default value of DESCRIPTION so that we can see where it's not set
|
||||
tinfoil.config_data.setVar('DESCRIPTION', '')
|
||||
# Clear the default value of HOMEPAGE ('unknown')
|
||||
tinfoil.config_data.setVar('HOMEPAGE', '')
|
||||
# Set a blank value for LICENSE so that it doesn't cause the parser to die
|
||||
# (e.g. with meta-ti -, why won't they just fix that?!)
|
||||
tinfoil.config_data.setVar('LICENSE', '')
|
||||
|
||||
return (tinfoil, tempdir)
|
||||
|
||||
if __name__ == "__main__":
|
||||
main()
|
Loading…
Reference in New Issue
Block a user