#! /usr/bin/env python # ex:ts=4:sw=4:sts=4:et # -*- tab-width: 4; c-basic-offset: 4; indent-tabs-mode: nil -*- # # BitBake Toaster Implementation # # Copyright (C) 2016 Intel Corporation # # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License version 2 as # published by the Free Software Foundation. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License along # with this program; if not, write to the Free Software Foundation, Inc., # 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA. # Tests were part of openembedded-core oe selftest Authored by: Lucian Musat # Ionut Chisanovici, Paul Eggleton and Cristian Iorga import os from django.db.models import Q from orm.models import Target_Image_File, Target_Installed_Package, Task from orm.models import Package_Dependency, Recipe_Dependency, Build from orm.models import Task_Dependency, Package, Target, Recipe from orm.models import CustomImagePackage from tests.builds.buildtest import BuildTest class BuildCoreImageMinimal(BuildTest): """Build core-image-minimal and test the results""" def setUp(self): self.completed_build = self.build("core-image-minimal") # Check if build name is unique - tc_id=795 def test_Build_Unique_Name(self): all_builds = Build.objects.all().count() distinct_builds = Build.objects.values('id').distinct().count() self.assertEqual(distinct_builds, all_builds, msg='Build name is not unique') # Check if build cooker log path is unique - tc_id=819 def test_Build_Unique_Cooker_Log_Path(self): distinct_path = Build.objects.values( 'cooker_log_path').distinct().count() total_builds = Build.objects.values('id').count() self.assertEqual(distinct_path, total_builds, msg='Build cooker log path is not unique') # Check if task order is unique for one build - tc=824 def test_Task_Unique_Order(self): total_task_order = Task.objects.filter( build=self.built).values('order').count() distinct_task_order = Task.objects.filter( build=self.completed_build).values('order').distinct().count() self.assertEqual(total_task_order, distinct_task_order, msg='Errors task order is not unique') # Check task order sequence for one build - tc=825 def test_Task_Order_Sequence(self): cnt_err = [] tasks = Task.objects.filter( Q(build=self.completed_build), ~Q(order=None), ~Q(task_name__contains='_setscene') ).values('id', 'order').order_by("order") cnt_tasks = 0 for task in tasks: cnt_tasks += 1 if (task['order'] != cnt_tasks): cnt_err.append(task['id']) self.assertEqual( len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) # Check if disk_io matches the difference between EndTimeIO and # StartTimeIO in build stats - tc=828 # def test_Task_Disk_IO_TC828(self): # Check if outcome = 2 (SSTATE) then sstate_result must be 3 (RESTORED) - # tc=832 def test_Task_If_Outcome_2_Sstate_Result_Must_Be_3(self): tasks = Task.objects.filter(outcome=2).values('id', 'sstate_result') cnt_err = [] for task in tasks: if (task['sstate_result'] != 3): cnt_err.append(task['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) # Check if outcome = 1 (COVERED) or 3 (EXISTING) then sstate_result must # be 0 (SSTATE_NA) - tc=833 def test_Task_If_Outcome_1_3_Sstate_Result_Must_Be_0(self): tasks = Task.objects.filter( outcome__in=(Task.OUTCOME_COVERED, Task.OUTCOME_PREBUILT)).values('id', 'task_name', 'sstate_result') cnt_err = [] for task in tasks: if (task['sstate_result'] != Task.SSTATE_NA and task['sstate_result'] != Task.SSTATE_MISS): cnt_err.append({'id': task['id'], 'name': task['task_name'], 'sstate_result': task['sstate_result']}) self.assertEqual(len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) # Check if outcome is 0 (SUCCESS) or 4 (FAILED) then sstate_result must be # 0 (NA), 1 (MISS) or 2 (FAILED) - tc=834 def test_Task_If_Outcome_0_4_Sstate_Result_Must_Be_0_1_2(self): tasks = Task.objects.filter( outcome__in=(0, 4)).values('id', 'sstate_result') cnt_err = [] for task in tasks: if (task['sstate_result'] not in [0, 1, 2]): cnt_err.append(task['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) # Check if task_executed = TRUE (1), script_type must be 0 (CODING_NA), 2 # (CODING_PYTHON), 3 (CODING_SHELL) - tc=891 def test_Task_If_Task_Executed_True_Script_Type_0_2_3(self): tasks = Task.objects.filter( task_executed=1).values('id', 'script_type') cnt_err = [] for task in tasks: if (task['script_type'] not in [0, 2, 3]): cnt_err.append(task['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) # Check if task_executed = TRUE (1), outcome must be 0 (SUCCESS) or 4 # (FAILED) - tc=836 def test_Task_If_Task_Executed_True_Outcome_0_4(self): tasks = Task.objects.filter(task_executed=1).values('id', 'outcome') cnt_err = [] for task in tasks: if (task['outcome'] not in [0, 4]): cnt_err.append(task['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) # Check if task_executed = FALSE (0), script_type must be 0 - tc=890 def test_Task_If_Task_Executed_False_Script_Type_0(self): tasks = Task.objects.filter( task_executed=0).values('id', 'script_type') cnt_err = [] for task in tasks: if (task['script_type'] != 0): cnt_err.append(task['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) # Check if task_executed = FALSE (0) and build outcome = SUCCEEDED (0), # task outcome must be 1 (COVERED), 2 (CACHED), 3 (PREBUILT), 5 (EMPTY) - # tc=837 def test_Task_If_Task_Executed_False_Outcome_1_2_3_5(self): builds = Build.objects.filter(outcome=0).values('id') cnt_err = [] for build in builds: tasks = Task.objects.filter( build=build['id'], task_executed=0).values('id', 'outcome') for task in tasks: if (task['outcome'] not in [1, 2, 3, 5]): cnt_err.append(task['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for task id: %s' % cnt_err) # Key verification - tc=888 def test_Target_Installed_Package(self): rows = Target_Installed_Package.objects.values('id', 'target_id', 'package_id') cnt_err = [] for row in rows: target = Target.objects.filter(id=row['target_id']).values('id') package = Package.objects.filter(id=row['package_id']).values('id') if (not target or not package): cnt_err.append(row['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for target installed package id: %s' % cnt_err) # Key verification - tc=889 def test_Task_Dependency(self): rows = Task_Dependency.objects.values('id', 'task_id', 'depends_on_id') cnt_err = [] for row in rows: task_id = Task.objects.filter(id=row['task_id']).values('id') depends_on_id = Task.objects.filter( id=row['depends_on_id']).values('id') if (not task_id or not depends_on_id): cnt_err.append(row['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for task dependency id: %s' % cnt_err) # Check if build target file_name is populated only if is_image=true AND # orm_build.outcome=0 then if the file exists and its size matches # the file_size value. Need to add the tc in the test run def test_Target_File_Name_Populated(self): builds = Build.objects.filter(outcome=0).values('id') for build in builds: targets = Target.objects.filter( build_id=build['id'], is_image=1).values('id') for target in targets: target_files = Target_Image_File.objects.filter( target_id=target['id']).values('id', 'file_name', 'file_size') cnt_err = [] for file_info in target_files: target_id = file_info['id'] target_file_name = file_info['file_name'] target_file_size = file_info['file_size'] if (not target_file_name or not target_file_size): cnt_err.append(target_id) else: if (not os.path.exists(target_file_name)): cnt_err.append(target_id) else: if (os.path.getsize(target_file_name) != target_file_size): cnt_err.append(target_id) self.assertEqual(len(cnt_err), 0, msg='Errors for target image file id: %s' % cnt_err) # Key verification - tc=884 def test_Package_Dependency(self): cnt_err = [] deps = Package_Dependency.objects.values( 'id', 'package_id', 'depends_on_id') for dep in deps: if (dep['package_id'] == dep['depends_on_id']): cnt_err.append(dep['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for package dependency id: %s' % cnt_err) # Recipe key verification, recipe name does not depends on a recipe having # the same name - tc=883 def test_Recipe_Dependency(self): deps = Recipe_Dependency.objects.values( 'id', 'recipe_id', 'depends_on_id') cnt_err = [] for dep in deps: if (not dep['recipe_id'] or not dep['depends_on_id']): cnt_err.append(dep['id']) else: name = Recipe.objects.filter( id=dep['recipe_id']).values('name') dep_name = Recipe.objects.filter( id=dep['depends_on_id']).values('name') if (name == dep_name): cnt_err.append(dep['id']) self.assertEqual(len(cnt_err), 0, msg='Errors for recipe dependency id: %s' % cnt_err) # Check if package name does not start with a number (0-9) - tc=846 def test_Package_Name_For_Number(self): packages = Package.objects.filter(~Q(size=-1)).values('id', 'name') cnt_err = [] for package in packages: if (package['name'][0].isdigit() is True): cnt_err.append(package['id']) self.assertEqual( len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) # Check if package version starts with a number (0-9) - tc=847 def test_Package_Version_Starts_With_Number(self): packages = Package.objects.filter( ~Q(size=-1)).values('id', 'version') cnt_err = [] for package in packages: if (package['version'][0].isdigit() is False): cnt_err.append(package['id']) self.assertEqual( len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) # Check if package revision starts with 'r' - tc=848 def test_Package_Revision_Starts_With_r(self): packages = Package.objects.filter( ~Q(size=-1)).values('id', 'revision') cnt_err = [] for package in packages: if (package['revision'][0].startswith("r") is False): cnt_err.append(package['id']) self.assertEqual( len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) # Check the validity of the package build_id # TC must be added in test run def test_Package_Build_Id(self): packages = Package.objects.filter( ~Q(size=-1)).values('id', 'build_id') cnt_err = [] for package in packages: build_id = Build.objects.filter( id=package['build_id']).values('id') if (not build_id): # They have no build_id but if they are # CustomImagePackage that's expected try: CustomImagePackage.objects.get(pk=package['id']) except CustomImagePackage.DoesNotExist: cnt_err.append(package['id']) self.assertEqual(len(cnt_err), 0, msg="Errors for package id: %s they have no build" "associated with them" % cnt_err) # Check the validity of package recipe_id # TC must be added in test run def test_Package_Recipe_Id(self): packages = Package.objects.filter( ~Q(size=-1)).values('id', 'recipe_id') cnt_err = [] for package in packages: recipe_id = Recipe.objects.filter( id=package['recipe_id']).values('id') if (not recipe_id): cnt_err.append(package['id']) self.assertEqual( len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) # Check if package installed_size field is not null # TC must be aded in test run def test_Package_Installed_Size_Not_NULL(self): packages = Package.objects.filter( installed_size__isnull=True).values('id') cnt_err = [] for package in packages: cnt_err.append(package['id']) self.assertEqual( len(cnt_err), 0, msg='Errors for package id: %s' % cnt_err) def test_custom_packages_generated(self): """Test if there is a corresponding generated CustomImagePackage""" """ for each of the packages generated""" missing_packages = [] for package in Package.objects.all(): try: CustomImagePackage.objects.get(name=package.name) except CustomImagePackage.DoesNotExist: missing_packages.append(package.name) self.assertEqual(len(missing_packages), 0, "Some package were created from the build but their" " corresponding CustomImagePackage was not found")