Merge "Add dependency information to the ide_query output." into main
diff --git a/Changes.md b/Changes.md
index fc15e60..9f2449c 100644
--- a/Changes.md
+++ b/Changes.md
@@ -43,14 +43,9 @@
The path set when running builds now makes the `python` executable point to python 3,
whereas on previous versions it pointed to python 2. If you still have python 2 scripts,
you can change the shebang line to use `python2` explicitly. This only applies for
-scripts run directly from makefiles, or from soong genrules. This behavior can be
-temporarily overridden by setting the `BUILD_BROKEN_PYTHON_IS_PYTHON2` environment
-variable to `true`. It's only an environment variable and not a product config variable
-because product config sometimes calls python code.
+scripts run directly from makefiles, or from soong genrules.
-In addition, `python_*` soong modules no longer allow python 2. This can be temporarily
-overridden by setting the `BUILD_BROKEN_USES_SOONG_PYTHON2_MODULES` product configuration
-variable to `true`.
+In addition, `python_*` soong modules no longer allow python 2.
Python 2 is slated for complete removal in V.
diff --git a/ci/Android.bp b/ci/Android.bp
index 104f517..6d4ac35 100644
--- a/ci/Android.bp
+++ b/ci/Android.bp
@@ -71,11 +71,37 @@
},
}
+python_test_host {
+ name: "optimized_targets_test",
+ main: "optimized_targets_test.py",
+ pkg_path: "testdata",
+ srcs: [
+ "optimized_targets_test.py",
+ ],
+ libs: [
+ "build_test_suites",
+ "pyfakefs",
+ ],
+ test_options: {
+ unit_test: true,
+ },
+ data: [
+ ":py3-cmd",
+ ],
+ version: {
+ py3: {
+ embedded_launcher: true,
+ },
+ },
+}
+
python_library_host {
name: "build_test_suites",
srcs: [
"build_test_suites.py",
"optimized_targets.py",
+ "test_mapping_module_retriever.py",
+ "build_context.py",
],
}
diff --git a/ci/build_context.py b/ci/build_context.py
new file mode 100644
index 0000000..cc48d53
--- /dev/null
+++ b/ci/build_context.py
@@ -0,0 +1,64 @@
+# Copyright 2024, The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+"""Container class for build context with utility functions."""
+
+import re
+
+
+class BuildContext:
+
+ def __init__(self, build_context_dict: dict[str, any]):
+ self.enabled_build_features = set()
+ for opt in build_context_dict.get('enabledBuildFeatures', []):
+ self.enabled_build_features.add(opt.get('name'))
+ self.test_infos = set()
+ for test_info_dict in build_context_dict.get('testContext', dict()).get(
+ 'testInfos', []
+ ):
+ self.test_infos.add(self.TestInfo(test_info_dict))
+
+ def build_target_used(self, target: str) -> bool:
+ return any(test.build_target_used(target) for test in self.test_infos)
+
+ class TestInfo:
+
+ _DOWNLOAD_OPTS = {
+ 'test-config-only-zip',
+ 'test-zip-file-filter',
+ 'extra-host-shared-lib-zip',
+ 'sandbox-tests-zips',
+ 'additional-files-filter',
+ 'cts-package-name',
+ }
+
+ def __init__(self, test_info_dict: dict[str, any]):
+ self.is_test_mapping = False
+ self.test_mapping_test_groups = set()
+ self.file_download_options = set()
+ for opt in test_info_dict.get('extraOptions', []):
+ key = opt.get('key')
+ if key == 'test-mapping-test-group':
+ self.is_test_mapping = True
+ self.test_mapping_test_groups.update(opt.get('values', set()))
+
+ if key in self._DOWNLOAD_OPTS:
+ self.file_download_options.update(opt.get('values', set()))
+
+ def build_target_used(self, target: str) -> bool:
+ # For all of a targets' outputs, check if any of the regexes used by tests
+ # to download artifacts would match it. If any of them do then this target
+ # is necessary.
+ regex = r'\b(%s)\b' % re.escape(target)
+ return any(re.search(regex, opt) for opt in self.file_download_options)
diff --git a/ci/build_metadata b/ci/build_metadata
new file mode 100755
index 0000000..a8eb65d
--- /dev/null
+++ b/ci/build_metadata
@@ -0,0 +1,25 @@
+#/bin/bash
+
+# Copyright 2024, The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+set -ex
+
+export TARGET_PRODUCT=aosp_arm64
+export TARGET_RELEASE=trunk_staging
+export TARGET_BUILD_VARIANT=eng
+
+build/soong/bin/m dist \
+ all_teams
+
diff --git a/ci/build_test_suites.py b/ci/build_test_suites.py
index deb1f1d..b8c4a38 100644
--- a/ci/build_test_suites.py
+++ b/ci/build_test_suites.py
@@ -20,10 +20,10 @@
import logging
import os
import pathlib
-import re
import subprocess
import sys
from typing import Callable
+from build_context import BuildContext
import optimized_targets
@@ -53,18 +53,9 @@
any output zip files needed by the build.
"""
- _DOWNLOAD_OPTS = {
- 'test-config-only-zip',
- 'test-zip-file-filter',
- 'extra-host-shared-lib-zip',
- 'sandbox-tests-zips',
- 'additional-files-filter',
- 'cts-package-name',
- }
-
def __init__(
self,
- build_context: dict[str, any],
+ build_context: BuildContext,
args: argparse.Namespace,
target_optimizations: dict[str, optimized_targets.OptimizedBuildTarget],
):
@@ -74,18 +65,15 @@
def create_build_plan(self):
- if 'optimized_build' not in self.build_context.get(
- 'enabledBuildFeatures', []
- ):
+ if 'optimized_build' not in self.build_context.enabled_build_features:
return BuildPlan(set(self.args.extra_targets), set())
build_targets = set()
- packaging_functions = set()
- self.file_download_options = self._aggregate_file_download_options()
+ packaging_commands_getters = []
for target in self.args.extra_targets:
if self._unused_target_exclusion_enabled(
target
- ) and not self._build_target_used(target):
+ ) and not self.build_context.build_target_used(target):
continue
target_optimizer_getter = self.target_optimizations.get(target, None)
@@ -97,44 +85,23 @@
target, self.build_context, self.args
)
build_targets.update(target_optimizer.get_build_targets())
- packaging_functions.add(target_optimizer.package_outputs)
+ packaging_commands_getters.append(
+ target_optimizer.get_package_outputs_commands
+ )
- return BuildPlan(build_targets, packaging_functions)
+ return BuildPlan(build_targets, packaging_commands_getters)
def _unused_target_exclusion_enabled(self, target: str) -> bool:
- return f'{target}_unused_exclusion' in self.build_context.get(
- 'enabledBuildFeatures', []
+ return (
+ f'{target}_unused_exclusion'
+ in self.build_context.enabled_build_features
)
- def _build_target_used(self, target: str) -> bool:
- """Determines whether this target's outputs are used by the test configurations listed in the build context."""
- # For all of a targets' outputs, check if any of the regexes used by tests
- # to download artifacts would match it. If any of them do then this target
- # is necessary.
- regex = r'\b(%s)\b' % re.escape(target)
- return any(re.search(regex, opt) for opt in self.file_download_options)
-
- def _aggregate_file_download_options(self) -> set[str]:
- """Lists out all test config options to specify targets to download.
-
- These come in the form of regexes.
- """
- all_options = set()
- for test_info in self._get_test_infos():
- for opt in test_info.get('extraOptions', []):
- # check the known list of options for downloading files.
- if opt.get('key') in self._DOWNLOAD_OPTS:
- all_options.update(opt.get('values', []))
- return all_options
-
- def _get_test_infos(self):
- return self.build_context.get('testContext', dict()).get('testInfos', [])
-
@dataclass(frozen=True)
class BuildPlan:
build_targets: set[str]
- packaging_functions: set[Callable[..., None]]
+ packaging_commands_getters: list[Callable[[], list[list[str]]]]
def build_test_suites(argv: list[str]) -> int:
@@ -148,7 +115,7 @@
"""
args = parse_args(argv)
check_required_env()
- build_context = load_build_context()
+ build_context = BuildContext(load_build_context())
build_planner = BuildPlanner(
build_context, args, optimized_targets.OPTIMIZED_BUILD_TARGETS
)
@@ -216,8 +183,12 @@
except subprocess.CalledProcessError as e:
raise BuildFailureError(e.returncode) from e
- for packaging_function in build_plan.packaging_functions:
- packaging_function()
+ for packaging_commands_getter in build_plan.packaging_commands_getters:
+ try:
+ for packaging_command in packaging_commands_getter():
+ run_command(packaging_command)
+ except subprocess.CalledProcessError as e:
+ raise BuildFailureError(e.returncode) from e
def get_top() -> pathlib.Path:
diff --git a/ci/build_test_suites_test.py b/ci/build_test_suites_test.py
index 463bdd0..2afaab7 100644
--- a/ci/build_test_suites_test.py
+++ b/ci/build_test_suites_test.py
@@ -32,6 +32,7 @@
from typing import Callable
import unittest
from unittest import mock
+from build_context import BuildContext
import build_test_suites
import ci_test_lib
import optimized_targets
@@ -240,17 +241,17 @@
class TestOptimizedBuildTarget(optimized_targets.OptimizedBuildTarget):
def __init__(
- self, target, build_context, args, output_targets, packaging_outputs
+ self, target, build_context, args, output_targets, packaging_commands
):
super().__init__(target, build_context, args)
self.output_targets = output_targets
- self.packaging_outputs = packaging_outputs
+ self.packaging_commands = packaging_commands
def get_build_targets_impl(self):
return self.output_targets
- def package_outputs_impl(self):
- self.packaging_outputs.add(f'packaging {" ".join(self.output_targets)}')
+ def get_package_outputs_commands_impl(self):
+ return self.packaging_commands
def get_enabled_flag(self):
return f'{self.target}_enabled'
@@ -275,14 +276,15 @@
build_plan = build_planner.create_build_plan()
- self.assertEqual(len(build_plan.packaging_functions), 0)
+ for packaging_command in self.run_packaging_commands(build_plan):
+ self.assertEqual(len(packaging_command), 0)
def test_build_optimization_on_optimizes_target(self):
build_targets = {'target_1', 'target_2'}
build_planner = self.create_build_planner(
build_targets=build_targets,
build_context=self.create_build_context(
- enabled_build_features={self.get_target_flag('target_1')}
+ enabled_build_features=[{'name': self.get_target_flag('target_1')}]
),
)
@@ -293,20 +295,19 @@
def test_build_optimization_on_packages_target(self):
build_targets = {'target_1', 'target_2'}
- packaging_outputs = set()
+ optimized_target_name = self.get_optimized_target_name('target_1')
+ packaging_commands = [[f'packaging {optimized_target_name}']]
build_planner = self.create_build_planner(
build_targets=build_targets,
build_context=self.create_build_context(
- enabled_build_features={self.get_target_flag('target_1')},
+ enabled_build_features=[{'name': self.get_target_flag('target_1')}]
),
- packaging_outputs=packaging_outputs,
+ packaging_commands=packaging_commands,
)
build_plan = build_planner.create_build_plan()
- self.run_packaging_functions(build_plan)
- optimized_target_name = self.get_optimized_target_name('target_1')
- self.assertIn(f'packaging {optimized_target_name}', packaging_outputs)
+ self.assertIn(packaging_commands, self.run_packaging_commands(build_plan))
def test_individual_build_optimization_off_doesnt_optimize(self):
build_targets = {'target_1', 'target_2'}
@@ -320,16 +321,16 @@
def test_individual_build_optimization_off_doesnt_package(self):
build_targets = {'target_1', 'target_2'}
- packaging_outputs = set()
+ packaging_commands = [['packaging command']]
build_planner = self.create_build_planner(
build_targets=build_targets,
- packaging_outputs=packaging_outputs,
+ packaging_commands=packaging_commands,
)
build_plan = build_planner.create_build_plan()
- self.run_packaging_functions(build_plan)
- self.assertFalse(packaging_outputs)
+ for packaging_command in self.run_packaging_commands(build_plan):
+ self.assertEqual(len(packaging_command), 0)
def test_target_output_used_target_built(self):
build_target = 'test_target'
@@ -337,7 +338,7 @@
build_targets={build_target},
build_context=self.create_build_context(
test_context=self.get_test_context(build_target),
- enabled_build_features={'test_target_unused_exclusion'},
+ enabled_build_features=[{'name': 'test_target_unused_exclusion'}],
),
)
@@ -356,7 +357,7 @@
build_targets={build_target},
build_context=self.create_build_context(
test_context=test_context,
- enabled_build_features={'test_target_unused_exclusion'},
+ enabled_build_features=[{'name': 'test_target_unused_exclusion'}],
),
)
@@ -372,7 +373,7 @@
build_targets={build_target},
build_context=self.create_build_context(
test_context=test_context,
- enabled_build_features={'test_target_unused_exclusion'},
+ enabled_build_features=[{'name': 'test_target_unused_exclusion'}],
),
)
@@ -391,7 +392,7 @@
build_targets={build_target},
build_context=self.create_build_context(
test_context=test_context,
- enabled_build_features={'test_target_unused_exclusion'},
+ enabled_build_features=[{'name': 'test_target_unused_exclusion'}],
),
)
@@ -402,12 +403,12 @@
def create_build_planner(
self,
build_targets: set[str],
- build_context: dict[str, any] = None,
+ build_context: BuildContext = None,
args: argparse.Namespace = None,
target_optimizations: dict[
str, optimized_targets.OptimizedBuildTarget
] = None,
- packaging_outputs: set[str] = set(),
+ packaging_commands: list[list[str]] = [],
) -> build_test_suites.BuildPlanner:
if not build_context:
build_context = self.create_build_context()
@@ -417,7 +418,7 @@
target_optimizations = self.create_target_optimizations(
build_context,
build_targets,
- packaging_outputs,
+ packaging_commands,
)
return build_test_suites.BuildPlanner(
build_context, args, target_optimizations
@@ -426,15 +427,17 @@
def create_build_context(
self,
optimized_build_enabled: bool = True,
- enabled_build_features: set[str] = set(),
+ enabled_build_features: list[dict[str, str]] = [],
test_context: dict[str, any] = {},
- ) -> dict[str, any]:
- build_context = {}
- build_context['enabledBuildFeatures'] = enabled_build_features
+ ) -> BuildContext:
+ build_context_dict = {}
+ build_context_dict['enabledBuildFeatures'] = enabled_build_features
if optimized_build_enabled:
- build_context['enabledBuildFeatures'].add('optimized_build')
- build_context['testContext'] = test_context
- return build_context
+ build_context_dict['enabledBuildFeatures'].append(
+ {'name': 'optimized_build'}
+ )
+ build_context_dict['testContext'] = test_context
+ return BuildContext(build_context_dict)
def create_args(
self, extra_build_targets: set[str] = set()
@@ -445,16 +448,16 @@
def create_target_optimizations(
self,
- build_context: dict[str, any],
+ build_context: BuildContext,
build_targets: set[str],
- packaging_outputs: set[str] = set(),
+ packaging_commands: list[list[str]] = [],
):
target_optimizations = dict()
for target in build_targets:
target_optimizations[target] = functools.partial(
self.TestOptimizedBuildTarget,
output_targets={self.get_optimized_target_name(target)},
- packaging_outputs=packaging_outputs,
+ packaging_commands=packaging_commands,
)
return target_optimizations
@@ -465,10 +468,6 @@
def get_optimized_target_name(self, target: str):
return f'{target}_optimized'
- def run_packaging_functions(self, build_plan: build_test_suites.BuildPlan):
- for packaging_function in build_plan.packaging_functions:
- packaging_function()
-
def get_test_context(self, target: str):
return {
'testInfos': [
@@ -488,6 +487,12 @@
],
}
+ def run_packaging_commands(self, build_plan: build_test_suites.BuildPlan):
+ return [
+ packaging_command_getter()
+ for packaging_command_getter in build_plan.packaging_commands_getters
+ ]
+
def wait_until(
condition_function: Callable[[], bool],
diff --git a/ci/buildbot.py b/ci/buildbot.py
new file mode 100644
index 0000000..97097be
--- /dev/null
+++ b/ci/buildbot.py
@@ -0,0 +1,43 @@
+# Copyright 2024, The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+"""Utilities for interacting with buildbot, with a simulation in a local environment"""
+
+import os
+import sys
+
+# Check that the script is running from the root of the tree. Prevents subtle
+# errors later, and CI always runs from the root of the tree.
+if not os.path.exists("build/make/ci/buildbot.py"):
+ raise Exception("CI script must be run from the root of the tree instead of: "
+ + os.getcwd())
+
+# Check that we are using the hermetic interpreter
+if "prebuilts/build-tools/" not in sys.executable:
+ raise Exception("CI script must be run using the hermetic interpreter from "
+ + "prebuilts/build-tools instead of: " + sys.executable)
+
+
+def OutDir():
+ "Get the out directory. Will create it if needed."
+ result = os.environ.get("OUT_DIR", "out")
+ os.makedirs(result, exist_ok=True)
+ return result
+
+def DistDir():
+ "Get the dist directory. Will create it if needed."
+ result = os.environ.get("DIST_DIR", os.path.join(OutDir(), "dist"))
+ os.makedirs(result, exist_ok=True)
+ return result
+
diff --git a/ci/dump_product_config b/ci/dump_product_config
new file mode 100755
index 0000000..77b51dd
--- /dev/null
+++ b/ci/dump_product_config
@@ -0,0 +1,353 @@
+#!prebuilts/build-tools/linux-x86/bin/py3-cmd -B
+
+# Copyright 2024, The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+"""Script to collect all of the make variables from all product config combos.
+
+This script must be run from the root of the source tree.
+
+See GetArgs() below or run dump_product_config for more information.
+"""
+
+import argparse
+import asyncio
+import contextlib
+import csv
+import dataclasses
+import json
+import multiprocessing
+import os
+import subprocess
+import sys
+import time
+from typing import List, Dict, Tuple, Optional
+
+import buildbot
+
+# We have some BIG variables
+csv.field_size_limit(sys.maxsize)
+
+
+class DataclassJSONEncoder(json.JSONEncoder):
+ """JSONEncoder for our custom types."""
+ def default(self, o):
+ if dataclasses.is_dataclass(o):
+ return dataclasses.asdict(o)
+ return super().default(o)
+
+
+def GetProducts():
+ """Get the all of the available TARGET_PRODUCT values."""
+ try:
+ stdout = subprocess.check_output(["build/soong/bin/list_products"], text=True)
+ except subprocess.CalledProcessError:
+ sys.exit(1)
+ return [s.strip() for s in stdout.splitlines() if s.strip()]
+
+
+def GetReleases(product):
+ """For a given product, get the release configs available to it."""
+ if True:
+ # Hard code the list
+ mainline_products = [
+ "module_arm",
+ "module_x86",
+ "module_arm64",
+ "module_riscv64",
+ "module_x86_64",
+ "module_arm64only",
+ "module_x86_64only",
+ ]
+ if product in mainline_products:
+ return ["trunk_staging", "trunk", "mainline"]
+ else:
+ return ["trunk_staging", "trunk", "next"]
+ else:
+ # Get it from the build system
+ try:
+ stdout = subprocess.check_output(["build/soong/bin/list_releases", product], text=True)
+ except subprocess.CalledProcessError:
+ sys.exit(1)
+ return [s.strip() for s in stdout.splitlines() if s.strip()]
+
+
+def GenerateAllLunchTargets():
+ """Generate the full list of lunch targets."""
+ for product in GetProducts():
+ for release in GetReleases(product):
+ for variant in ["user", "userdebug", "eng"]:
+ yield (product, release, variant)
+
+
+async def ParallelExec(parallelism, tasks):
+ '''
+ ParallelExec takes a parallelism number, and an iterator of tasks to run.
+ Then it will run all the tasks, but a maximum of parallelism will be run at
+ any given time. The tasks must be async functions that accept one argument,
+ which will be an integer id of the worker that they're running on.
+ '''
+ tasks = iter(tasks)
+
+ overall_start = time.monotonic()
+ # lists so they can be modified from the inner function
+ total_duration = [0]
+ count = [0]
+ async def dispatch(worker):
+ while True:
+ try:
+ task = next(tasks)
+ item_start = time.monotonic()
+ await task(worker)
+ now = time.monotonic()
+ item_duration = now - item_start
+ count[0] += 1
+ total_duration[0] += item_duration
+ sys.stderr.write(f"Timing: Items processed: {count[0]}, Wall time: {now-overall_start:0.1f} sec, Throughput: {(now-overall_start)/count[0]:0.3f} sec per item, Average duration: {total_duration[0]/count[0]:0.1f} sec\n")
+ except StopIteration:
+ return
+
+ await asyncio.gather(*[dispatch(worker) for worker in range(parallelism)])
+
+
+async def DumpProductConfigs(out, generator, out_dir):
+ """Collects all of the product config data and store it in file."""
+ # Write the outer json list by hand so we can stream it
+ out.write("[")
+ try:
+ first_result = [True] # a list so it can be modified from the inner function
+ def run(lunch):
+ async def curried(worker):
+ sys.stderr.write(f"running: {'-'.join(lunch)}\n")
+ result = await DumpOneProductConfig(lunch, os.path.join(out_dir, f"lunchable_{worker}"))
+ if first_result[0]:
+ out.write("\n")
+ first_result[0] = False
+ else:
+ out.write(",\n")
+ result.dumpToFile(out)
+ sys.stderr.write(f"finished: {'-'.join(lunch)}\n")
+ return curried
+
+ await ParallelExec(multiprocessing.cpu_count(), (run(lunch) for lunch in generator))
+ finally:
+ # Close the json regardless of how we exit
+ out.write("\n]\n")
+
+
+@dataclasses.dataclass(frozen=True)
+class Variable:
+ """A variable name, value and where it was set."""
+ name: str
+ value: str
+ location: str
+
+
+@dataclasses.dataclass(frozen=True)
+class ProductResult:
+ product: str
+ release: str
+ variant: str
+ board_includes: List[str]
+ product_includes: Dict[str, List[str]]
+ product_graph: List[Tuple[str, str]]
+ board_vars: List[Variable]
+ product_vars: List[Variable]
+
+ def dumpToFile(self, f):
+ json.dump(self, f, sort_keys=True, indent=2, cls=DataclassJSONEncoder)
+
+
+@dataclasses.dataclass(frozen=True)
+class ProductError:
+ product: str
+ release: str
+ variant: str
+ error: str
+
+ def dumpToFile(self, f):
+ json.dump(self, f, sort_keys=True, indent=2, cls=DataclassJSONEncoder)
+
+
+def NormalizeInheritGraph(lists):
+ """Flatten the inheritance graph to a simple list for easier querying."""
+ result = set()
+ for item in lists:
+ for i in range(len(item)):
+ result.add((item[i+1] if i < len(item)-1 else "", item[i]))
+ return sorted(list(result))
+
+
+def ParseDump(lunch, filename) -> ProductResult:
+ """Parses the csv and returns a tuple of the data."""
+ def diff(initial, final):
+ return [after for after in final.values() if
+ initial.get(after.name, Variable(after.name, "", "<unset>")).value != after.value]
+ product_initial = {}
+ product_final = {}
+ board_initial = {}
+ board_final = {}
+ inherit_product = [] # The stack of inherit-product calls
+ product_includes = {} # Other files included by each of the properly imported files
+ board_includes = [] # Files included by boardconfig
+ with open(filename) as f:
+ phase = ""
+ for line in csv.reader(f):
+ if line[0] == "phase":
+ phase = line[1]
+ elif line[0] == "val":
+ # TOOD: We should skip these somewhere else.
+ if line[3].startswith("_ALL_RELEASE_FLAGS"):
+ continue
+ if line[3].startswith("PRODUCTS."):
+ continue
+ if phase == "PRODUCTS":
+ if line[2] == "initial":
+ product_initial[line[3]] = Variable(line[3], line[4], line[5])
+ if phase == "PRODUCT-EXPAND":
+ if line[2] == "final":
+ product_final[line[3]] = Variable(line[3], line[4], line[5])
+ if phase == "BOARD":
+ if line[2] == "initial":
+ board_initial[line[3]] = Variable(line[3], line[4], line[5])
+ if line[2] == "final":
+ board_final[line[3]] = Variable(line[3], line[4], line[5])
+ elif line[0] == "imported":
+ imports = [s.strip() for s in line[1].split()]
+ if imports:
+ inherit_product.append(imports)
+ inc = [s.strip() for s in line[2].split()]
+ for f in inc:
+ product_includes.setdefault(imports[0], []).append(f)
+ elif line[0] == "board_config_files":
+ board_includes += [s.strip() for s in line[1].split()]
+ return ProductResult(
+ product = lunch[0],
+ release = lunch[1],
+ variant = lunch[2],
+ product_vars = diff(product_initial, product_final),
+ board_vars = diff(board_initial, board_final),
+ product_graph = NormalizeInheritGraph(inherit_product),
+ product_includes = product_includes,
+ board_includes = board_includes
+ )
+
+
+async def DumpOneProductConfig(lunch, out_dir) -> ProductResult | ProductError:
+ """Print a single config's lunch info to stdout."""
+ product, release, variant = lunch
+
+ dumpconfig_file = os.path.join(out_dir, f"{product}-{release}-{variant}.csv")
+
+ # Run get_build_var to bootstrap soong_ui for this target
+ env = dict(os.environ)
+ env["TARGET_PRODUCT"] = product
+ env["TARGET_RELEASE"] = release
+ env["TARGET_BUILD_VARIANT"] = variant
+ env["OUT_DIR"] = out_dir
+ process = await asyncio.create_subprocess_exec(
+ "build/soong/bin/get_build_var",
+ "TARGET_PRODUCT",
+ stdout=subprocess.PIPE,
+ stderr=subprocess.STDOUT,
+ env=env
+ )
+ stdout, _ = await process.communicate()
+ stdout = stdout.decode()
+
+ if process.returncode != 0:
+ return ProductError(
+ product = product,
+ release = release,
+ variant = variant,
+ error = stdout
+ )
+ else:
+ # Run kati to extract the data
+ process = await asyncio.create_subprocess_exec(
+ "prebuilts/build-tools/linux-x86/bin/ckati",
+ "-f",
+ "build/make/core/dumpconfig.mk",
+ f"TARGET_PRODUCT={product}",
+ f"TARGET_RELEASE={release}",
+ f"TARGET_BUILD_VARIANT={variant}",
+ f"DUMPCONFIG_FILE={dumpconfig_file}",
+ stdout=subprocess.PIPE,
+ stderr=subprocess.STDOUT,
+ env=env
+ )
+ stdout, _ = await process.communicate()
+ if process.returncode != 0:
+ stdout = stdout.decode()
+ return ProductError(
+ product = product,
+ release = release,
+ variant = variant,
+ error = stdout
+ )
+ else:
+ # Parse and record the output
+ return ParseDump(lunch, dumpconfig_file)
+
+
+def GetArgs():
+ """Parse command line arguments."""
+ parser = argparse.ArgumentParser(
+ description="Collect all of the make variables from product config.",
+ epilog="NOTE: This script must be run from the root of the source tree.")
+ parser.add_argument("--lunch", nargs="*")
+ parser.add_argument("--dist", action="store_true")
+
+ return parser.parse_args()
+
+
+async def main():
+ args = GetArgs()
+
+ out_dir = buildbot.OutDir()
+
+ if args.dist:
+ cm = open(os.path.join(buildbot.DistDir(), "all_product_config.json"), "w")
+ else:
+ cm = contextlib.nullcontext(sys.stdout)
+
+
+ with cm as out:
+ if args.lunch:
+ lunches = [lunch.split("-") for lunch in args.lunch]
+ fail = False
+ for i in range(len(lunches)):
+ if len(lunches[i]) != 3:
+ sys.stderr.write(f"Malformed lunch targets: {args.lunch[i]}\n")
+ fail = True
+ if fail:
+ sys.exit(1)
+ if len(lunches) == 1:
+ result = await DumpOneProductConfig(lunches[0], out_dir)
+ result.dumpToFile(out)
+ out.write("\n")
+ else:
+ await DumpProductConfigs(out, lunches, out_dir)
+ else:
+ # All configs mode. This will exec single config mode in parallel
+ # for each lunch combo. Write output to $DIST_DIR.
+ await DumpProductConfigs(out, GenerateAllLunchTargets(), out_dir)
+
+
+if __name__ == "__main__":
+ asyncio.run(main())
+
+
+# vim: set syntax=python ts=4 sw=4 sts=4:
+
diff --git a/ci/optimized_targets.py b/ci/optimized_targets.py
index 8a529c7..688bdd8 100644
--- a/ci/optimized_targets.py
+++ b/ci/optimized_targets.py
@@ -14,9 +14,16 @@
# limitations under the License.
from abc import ABC
-from typing import Self
import argparse
import functools
+import json
+import logging
+import os
+import pathlib
+import subprocess
+
+from build_context import BuildContext
+import test_mapping_module_retriever
class OptimizedBuildTarget(ABC):
@@ -27,10 +34,13 @@
build.
"""
+ _SOONG_UI_BASH_PATH = 'build/soong/soong_ui.bash'
+ _PREBUILT_SOONG_ZIP_PATH = 'prebuilts/build-tools/linux-x86/bin/soong_zip'
+
def __init__(
self,
target: str,
- build_context: dict[str, any],
+ build_context: BuildContext,
args: argparse.Namespace,
):
self.target = target
@@ -38,19 +48,25 @@
self.args = args
def get_build_targets(self) -> set[str]:
- features = self.build_context.get('enabledBuildFeatures', [])
+ features = self.build_context.enabled_build_features
if self.get_enabled_flag() in features:
- return self.get_build_targets_impl()
+ self.modules_to_build = self.get_build_targets_impl()
+ return self.modules_to_build
+
+ self.modules_to_build = {self.target}
return {self.target}
- def package_outputs(self):
- features = self.build_context.get('enabledBuildFeatures', [])
+ def get_package_outputs_commands(self) -> list[list[str]]:
+ features = self.build_context.enabled_build_features
if self.get_enabled_flag() in features:
- return self.package_outputs_impl()
+ return self.get_package_outputs_commands_impl()
- def package_outputs_impl(self):
+ return []
+
+ def get_package_outputs_commands_impl(self) -> list[list[str]]:
raise NotImplementedError(
- f'package_outputs_impl not implemented in {type(self).__name__}'
+ 'get_package_outputs_commands_impl not implemented in'
+ f' {type(self).__name__}'
)
def get_enabled_flag(self):
@@ -63,6 +79,88 @@
f'get_build_targets_impl not implemented in {type(self).__name__}'
)
+ def _generate_zip_options_for_items(
+ self,
+ prefix: str = '',
+ relative_root: str = '',
+ list_files: list[str] | None = None,
+ files: list[str] | None = None,
+ directories: list[str] | None = None,
+ ) -> list[str]:
+ if not list_files and not files and not directories:
+ raise RuntimeError(
+ f'No items specified to be added to zip! Prefix: {prefix}, Relative'
+ f' root: {relative_root}'
+ )
+ command_segment = []
+ # These are all soong_zip options so consult soong_zip --help for specifics.
+ if prefix:
+ command_segment.append('-P')
+ command_segment.append(prefix)
+ if relative_root:
+ command_segment.append('-C')
+ command_segment.append(relative_root)
+ if list_files:
+ for list_file in list_files:
+ command_segment.append('-l')
+ command_segment.append(list_file)
+ if files:
+ for file in files:
+ command_segment.append('-f')
+ command_segment.append(file)
+ if directories:
+ for directory in directories:
+ command_segment.append('-D')
+ command_segment.append(directory)
+
+ return command_segment
+
+ def _query_soong_vars(
+ self, src_top: pathlib.Path, soong_vars: list[str]
+ ) -> dict[str, str]:
+ process_result = subprocess.run(
+ args=[
+ f'{src_top / self._SOONG_UI_BASH_PATH}',
+ '--dumpvars-mode',
+ f'--abs-vars={" ".join(soong_vars)}',
+ ],
+ env=os.environ,
+ check=False,
+ capture_output=True,
+ text=True,
+ )
+ if not process_result.returncode == 0:
+ logging.error('soong dumpvars command failed! stderr:')
+ logging.error(process_result.stderr)
+ raise RuntimeError('Soong dumpvars failed! See log for stderr.')
+
+ if not process_result.stdout:
+ raise RuntimeError(
+ 'Necessary soong variables ' + soong_vars + ' not found.'
+ )
+
+ try:
+ return {
+ line.split('=')[0]: line.split('=')[1].strip("'")
+ for line in process_result.stdout.strip().split('\n')
+ }
+ except IndexError as e:
+ raise RuntimeError(
+ 'Error parsing soong dumpvars output! See output here:'
+ f' {process_result.stdout}',
+ e,
+ )
+
+ def _base_zip_command(
+ self, src_top: pathlib.Path, dist_dir: pathlib.Path, name: str
+ ) -> list[str]:
+ return [
+ f'{src_top / self._PREBUILT_SOONG_ZIP_PATH }',
+ '-d',
+ '-o',
+ f'{dist_dir / name}',
+ ]
+
class NullOptimizer(OptimizedBuildTarget):
"""No-op target optimizer.
@@ -77,15 +175,38 @@
def get_build_targets(self):
return {self.target}
- def package_outputs(self):
- pass
+ def get_package_outputs_commands(self):
+ return []
+
+
+class ChangeInfo:
+
+ def __init__(self, change_info_file_path):
+ try:
+ with open(change_info_file_path) as change_info_file:
+ change_info_contents = json.load(change_info_file)
+ except json.decoder.JSONDecodeError:
+ logging.error(f'Failed to load CHANGE_INFO: {change_info_file_path}')
+ raise
+
+ self._change_info_contents = change_info_contents
+
+ def find_changed_files(self) -> set[str]:
+ changed_files = set()
+
+ for change in self._change_info_contents['changes']:
+ project_path = change.get('projectPath') + '/'
+
+ for revision in change.get('revisions'):
+ for file_info in revision.get('fileInfos'):
+ changed_files.add(project_path + file_info.get('path'))
+
+ return changed_files
class GeneralTestsOptimizer(OptimizedBuildTarget):
"""general-tests optimizer
- TODO(b/358215235): Implement
-
This optimizer reads in the list of changed files from the file located in
env[CHANGE_INFO] and uses this list alongside the normal TEST MAPPING logic to
determine what test mapping modules will run for the given changes. It then
@@ -93,8 +214,264 @@
normally built.
"""
+ # List of modules that are built alongside general-tests as dependencies.
+ _REQUIRED_MODULES = frozenset([
+ 'cts-tradefed',
+ 'vts-tradefed',
+ 'compatibility-host-util',
+ 'general-tests-shared-libs',
+ ])
+
+ def get_build_targets_impl(self) -> set[str]:
+ change_info_file_path = os.environ.get('CHANGE_INFO')
+ if not change_info_file_path:
+ logging.info(
+ 'No CHANGE_INFO env var found, general-tests optimization disabled.'
+ )
+ return {'general-tests'}
+
+ test_infos = self.build_context.test_infos
+ test_mapping_test_groups = set()
+ for test_info in test_infos:
+ is_test_mapping = test_info.is_test_mapping
+ current_test_mapping_test_groups = test_info.test_mapping_test_groups
+ uses_general_tests = test_info.build_target_used('general-tests')
+
+ if uses_general_tests and not is_test_mapping:
+ logging.info(
+ 'Test uses general-tests.zip but is not test-mapping, general-tests'
+ ' optimization disabled.'
+ )
+ return {'general-tests'}
+
+ if is_test_mapping:
+ test_mapping_test_groups.update(current_test_mapping_test_groups)
+
+ change_info = ChangeInfo(change_info_file_path)
+ changed_files = change_info.find_changed_files()
+
+ test_mappings = test_mapping_module_retriever.GetTestMappings(
+ changed_files, set()
+ )
+
+ modules_to_build = set(self._REQUIRED_MODULES)
+
+ modules_to_build.update(
+ test_mapping_module_retriever.FindAffectedModules(
+ test_mappings, changed_files, test_mapping_test_groups
+ )
+ )
+
+ return modules_to_build
+
+ def get_package_outputs_commands_impl(self):
+ src_top = pathlib.Path(os.environ.get('TOP', os.getcwd()))
+ dist_dir = pathlib.Path(os.environ.get('DIST_DIR'))
+
+ soong_vars = self._query_soong_vars(
+ src_top,
+ [
+ 'HOST_OUT_TESTCASES',
+ 'TARGET_OUT_TESTCASES',
+ 'PRODUCT_OUT',
+ 'SOONG_HOST_OUT',
+ 'HOST_OUT',
+ ],
+ )
+ host_out_testcases = pathlib.Path(soong_vars.get('HOST_OUT_TESTCASES'))
+ target_out_testcases = pathlib.Path(soong_vars.get('TARGET_OUT_TESTCASES'))
+ product_out = pathlib.Path(soong_vars.get('PRODUCT_OUT'))
+ soong_host_out = pathlib.Path(soong_vars.get('SOONG_HOST_OUT'))
+ host_out = pathlib.Path(soong_vars.get('HOST_OUT'))
+
+ host_paths = []
+ target_paths = []
+ host_config_files = []
+ target_config_files = []
+ for module in self.modules_to_build:
+ # The required modules are handled separately, no need to package.
+ if module in self._REQUIRED_MODULES:
+ continue
+
+ host_path = host_out_testcases / module
+ if os.path.exists(host_path):
+ host_paths.append(host_path)
+ self._collect_config_files(src_top, host_path, host_config_files)
+
+ target_path = target_out_testcases / module
+ if os.path.exists(target_path):
+ target_paths.append(target_path)
+ self._collect_config_files(src_top, target_path, target_config_files)
+
+ if not os.path.exists(host_path) and not os.path.exists(target_path):
+ logging.info(f'No host or target build outputs found for {module}.')
+
+ zip_commands = []
+
+ zip_commands.extend(
+ self._get_zip_test_configs_zips_commands(
+ src_top,
+ dist_dir,
+ host_out,
+ product_out,
+ host_config_files,
+ target_config_files,
+ )
+ )
+
+ zip_command = self._base_zip_command(src_top, dist_dir, 'general-tests.zip')
+
+ # Add host testcases.
+ if host_paths:
+ zip_command.extend(
+ self._generate_zip_options_for_items(
+ prefix='host',
+ relative_root=f'{src_top / soong_host_out}',
+ directories=host_paths,
+ )
+ )
+
+ # Add target testcases.
+ if target_paths:
+ zip_command.extend(
+ self._generate_zip_options_for_items(
+ prefix='target',
+ relative_root=f'{src_top / product_out}',
+ directories=target_paths,
+ )
+ )
+
+ # TODO(lucafarsi): Push this logic into a general-tests-minimal build command
+ # Add necessary tools. These are also hardcoded in general-tests.mk.
+ framework_path = soong_host_out / 'framework'
+
+ zip_command.extend(
+ self._generate_zip_options_for_items(
+ prefix='host/tools',
+ relative_root=str(framework_path),
+ files=[
+ f"{framework_path / 'cts-tradefed.jar'}",
+ f"{framework_path / 'compatibility-host-util.jar'}",
+ f"{framework_path / 'vts-tradefed.jar'}",
+ ],
+ )
+ )
+
+ zip_commands.append(zip_command)
+ return zip_commands
+
+ def _collect_config_files(
+ self,
+ src_top: pathlib.Path,
+ root_dir: pathlib.Path,
+ config_files: list[str],
+ ):
+ for root, dirs, files in os.walk(src_top / root_dir):
+ for file in files:
+ if file.endswith('.config'):
+ config_files.append(root_dir / file)
+
+ def _get_zip_test_configs_zips_commands(
+ self,
+ src_top: pathlib.Path,
+ dist_dir: pathlib.Path,
+ host_out: pathlib.Path,
+ product_out: pathlib.Path,
+ host_config_files: list[str],
+ target_config_files: list[str],
+ ) -> tuple[list[str], list[str]]:
+ """Generate general-tests_configs.zip and general-tests_list.zip.
+
+ general-tests_configs.zip contains all of the .config files that were
+ built and general-tests_list.zip contains a text file which lists
+ all of the .config files that are in general-tests_configs.zip.
+
+ general-tests_configs.zip is organized as follows:
+ /
+ host/
+ testcases/
+ test_1.config
+ test_2.config
+ ...
+ target/
+ testcases/
+ test_1.config
+ test_2.config
+ ...
+
+ So the process is we write out the paths to all the host config files into
+ one
+ file and all the paths to the target config files in another. We also write
+ the paths to all the config files into a third file to use for
+ general-tests_list.zip.
+
+ Args:
+ dist_dir: dist directory.
+ host_out: host out directory.
+ product_out: product out directory.
+ host_config_files: list of all host config files.
+ target_config_files: list of all target config files.
+
+ Returns:
+ The commands to generate general-tests_configs.zip and
+ general-tests_list.zip
+ """
+ with open(
+ f"{host_out / 'host_general-tests_list'}", 'w'
+ ) as host_list_file, open(
+ f"{product_out / 'target_general-tests_list'}", 'w'
+ ) as target_list_file, open(
+ f"{host_out / 'general-tests_list'}", 'w'
+ ) as list_file:
+
+ for config_file in host_config_files:
+ host_list_file.write(f'{config_file}' + '\n')
+ list_file.write('host/' + os.path.relpath(config_file, host_out) + '\n')
+
+ for config_file in target_config_files:
+ target_list_file.write(f'{config_file}' + '\n')
+ list_file.write(
+ 'target/' + os.path.relpath(config_file, product_out) + '\n'
+ )
+
+ zip_commands = []
+
+ tests_config_zip_command = self._base_zip_command(
+ src_top, dist_dir, 'general-tests_configs.zip'
+ )
+ tests_config_zip_command.extend(
+ self._generate_zip_options_for_items(
+ prefix='host',
+ relative_root=str(host_out),
+ list_files=[f"{host_out / 'host_general-tests_list'}"],
+ )
+ )
+
+ tests_config_zip_command.extend(
+ self._generate_zip_options_for_items(
+ prefix='target',
+ relative_root=str(product_out),
+ list_files=[f"{product_out / 'target_general-tests_list'}"],
+ ),
+ )
+
+ zip_commands.append(tests_config_zip_command)
+
+ tests_list_zip_command = self._base_zip_command(
+ src_top, dist_dir, 'general-tests_list.zip'
+ )
+ tests_list_zip_command.extend(
+ self._generate_zip_options_for_items(
+ relative_root=str(host_out),
+ files=[f"{host_out / 'general-tests_list'}"],
+ )
+ )
+ zip_commands.append(tests_list_zip_command)
+
+ return zip_commands
+
def get_enabled_flag(self):
- return 'general-tests-optimized'
+ return 'general_tests_optimized'
@classmethod
def get_optimized_targets(cls) -> dict[str, OptimizedBuildTarget]:
diff --git a/ci/optimized_targets_test.py b/ci/optimized_targets_test.py
new file mode 100644
index 0000000..0b0c0ec
--- /dev/null
+++ b/ci/optimized_targets_test.py
@@ -0,0 +1,350 @@
+# Copyright 2024, The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+"""Tests for optimized_targets.py"""
+
+import json
+import logging
+import os
+import pathlib
+import re
+import subprocess
+import textwrap
+import unittest
+from unittest import mock
+from build_context import BuildContext
+import optimized_targets
+from pyfakefs import fake_filesystem_unittest
+
+
+class GeneralTestsOptimizerTest(fake_filesystem_unittest.TestCase):
+
+ def setUp(self):
+ self.setUpPyfakefs()
+
+ os_environ_patcher = mock.patch.dict('os.environ', {})
+ self.addCleanup(os_environ_patcher.stop)
+ self.mock_os_environ = os_environ_patcher.start()
+
+ self._setup_working_build_env()
+ self._write_change_info_file()
+ test_mapping_dir = pathlib.Path('/project/path/file/path')
+ test_mapping_dir.mkdir(parents=True)
+ self._write_test_mapping_file()
+
+ def _setup_working_build_env(self):
+ self.change_info_file = pathlib.Path('/tmp/change_info')
+ self._write_soong_ui_file()
+ self._host_out_testcases = pathlib.Path('/tmp/top/host_out_testcases')
+ self._host_out_testcases.mkdir(parents=True)
+ self._target_out_testcases = pathlib.Path('/tmp/top/target_out_testcases')
+ self._target_out_testcases.mkdir(parents=True)
+ self._product_out = pathlib.Path('/tmp/top/product_out')
+ self._product_out.mkdir(parents=True)
+ self._soong_host_out = pathlib.Path('/tmp/top/soong_host_out')
+ self._soong_host_out.mkdir(parents=True)
+ self._host_out = pathlib.Path('/tmp/top/host_out')
+ self._host_out.mkdir(parents=True)
+
+ self._dist_dir = pathlib.Path('/tmp/top/out/dist')
+ self._dist_dir.mkdir(parents=True)
+
+ self.mock_os_environ.update({
+ 'CHANGE_INFO': str(self.change_info_file),
+ 'TOP': '/tmp/top',
+ 'DIST_DIR': '/tmp/top/out/dist',
+ })
+
+ def _write_soong_ui_file(self):
+ soong_path = pathlib.Path('/tmp/top/build/soong')
+ soong_path.mkdir(parents=True)
+ with open(os.path.join(soong_path, 'soong_ui.bash'), 'w') as f:
+ f.write("""
+ #/bin/bash
+ echo HOST_OUT_TESTCASES='/tmp/top/host_out_testcases'
+ echo TARGET_OUT_TESTCASES='/tmp/top/target_out_testcases'
+ echo PRODUCT_OUT='/tmp/top/product_out'
+ echo SOONG_HOST_OUT='/tmp/top/soong_host_out'
+ echo HOST_OUT='/tmp/top/host_out'
+ """)
+ os.chmod(os.path.join(soong_path, 'soong_ui.bash'), 0o666)
+
+ def _write_change_info_file(self):
+ change_info_contents = {
+ 'changes': [{
+ 'projectPath': '/project/path',
+ 'revisions': [{
+ 'fileInfos': [{
+ 'path': 'file/path/file_name',
+ }],
+ }],
+ }]
+ }
+
+ with open(self.change_info_file, 'w') as f:
+ json.dump(change_info_contents, f)
+
+ def _write_test_mapping_file(self):
+ test_mapping_contents = {
+ 'test-mapping-group': [
+ {
+ 'name': 'test_mapping_module',
+ },
+ ],
+ }
+
+ with open('/project/path/file/path/TEST_MAPPING', 'w') as f:
+ json.dump(test_mapping_contents, f)
+
+ def test_general_tests_optimized(self):
+ optimizer = self._create_general_tests_optimizer()
+
+ build_targets = optimizer.get_build_targets()
+
+ expected_build_targets = set(
+ optimized_targets.GeneralTestsOptimizer._REQUIRED_MODULES
+ )
+ expected_build_targets.add('test_mapping_module')
+
+ self.assertSetEqual(build_targets, expected_build_targets)
+
+ def test_no_change_info_no_optimization(self):
+ del os.environ['CHANGE_INFO']
+
+ optimizer = self._create_general_tests_optimizer()
+
+ build_targets = optimizer.get_build_targets()
+
+ self.assertSetEqual(build_targets, {'general-tests'})
+
+ def test_mapping_groups_unused_module_not_built(self):
+ test_context = self._create_test_context()
+ test_context['testInfos'][0]['extraOptions'] = [
+ {
+ 'key': 'additional-files-filter',
+ 'values': ['general-tests.zip'],
+ },
+ {
+ 'key': 'test-mapping-test-group',
+ 'values': ['unused-test-mapping-group'],
+ },
+ ]
+ optimizer = self._create_general_tests_optimizer(
+ build_context=self._create_build_context(test_context=test_context)
+ )
+
+ build_targets = optimizer.get_build_targets()
+
+ expected_build_targets = set(
+ optimized_targets.GeneralTestsOptimizer._REQUIRED_MODULES
+ )
+ self.assertSetEqual(build_targets, expected_build_targets)
+
+ def test_general_tests_used_by_non_test_mapping_test_no_optimization(self):
+ test_context = self._create_test_context()
+ test_context['testInfos'][0]['extraOptions'] = [{
+ 'key': 'additional-files-filter',
+ 'values': ['general-tests.zip'],
+ }]
+ optimizer = self._create_general_tests_optimizer(
+ build_context=self._create_build_context(test_context=test_context)
+ )
+
+ build_targets = optimizer.get_build_targets()
+
+ self.assertSetEqual(build_targets, {'general-tests'})
+
+ def test_malformed_change_info_raises(self):
+ with open(self.change_info_file, 'w') as f:
+ f.write('not change info')
+
+ optimizer = self._create_general_tests_optimizer()
+
+ with self.assertRaises(json.decoder.JSONDecodeError):
+ build_targets = optimizer.get_build_targets()
+
+ def test_malformed_test_mapping_raises(self):
+ with open('/project/path/file/path/TEST_MAPPING', 'w') as f:
+ f.write('not test mapping')
+
+ optimizer = self._create_general_tests_optimizer()
+
+ with self.assertRaises(json.decoder.JSONDecodeError):
+ build_targets = optimizer.get_build_targets()
+
+ @mock.patch('subprocess.run')
+ def test_packaging_outputs_success(self, subprocess_run):
+ subprocess_run.return_value = self._get_soong_vars_output()
+ optimizer = self._create_general_tests_optimizer()
+ self._set_up_build_outputs(['test_mapping_module'])
+
+ targets = optimizer.get_build_targets()
+ package_commands = optimizer.get_package_outputs_commands()
+
+ self._verify_soong_zip_commands(package_commands, ['test_mapping_module'])
+
+ @mock.patch('subprocess.run')
+ def test_get_soong_dumpvars_fails_raises(self, subprocess_run):
+ subprocess_run.return_value = self._get_soong_vars_output(return_code=-1)
+ optimizer = self._create_general_tests_optimizer()
+ self._set_up_build_outputs(['test_mapping_module'])
+
+ targets = optimizer.get_build_targets()
+
+ with self.assertRaisesRegex(RuntimeError, 'Soong dumpvars failed!'):
+ package_commands = optimizer.get_package_outputs_commands()
+
+ @mock.patch('subprocess.run')
+ def test_get_soong_dumpvars_bad_output_raises(self, subprocess_run):
+ subprocess_run.return_value = self._get_soong_vars_output(
+ stdout='This output is bad'
+ )
+ optimizer = self._create_general_tests_optimizer()
+ self._set_up_build_outputs(['test_mapping_module'])
+
+ targets = optimizer.get_build_targets()
+
+ with self.assertRaisesRegex(
+ RuntimeError, 'Error parsing soong dumpvars output'
+ ):
+ package_commands = optimizer.get_package_outputs_commands()
+
+ def _create_general_tests_optimizer(self, build_context: BuildContext = None):
+ if not build_context:
+ build_context = self._create_build_context()
+ return optimized_targets.GeneralTestsOptimizer(
+ 'general-tests', build_context, None
+ )
+
+ def _create_build_context(
+ self,
+ general_tests_optimized: bool = True,
+ test_context: dict[str, any] = None,
+ ) -> BuildContext:
+ if not test_context:
+ test_context = self._create_test_context()
+ build_context_dict = {}
+ build_context_dict['enabledBuildFeatures'] = [{'name': 'optimized_build'}]
+ if general_tests_optimized:
+ build_context_dict['enabledBuildFeatures'].append(
+ {'name': 'general_tests_optimized'}
+ )
+ build_context_dict['testContext'] = test_context
+ return BuildContext(build_context_dict)
+
+ def _create_test_context(self):
+ return {
+ 'testInfos': [
+ {
+ 'name': 'atp_test',
+ 'target': 'test_target',
+ 'branch': 'branch',
+ 'extraOptions': [
+ {
+ 'key': 'additional-files-filter',
+ 'values': ['general-tests.zip'],
+ },
+ {
+ 'key': 'test-mapping-test-group',
+ 'values': ['test-mapping-group'],
+ },
+ ],
+ 'command': '/tf/command',
+ 'extraBuildTargets': [
+ 'extra_build_target',
+ ],
+ },
+ ],
+ }
+
+ def _get_soong_vars_output(
+ self, return_code: int = 0, stdout: str = ''
+ ) -> subprocess.CompletedProcess:
+ return_value = subprocess.CompletedProcess(args=[], returncode=return_code)
+ if not stdout:
+ stdout = textwrap.dedent(f"""\
+ HOST_OUT_TESTCASES='{self._host_out_testcases}'
+ TARGET_OUT_TESTCASES='{self._target_out_testcases}'
+ PRODUCT_OUT='{self._product_out}'
+ SOONG_HOST_OUT='{self._soong_host_out}'
+ HOST_OUT='{self._host_out}'""")
+
+ return_value.stdout = stdout
+ return return_value
+
+ def _set_up_build_outputs(self, targets: list[str]):
+ for target in targets:
+ host_dir = self._host_out_testcases / target
+ host_dir.mkdir()
+ (host_dir / f'{target}.config').touch()
+ (host_dir / f'test_file').touch()
+
+ target_dir = self._target_out_testcases / target
+ target_dir.mkdir()
+ (target_dir / f'{target}.config').touch()
+ (target_dir / f'test_file').touch()
+
+ def _verify_soong_zip_commands(self, commands: list[str], targets: list[str]):
+ """Verify the structure of the zip commands.
+
+ Zip commands have to start with the soong_zip binary path, then are followed
+ by a couple of options and the name of the file being zipped. Depending on
+ which zip we are creating look for a few essential items being added in
+ those zips.
+
+ Args:
+ commands: list of command lists
+ targets: list of targets expected to be in general-tests.zip
+ """
+ for command in commands:
+ self.assertEqual(
+ '/tmp/top/prebuilts/build-tools/linux-x86/bin/soong_zip',
+ command[0],
+ )
+ self.assertEqual('-d', command[1])
+ self.assertEqual('-o', command[2])
+ match (command[3]):
+ case '/tmp/top/out/dist/general-tests_configs.zip':
+ self.assertIn(f'{self._host_out}/host_general-tests_list', command)
+ self.assertIn(
+ f'{self._product_out}/target_general-tests_list', command
+ )
+ return
+ case '/tmp/top/out/dist/general-tests_list.zip':
+ self.assertIn('-f', command)
+ self.assertIn(f'{self._host_out}/general-tests_list', command)
+ return
+ case '/tmp/top/out/dist/general-tests.zip':
+ for target in targets:
+ self.assertIn(f'{self._host_out_testcases}/{target}', command)
+ self.assertIn(f'{self._target_out_testcases}/{target}', command)
+ self.assertIn(
+ f'{self._soong_host_out}/framework/cts-tradefed.jar', command
+ )
+ self.assertIn(
+ f'{self._soong_host_out}/framework/compatibility-host-util.jar',
+ command,
+ )
+ self.assertIn(
+ f'{self._soong_host_out}/framework/vts-tradefed.jar', command
+ )
+ return
+ case _:
+ self.fail(f'malformed command: {command}')
+
+
+if __name__ == '__main__':
+ # Setup logging to be silent so unit tests can pass through TF.
+ logging.disable(logging.ERROR)
+ unittest.main()
diff --git a/ci/test_mapping_module_retriever.py b/ci/test_mapping_module_retriever.py
index d2c13c0..c93cdd5 100644
--- a/ci/test_mapping_module_retriever.py
+++ b/ci/test_mapping_module_retriever.py
@@ -17,11 +17,13 @@
modules are needed to build for the given list of changed files.
TODO(lucafarsi): Deduplicate from artifact_helper.py
"""
+# TODO(lucafarsi): Share this logic with the original logic in
+# test_mapping_test_retriever.py
-from typing import Any, Dict, Set, Text
import json
import os
import re
+from typing import Any
# Regex to extra test name from the path of test config file.
TEST_NAME_REGEX = r'(?:^|.*/)([^/]+)\.config'
@@ -39,7 +41,7 @@
_COMMENTS_RE = re.compile(r'(\"(?:[^\"\\]|\\.)*\"|(?=//))(?://.*)?')
-def FilterComments(test_mapping_file: Text) -> Text:
+def FilterComments(test_mapping_file: str) -> str:
"""Remove comments in TEST_MAPPING file to valid format.
Only '//' is regarded as comments.
@@ -52,8 +54,8 @@
"""
return re.sub(_COMMENTS_RE, r'\1', test_mapping_file)
-def GetTestMappings(paths: Set[Text],
- checked_paths: Set[Text]) -> Dict[Text, Dict[Text, Any]]:
+def GetTestMappings(paths: set[str],
+ checked_paths: set[str]) -> dict[str, dict[str, Any]]:
"""Get the affected TEST_MAPPING files.
TEST_MAPPING files in source code are packaged into a build artifact
@@ -123,3 +125,68 @@
pass
return test_mappings
+
+
+def FindAffectedModules(
+ test_mappings: dict[str, Any],
+ changed_files: set[str],
+ test_mapping_test_groups: set[str],
+) -> set[str]:
+ """Find affected test modules.
+
+ Find the affected set of test modules that would run in a test mapping run based on the given test mappings, changed files, and test mapping test group.
+
+ Args:
+ test_mappings: A set of test mappings returned by GetTestMappings in the following format:
+ {
+ 'test_mapping_file_path': {
+ 'group_name' : [
+ 'name': 'module_name',
+ ],
+ }
+ }
+ changed_files: A set of files changed for the given run.
+ test_mapping_test_groups: A set of test mapping test groups that are being considered for the given run.
+
+ Returns:
+ A set of test module names which would run for a test mapping test run with the given parameters.
+ """
+
+ modules = set()
+
+ for test_mapping in test_mappings.values():
+ for group_name, group in test_mapping.items():
+ # If a module is not in any of the test mapping groups being tested skip
+ # it.
+ if group_name not in test_mapping_test_groups:
+ continue
+
+ for entry in group:
+ module_name = entry.get('name')
+
+ if not module_name:
+ continue
+
+ file_patterns = entry.get('file_patterns')
+ if not file_patterns:
+ modules.add(module_name)
+ continue
+
+ if matches_file_patterns(file_patterns, changed_files):
+ modules.add(module_name)
+
+ return modules
+
+def MatchesFilePatterns(
+ file_patterns: list[set], changed_files: set[str]
+) -> bool:
+ """Checks if any of the changed files match any of the file patterns.
+
+ Args:
+ file_patterns: A list of file patterns to match against.
+ changed_files: A set of files to check against the file patterns.
+
+ Returns:
+ True if any of the changed files match any of the file patterns.
+ """
+ return any(re.search(pattern, "|".join(changed_files)) for pattern in file_patterns)
diff --git a/cogsetup.sh b/cogsetup.sh
deleted file mode 100644
index ef1485d..0000000
--- a/cogsetup.sh
+++ /dev/null
@@ -1,71 +0,0 @@
-#
-# Copyright (C) 2023 The Android Open Source Project
-#
-# Licensed under the Apache License, Version 2.0 (the "License");
-# you may not use this file except in compliance with the License.
-# You may obtain a copy of the License at
-#
-# http://www.apache.org/licenses/LICENSE-2.0
-#
-# Unless required by applicable law or agreed to in writing, software
-# distributed under the License is distributed on an "AS IS" BASIS,
-# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-# See the License for the specific language governing permissions and
-# limitations under the License.
-#
-
-# This file is executed by build/envsetup.sh, and can use anything
-# defined in envsetup.sh.
-function _create_out_symlink_for_cog() {
- if [[ "${OUT_DIR}" == "" ]]; then
- OUT_DIR="out"
- fi
-
- # getoutdir ensures paths are absolute. envsetup could be called from a
- # directory other than the root of the source tree
- local outdir=$(getoutdir)
- if [[ -L "${outdir}" ]]; then
- return
- fi
- if [ -d "${outdir}" ]; then
- echo -e "\tOutput directory ${outdir} cannot be present in a Cog workspace."
- echo -e "\tDelete \"${outdir}\" or create a symlink from \"${outdir}\" to a directory outside your workspace."
- return 1
- fi
-
- DEFAULT_OUTPUT_DIR="${HOME}/.cog/android-build-out"
- mkdir -p ${DEFAULT_OUTPUT_DIR}
- ln -s ${DEFAULT_OUTPUT_DIR} ${outdir}
-}
-
-# This function sets up the build environment to be appropriate for Cog.
-function _setup_cog_env() {
- _create_out_symlink_for_cog
- if [ "$?" -eq "1" ]; then
- echo -e "\e[0;33mWARNING:\e[00m Cog environment setup failed!"
- return 1
- fi
-
- export ANDROID_BUILD_ENVIRONMENT_CONFIG="googler-cog"
-
- # Running repo command within Cog workspaces is not supported, so override
- # it with this function. If the user is running repo within a Cog workspace,
- # we'll fail with an error, otherwise, we run the original repo command with
- # the given args.
- if ! ORIG_REPO_PATH=`which repo`; then
- return 0
- fi
- function repo {
- if [[ "${PWD}" == /google/cog/* ]]; then
- echo "\e[01;31mERROR:\e[0mrepo command is disallowed within Cog workspaces."
- return 1
- fi
- ${ORIG_REPO_PATH} "$@"
- }
-}
-
-if [[ "${PWD}" != /google/cog/* ]]; then
- echo -e "\e[01;31mERROR:\e[0m This script must be run from a Cog workspace."
-fi
-
-_setup_cog_env
diff --git a/core/Makefile b/core/Makefile
index d2e712e..db038c3 100644
--- a/core/Makefile
+++ b/core/Makefile
@@ -1964,7 +1964,7 @@
installed_system_dlkm_notice_xml_gz := $(TARGET_OUT_SYSTEM_DLKM)/etc/NOTICE.xml.gz
ALL_INSTALLED_NOTICE_FILES := \
- $(installed_notice_html_or_xml_gz) \
+ $(if $(USE_SOONG_DEFINED_SYSTEM_IMAGE),,$(installed_notice_html_or_xml_gz)) \
$(installed_vendor_notice_xml_gz) \
$(installed_product_notice_xml_gz) \
$(installed_system_ext_notice_xml_gz) \
@@ -2051,7 +2051,9 @@
endif # PRODUCT_NOTICE_SPLIT
+ifneq ($(USE_SOONG_DEFINED_SYSTEM_IMAGE),true)
ALL_DEFAULT_INSTALLED_MODULES += $(installed_notice_html_or_xml_gz)
+endif
need_vendor_notice:=false
ifeq ($(BUILDING_VENDOR_BOOT_IMAGE),true)
@@ -3416,8 +3418,10 @@
# system image
INSTALLED_FILES_OUTSIDE_IMAGES := $(filter-out $(TARGET_OUT)/%, $(INSTALLED_FILES_OUTSIDE_IMAGES))
+ifdef BUILDING_SYSTEM_IMAGE
INTERNAL_SYSTEMIMAGE_FILES := $(sort $(filter $(TARGET_OUT)/%, \
$(ALL_DEFAULT_INSTALLED_MODULES)))
+endif
# Create symlink /system/vendor to /vendor if necessary.
ifdef BOARD_USES_VENDORIMAGE
@@ -3509,6 +3513,8 @@
--output $@ --value "$(STUB_LIBRARIES)" --system "$(TARGET_OUT)"
$(HOST_OUT_EXECUTABLES)/conv_linker_config append --source $@ --output $@ --key requireLibs \
--value "$(foreach lib,$(LLNDK_MOVED_TO_APEX_LIBRARIES), $(lib).so)"
+ $(HOST_OUT_EXECUTABLES)/conv_linker_config append --source $@ --output $@ --key provideLibs \
+ --value "$(foreach lib,$(PRODUCT_EXTRA_STUB_LIBRARIES), $(lib).so)"
$(call declare-1p-target,$(SYSTEM_LINKER_CONFIG),)
$(call declare-license-deps,$(SYSTEM_LINKER_CONFIG),$(INTERNAL_SYSTEMIMAGE_FILES) $(SYSTEM_LINKER_CONFIG_SOURCE))
@@ -3586,10 +3592,10 @@
ifeq ($(PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE),)
$(error PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE must be set if USE_SOONG_DEFINED_SYSTEM_IMAGE is true)
endif
-soong_defined_system_image := $(call intermediates-dir-for,ETC,$(PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE))/$(PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE)
-$(BUILT_SYSTEMIMAGE): $(INSTALLED_FILES_FILE) $(systemimage_intermediates)/file_list.txt $(soong_defined_system_image)
-$(eval $(call copy-one-file, $(soong_defined_system_image), $(BUILT_SYSTEMIMAGE)))
-soong_defined_system_image :=
+SOONG_DEFINED_SYSTEM_IMAGE_PATH := $(call intermediates-dir-for,ETC,$(PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE))/$(PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE)
+SOONG_DEFINED_SYSTEM_IMAGE_BASE := $(dir $(ALL_MODULES.$(PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE).FILESYSTEM_FILELIST))
+$(BUILT_SYSTEMIMAGE): $(INSTALLED_FILES_FILE) $(systemimage_intermediates)/file_list.txt $(SOONG_DEFINED_SYSTEM_IMAGE_PATH)
+$(eval $(call copy-one-file, $(SOONG_DEFINED_SYSTEM_IMAGE_PATH), $(BUILT_SYSTEMIMAGE)))
else
$(BUILT_SYSTEMIMAGE): $(FULL_SYSTEMIMAGE_DEPS) $(INSTALLED_FILES_FILE) $(systemimage_intermediates)/file_list.txt
$(call build-systemimage-target,$@)
@@ -3674,10 +3680,10 @@
# -----------------------------------------------------------------
# data partition image
INSTALLED_FILES_OUTSIDE_IMAGES := $(filter-out $(TARGET_OUT_DATA)/%, $(INSTALLED_FILES_OUTSIDE_IMAGES))
+ifdef BUILDING_USERDATA_IMAGE
INTERNAL_USERDATAIMAGE_FILES := \
$(filter $(TARGET_OUT_DATA)/%,$(ALL_DEFAULT_INSTALLED_MODULES))
-ifdef BUILDING_USERDATA_IMAGE
userdataimage_intermediates := \
$(call intermediates-dir-for,PACKAGING,userdata)
BUILT_USERDATAIMAGE_TARGET := $(PRODUCT_OUT)/userdata.img
@@ -6132,6 +6138,9 @@
$(BUILT_TARGET_FILES_DIR): zip_root := $(intermediates)/$(name)
$(BUILT_TARGET_FILES_DIR): intermediates := $(intermediates)
+ifneq ($(SOONG_DEFINED_SYSTEM_IMAGE_PATH),)
+ $(BUILT_TARGET_FILES_DIR): $(SOONG_DEFINED_SYSTEM_IMAGE_PATH)
+endif
# $(1): Directory to copy
# $(2): Location to copy it to
@@ -6460,8 +6469,11 @@
$(INSTALLED_RAMDISK_TARGET) \
$(INSTALLED_DTBIMAGE_TARGET) \
$(INSTALLED_2NDBOOTLOADER_TARGET) \
+ $(INSTALLED_VENDOR_KERNEL_RAMDISK_TARGET) \
$(BUILT_RAMDISK_16K_TARGET) \
$(BUILT_KERNEL_16K_TARGET) \
+ $(BUILT_BOOTIMAGE_16K_TARGET) \
+ $(INSTALLED_DTBOIMAGE_16KB_TARGET) \
$(BOARD_PREBUILT_DTBOIMAGE) \
$(BOARD_PREBUILT_RECOVERY_DTBOIMAGE) \
$(BOARD_RECOVERY_ACPIO) \
@@ -6615,8 +6627,13 @@
endif # INSTALLED_VENDOR_BOOTIMAGE_TARGET
ifdef BUILDING_SYSTEM_IMAGE
@# Contents of the system image
+ifneq ($(SOONG_DEFINED_SYSTEM_IMAGE_PATH),)
+ $(hide) $(call package_files-copy-root, \
+ $(SOONG_DEFINED_SYSTEM_IMAGE_BASE)/root/system,$(zip_root)/SYSTEM)
+else
$(hide) $(call package_files-copy-root, \
$(SYSTEMIMAGE_SOURCE_DIR),$(zip_root)/SYSTEM)
+endif
else ifdef INSTALLED_BUILD_PROP_TARGET
@# Copy the system build.prop even if not building a system image
@# because add_img_to_target_files may need it to build other partition
@@ -6810,14 +6827,22 @@
$(hide) mkdir -p $(zip_root)/PREBUILT_IMAGES
$(hide) cp $(INSTALLED_DTBOIMAGE_TARGET) $(zip_root)/PREBUILT_IMAGES/
endif # BOARD_PREBUILT_DTBOIMAGE
-ifdef BUILT_KERNEL_16K_TARGET
+ifdef BOARD_KERNEL_PATH_16K
$(hide) mkdir -p $(zip_root)/PREBUILT_IMAGES
$(hide) cp $(BUILT_KERNEL_16K_TARGET) $(zip_root)/PREBUILT_IMAGES/
-endif # BUILT_KERNEL_16K_TARGET
-ifdef BUILT_RAMDISK_16K_TARGET
+endif # BOARD_KERNEL_PATH_16K
+ifdef BOARD_KERNEL_MODULES_16K
$(hide) mkdir -p $(zip_root)/PREBUILT_IMAGES
$(hide) cp $(BUILT_RAMDISK_16K_TARGET) $(zip_root)/PREBUILT_IMAGES/
-endif # BUILT_RAMDISK_16K_TARGET
+endif # BOARD_KERNEL_MODULES_16K
+ifdef BUILT_BOOTIMAGE_16K_TARGET
+ $(hide) mkdir -p $(zip_root)/IMAGES
+ $(hide) cp $(BUILT_BOOTIMAGE_16K_TARGET) $(zip_root)/IMAGES/
+endif # BUILT_BOOTIMAGE_16K_TARGET
+ifdef INSTALLED_DTBOIMAGE_16KB_TARGET
+ $(hide) mkdir -p $(zip_root)/IMAGES
+ $(hide) cp $(INSTALLED_DTBOIMAGE_16KB_TARGET) $(zip_root)/IMAGES/
+endif # INSTALLED_DTBOIMAGE_16KB_TARGET
ifeq ($(BOARD_USES_PVMFWIMAGE),true)
$(hide) mkdir -p $(zip_root)/PREBUILT_IMAGES
$(hide) cp $(INSTALLED_PVMFWIMAGE_TARGET) $(zip_root)/PREBUILT_IMAGES/
@@ -7925,9 +7950,14 @@
# Desktop pack recovery image hook.
ifneq (,$(strip $(PACK_DESKTOP_RECOVERY_IMAGE)))
PACK_RECOVERY_IMAGE_TARGET := $(PRODUCT_OUT)/android-desktop_recovery_image.bin
+PACK_RECOVERY_IMAGE_ARGS := --noarchive --recovery
+
+ifneq (,$(strip $(PACK_RECOVERY_IMAGE_EXPERIMENTAL)))
+PACK_RECOVERY_IMAGE_ARGS += --experimental
+endif # PACK_RECOVERY_IMAGE_EXPERIMENTAL
$(PACK_RECOVERY_IMAGE_TARGET): $(IMAGES) $(PACK_IMAGE_SCRIPT)
- $(PACK_IMAGE_SCRIPT) --out_dir $(PRODUCT_OUT) --noarchive --recovery
+ $(PACK_IMAGE_SCRIPT) --out_dir $(PRODUCT_OUT) $(PACK_RECOVERY_IMAGE_ARGS)
PACKED_RECOVERY_IMAGE_ARCHIVE_TARGET := $(PACK_RECOVERY_IMAGE_TARGET).gz
@@ -7942,6 +7972,26 @@
endif # PACK_DESKTOP_RECOVERY_IMAGE
# -----------------------------------------------------------------
+# Desktop pack update image hook.
+ifneq (,$(strip $(PACK_DESKTOP_UPDATE_IMAGE)))
+PACK_UPDATE_IMAGE_TARGET := $(PRODUCT_OUT)/android-desktop_update_image.bin
+
+$(PACK_UPDATE_IMAGE_TARGET): $(IMAGES) $(PACK_IMAGE_SCRIPT)
+ $(PACK_IMAGE_SCRIPT) --out_dir $(PRODUCT_OUT) --noarchive --update
+
+PACKED_UPDATE_IMAGE_ARCHIVE_TARGET := $(PACK_UPDATE_IMAGE_TARGET).gz
+
+$(PACKED_UPDATE_IMAGE_ARCHIVE_TARGET): $(PACK_UPDATE_IMAGE_TARGET) | $(GZIP)
+ $(GZIP) -fk $(PACK_UPDATE_IMAGE_TARGET)
+
+$(call dist-for-goals,dist_files,$(PACKED_UPDATE_IMAGE_ARCHIVE_TARGET))
+
+.PHONY: pack-update-image
+pack-update-image: $(PACK_UPDATE_IMAGE_TARGET)
+
+endif # PACK_DESKTOP_UPDATE_IMAGE
+
+# -----------------------------------------------------------------
# OS Licensing
include $(BUILD_SYSTEM)/os_licensing.mk
diff --git a/core/android_soong_config_vars.mk b/core/android_soong_config_vars.mk
index 5fc8fd4..f66bdd9 100644
--- a/core/android_soong_config_vars.mk
+++ b/core/android_soong_config_vars.mk
@@ -47,6 +47,8 @@
$(call soong_config_set_bool,ANDROID,CLANG_COVERAGE,$(CLANG_COVERAGE))
$(call soong_config_set,ANDROID,SCUDO_ALLOCATION_RING_BUFFER_SIZE,$(PRODUCT_SCUDO_ALLOCATION_RING_BUFFER_SIZE))
+$(call soong_config_set_bool,ANDROID,EMMA_INSTRUMENT,$(if $(filter true,$(EMMA_INSTRUMENT)),true,false))
+
# PRODUCT_PRECOMPILED_SEPOLICY defaults to true. Explicitly check if it's "false" or not.
$(call soong_config_set_bool,ANDROID,PRODUCT_PRECOMPILED_SEPOLICY,$(if $(filter false,$(PRODUCT_PRECOMPILED_SEPOLICY)),false,true))
@@ -182,3 +184,23 @@
# Add target_use_pan_display flag for hardware/libhardware:gralloc.default
$(call soong_config_set_bool,gralloc,target_use_pan_display,$(if $(filter true,$(TARGET_USE_PAN_DISPLAY)),true,false))
+
+# Add use_camera_v4l2_hal flag for hardware/libhardware/modules/camera/3_4:camera.v4l2
+$(call soong_config_set_bool,camera,use_camera_v4l2_hal,$(if $(filter true,$(USE_CAMERA_V4L2_HAL)),true,false))
+
+# Add audioserver_multilib flag for hardware/interfaces/soundtrigger/2.0/default:android.hardware.soundtrigger@2.0-impl
+ifneq ($(strip $(AUDIOSERVER_MULTILIB)),)
+ $(call soong_config_set,soundtrigger,audioserver_multilib,$(AUDIOSERVER_MULTILIB))
+endif
+
+# Add sim_count, disable_rild_oem_hook, and use_aosp_rild flag for ril related modules
+$(call soong_config_set,ril,sim_count,$(SIM_COUNT))
+ifneq ($(DISABLE_RILD_OEM_HOOK), false)
+ $(call soong_config_set_bool,ril,disable_rild_oem_hook,true)
+endif
+ifneq ($(ENABLE_VENDOR_RIL_SERVICE), true)
+ $(call soong_config_set_bool,ril,use_aosp_rild,true)
+endif
+
+# Export target_board_platform to soong for hardware/google/graphics/common/libmemtrack:memtrack.$(TARGET_BOARD_PLATFORM)
+$(call soong_config_set,ANDROID,target_board_platform,$(TARGET_BOARD_PLATFORM))
diff --git a/core/base_rules.mk b/core/base_rules.mk
index 1135003..5363e0f 100644
--- a/core/base_rules.mk
+++ b/core/base_rules.mk
@@ -340,7 +340,7 @@
ifneq (,$(LOCAL_SOONG_INSTALLED_MODULE))
ifneq ($(LOCAL_MODULE_MAKEFILE),$(SOONG_ANDROID_MK))
- $(call pretty-error, LOCAL_SOONG_INSTALLED_MODULE can only be used from $(SOONG_ANDROID_MK))
+ $(call pretty-error, LOCAL_MODULE_MAKEFILE can only be used from $(SOONG_ANDROID_MK))
endif
# Use the install path requested by Soong.
LOCAL_INSTALLED_MODULE := $(LOCAL_SOONG_INSTALLED_MODULE)
diff --git a/core/binary.mk b/core/binary.mk
index 1e98bc0..3481144 100644
--- a/core/binary.mk
+++ b/core/binary.mk
@@ -330,18 +330,20 @@
ifneq ($(LOCAL_IN_VENDOR),)
# Vendor modules have LOCAL_IN_VENDOR
my_cflags += -D__ANDROID_VENDOR__
-
- ifeq ($(BOARD_API_LEVEL),)
- # TODO(b/314036847): This is a fallback for UDC targets.
- # This must be a build failure when UDC is no longer built from this source tree.
- my_cflags += -D__ANDROID_VENDOR_API__=$(PLATFORM_SDK_VERSION)
- else
- my_cflags += -D__ANDROID_VENDOR_API__=$(BOARD_API_LEVEL)
- endif
else ifneq ($(LOCAL_IN_PRODUCT),)
# Product modules have LOCAL_IN_PRODUCT
my_cflags += -D__ANDROID_PRODUCT__
endif
+
+ # Define __ANDROID_VENDOR_API__ for both product and vendor variants because
+ # they both use the same LLNDK libraries.
+ ifeq ($(BOARD_API_LEVEL),)
+ # TODO(b/314036847): This is a fallback for UDC targets.
+ # This must be a build failure when UDC is no longer built from this source tree.
+ my_cflags += -D__ANDROID_VENDOR_API__=$(PLATFORM_SDK_VERSION)
+ else
+ my_cflags += -D__ANDROID_VENDOR_API__=$(BOARD_API_LEVEL)
+ endif
endif
ifndef LOCAL_IS_HOST_MODULE
diff --git a/core/board_config.mk b/core/board_config.mk
index d3f0493..5606964 100644
--- a/core/board_config.mk
+++ b/core/board_config.mk
@@ -237,6 +237,7 @@
.KATI_READONLY := TARGET_DEVICE_DIR
endif
+$(call dump-phase-start,BOARD,,,, build/make/core/board_config.mk)
ifndef RBC_PRODUCT_CONFIG
include $(board_config_mk)
else
@@ -261,6 +262,7 @@
include $(OUT_DIR)/rbc/rbc_board_config_results.mk
endif
+$(call dump-phase-end, build/make/core/board_config.mk)
ifneq (,$(and $(TARGET_ARCH),$(TARGET_ARCH_SUITE)))
$(error $(board_config_mk) erroneously sets both TARGET_ARCH and TARGET_ARCH_SUITE)
diff --git a/core/combo/arch/arm64/armv9-2a.mk b/core/combo/arch/arm64/armv9-2a.mk
new file mode 100644
index 0000000..69ffde0
--- /dev/null
+++ b/core/combo/arch/arm64/armv9-2a.mk
@@ -0,0 +1,18 @@
+#
+# Copyright (C) 2023 The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+#
+
+# .mk file required to support build for the ARMv9.2-A arch variant.
+# The file just needs to be present, it does not need to contain anything.
diff --git a/core/combo/arch/x86/alderlake.mk b/core/combo/arch/x86/alderlake.mk
new file mode 100644
index 0000000..a7ae6ed
--- /dev/null
+++ b/core/combo/arch/x86/alderlake.mk
@@ -0,0 +1,6 @@
+# Configuration for Linux on x86.
+# Generating binaries for processors
+# that have AVX2 feature flag
+#
+
+ARCH_X86_HAVE_SSE4_1 := true
diff --git a/core/combo/arch/x86_64/alderlake.mk b/core/combo/arch/x86_64/alderlake.mk
new file mode 100644
index 0000000..a7ae6ed
--- /dev/null
+++ b/core/combo/arch/x86_64/alderlake.mk
@@ -0,0 +1,6 @@
+# Configuration for Linux on x86.
+# Generating binaries for processors
+# that have AVX2 feature flag
+#
+
+ARCH_X86_HAVE_SSE4_1 := true
diff --git a/core/config.mk b/core/config.mk
index 2f18add..192c8b2 100644
--- a/core/config.mk
+++ b/core/config.mk
@@ -173,6 +173,7 @@
$(KATI_obsolete_var BUILDING_PVMFW_IMAGE,BUILDING_PVMFW_IMAGE is no longer used)
$(KATI_obsolete_var BOARD_BUILD_SYSTEM_ROOT_IMAGE)
$(KATI_obsolete_var FS_GET_STATS)
+$(KATI_obsolete_var BUILD_BROKEN_USES_SOONG_PYTHON2_MODULES)
# Used to force goals to build. Only use for conditionally defined goals.
.PHONY: FORCE
@@ -363,8 +364,7 @@
# configs, generally for cross-cutting features.
# Build broken variables that should be treated as booleans
-_build_broken_bool_vars := \
- BUILD_BROKEN_USES_SOONG_PYTHON2_MODULES \
+_build_broken_bool_vars :=
# Build broken variables that should be treated as lists
_build_broken_list_vars := \
@@ -811,6 +811,12 @@
BOARD_PROPERTY_OVERRIDES_SPLIT_ENABLED ?= true
endif
+ifneq ($(call math_gt_or_eq,$(PRODUCT_SHIPPING_API_LEVEL),36),)
+ ifneq ($(NEED_AIDL_NDK_PLATFORM_BACKEND),)
+ $(error Must not set NEED_AIDL_NDK_PLATFORM_BACKEND, but it is set to: $(NEED_AIDL_NDK_PLATFORM_BACKEND). Support will be removed.)
+ endif
+endif
+
# Set BOARD_SYSTEMSDK_VERSIONS to the latest SystemSDK version starting from P-launching
# devices if unset.
ifndef BOARD_SYSTEMSDK_VERSIONS
@@ -1201,6 +1207,11 @@
APPS_DEFAULT_VERSION_NAME := $(PLATFORM_VERSION)
+# Add BUILD_NUMBER to apps if PRODUCT_BUILD_APPS_WITH_BUILD_NUMBER is defined.
+ifeq ($(PRODUCT_BUILD_APPS_WITH_BUILD_NUMBER),true)
+ APPS_DEFAULT_VERSION_NAME := $(PLATFORM_VERSION)-$(BUILD_NUMBER_FROM_FILE)
+endif
+
# ANDROID_WARNING_ALLOWED_PROJECTS is generated by build/soong.
define find_warning_allowed_projects
$(filter $(ANDROID_WARNING_ALLOWED_PROJECTS),$(1)/)
@@ -1258,7 +1269,15 @@
TARGET_PRODUCT_PROP := $(wildcard $(TARGET_DEVICE_DIR)/product.prop)
endif
-.KATI_READONLY := TARGET_SYSTEM_PROP TARGET_SYSTEM_EXT_PROP TARGET_PRODUCT_PROP
+ifeq ($(TARGET_ODM_PROP),)
+TARGET_ODM_PROP := $(wildcard $(TARGET_DEVICE_DIR)/odm.prop)
+endif
+
+.KATI_READONLY := \
+ TARGET_SYSTEM_PROP \
+ TARGET_SYSTEM_EXT_PROP \
+ TARGET_PRODUCT_PROP \
+ TARGET_ODM_PROP \
include $(BUILD_SYSTEM)/sysprop_config.mk
@@ -1266,8 +1285,15 @@
# consistency with those defined in BoardConfig.mk files.
include $(BUILD_SYSTEM)/android_soong_config_vars.mk
-SOONG_VARIABLES := $(SOONG_OUT_DIR)/soong.$(TARGET_PRODUCT).variables
-SOONG_EXTRA_VARIABLES := $(SOONG_OUT_DIR)/soong.$(TARGET_PRODUCT).extra.variables
+# EMMA_INSTRUMENT is set to true when coverage is enabled. Creates a suffix to
+# differeciate the coverage version of ninja files. This will save 5 minutes of
+# build time used to regenerate ninja.
+ifeq (true,$(EMMA_INSTRUMENT))
+COVERAGE_SUFFIX := .coverage
+endif
+
+SOONG_VARIABLES := $(SOONG_OUT_DIR)/soong.$(TARGET_PRODUCT)$(COVERAGE_SUFFIX).variables
+SOONG_EXTRA_VARIABLES := $(SOONG_OUT_DIR)/soong.$(TARGET_PRODUCT)$(COVERAGE_SUFFIX).extra.variables
ifeq ($(CALLED_FROM_SETUP),true)
include $(BUILD_SYSTEM)/ninja_config.mk
diff --git a/core/config_sanitizers.mk b/core/config_sanitizers.mk
index ab2d5c1..c0f2c68 100644
--- a/core/config_sanitizers.mk
+++ b/core/config_sanitizers.mk
@@ -284,9 +284,9 @@
ifneq ($(filter memtag_stack,$(my_sanitize)),)
my_cflags += -fsanitize=memtag-stack
my_ldflags += -fsanitize=memtag-stack
- my_cflags += -march=armv8a+memtag
- my_ldflags += -march=armv8a+memtag
- my_asflags += -march=armv8a+memtag
+ my_cflags += -Xclang -target-feature -Xclang +mte
+ my_ldflags += -Xclang -target-feature -Xclang +mte
+ my_asflags += -Xclang -target-feature -Xclang +mte
my_sanitize := $(filter-out memtag_stack,$(my_sanitize))
endif
diff --git a/core/dex_preopt.mk b/core/dex_preopt.mk
index 906d7f0..88e0cc7 100644
--- a/core/dex_preopt.mk
+++ b/core/dex_preopt.mk
@@ -13,28 +13,6 @@
install-on-system-other = $(filter-out $(PRODUCT_DEXPREOPT_SPEED_APPS) $(PRODUCT_SYSTEM_SERVER_APPS),$(basename $(notdir $(filter $(foreach f,$(SYSTEM_OTHER_ODEX_FILTER),$(TARGET_OUT)/$(f)),$(1)))))
endif
-# Install boot images for testing on host. We exclude framework image as it is not part of art manifest.
-my_boot_image_arch := HOST_ARCH
-my_boot_image_out := $(HOST_OUT)
-my_boot_image_syms := $(HOST_OUT)/symbols
-HOST_BOOT_IMAGE_MODULE := \
- $(foreach my_boot_image_name,art_host,$(strip \
- $(eval include $(BUILD_SYSTEM)/dex_preopt_libart.mk) \
- $(my_boot_image_module)))
-HOST_BOOT_IMAGE := $(call module-installed-files,$(HOST_BOOT_IMAGE_MODULE))
-ifdef HOST_2ND_ARCH
- my_boot_image_arch := HOST_2ND_ARCH
- 2ND_HOST_BOOT_IMAGE_MODULE := \
- $(foreach my_boot_image_name,art_host,$(strip \
- $(eval include $(BUILD_SYSTEM)/dex_preopt_libart.mk) \
- $(my_boot_image_module)))
- 2ND_HOST_BOOT_IMAGE := $(call module-installed-files,$(2ND_HOST_BOOT_IMAGE_MODULE))
-endif
-my_boot_image_arch :=
-my_boot_image_out :=
-my_boot_image_syms :=
-my_boot_image_module :=
-
# Build the boot.zip which contains the boot jars and their compilation output
# We can do this only if preopt is enabled and if the product uses libart config (which sets the
# default properties for preopting).
diff --git a/core/dex_preopt_config.mk b/core/dex_preopt_config.mk
index d51de33..f1e9fb5 100644
--- a/core/dex_preopt_config.mk
+++ b/core/dex_preopt_config.mk
@@ -1,4 +1,4 @@
-DEX_PREOPT_CONFIG := $(SOONG_OUT_DIR)/dexpreopt.config
+DEX_PREOPT_CONFIG := $(SOONG_OUT_DIR)/dexpreopt${COVERAGE_SUFFIX}.config
ENABLE_PREOPT := true
ENABLE_PREOPT_BOOT_IMAGES := true
diff --git a/core/dex_preopt_libart.mk b/core/dex_preopt_libart.mk
deleted file mode 100644
index a2c9942..0000000
--- a/core/dex_preopt_libart.mk
+++ /dev/null
@@ -1,109 +0,0 @@
-####################################
-# ART boot image installation
-# Input variables:
-# my_boot_image_name: the boot image to install
-# my_boot_image_arch: the architecture to install (e.g. TARGET_ARCH, not expanded)
-# my_boot_image_out: the install directory (e.g. $(PRODUCT_OUT))
-# my_boot_image_syms: the symbols director (e.g. $(TARGET_OUT_UNSTRIPPED))
-#
-# Output variables:
-# my_boot_image_module: the created module name. Empty if no module is created.
-#
-# Install the boot images compiled by Soong.
-# Create a module named dexpreopt_bootjar.$(my_boot_image_name)_$($(my_boot_image_arch))
-# that installs all of boot image files.
-# If there is no file to install for $(my_boot_image_name), for example when
-# building an unbundled build, then no module is created.
-#
-####################################
-
-# Takes a list of src:dest install pairs and returns a new list with a path
-# prefixed to each dest value.
-# $(1): list of src:dest install pairs
-# $(2): path to prefix to each dest value
-define prefix-copy-many-files-dest
-$(foreach v,$(1),$(call word-colon,1,$(v)):$(2)$(call word-colon,2,$(v)))
-endef
-
-# Converts an architecture-specific vdex path into a location that can be shared
-# between architectures.
-define vdex-shared-install-path
-$(dir $(patsubst %/,%,$(dir $(1))))$(notdir $(1))
-endef
-
-# Takes a list of src:dest install pairs of vdex files and returns a new list
-# where each dest has been rewritten to the shared location for vdex files.
-define vdex-copy-many-files-shared-dest
-$(foreach v,$(1),$(call word-colon,1,$(v)):$(call vdex-shared-install-path,$(call word-colon,2,$(v))))
-endef
-
-# Creates a rule to symlink an architecture specific vdex file to the shared
-# location for that vdex file.
-define symlink-vdex-file
-$(strip \
- $(call symlink-file,\
- $(call vdex-shared-install-path,$(1)),\
- ../$(notdir $(1)),\
- $(1))\
- $(1))
-endef
-
-# Takes a list of src:dest install pairs of vdex files and creates rules to
-# symlink each dest to the shared location for that vdex file.
-define symlink-vdex-files
-$(foreach v,$(1),$(call symlink-vdex-file,$(call word-colon,2,$(v))))
-endef
-
-my_boot_image_module :=
-
-my_suffix := $(my_boot_image_name)_$($(my_boot_image_arch))
-my_copy_pairs := $(call prefix-copy-many-files-dest,$(DEXPREOPT_IMAGE_BUILT_INSTALLED_$(my_suffix)),$(my_boot_image_out))
-my_vdex_copy_pairs := $(call prefix-copy-many-files-dest,$(DEXPREOPT_IMAGE_VDEX_BUILT_INSTALLED_$(my_suffix)),$(my_boot_image_out))
-my_vdex_copy_shared_pairs := $(call vdex-copy-many-files-shared-dest,$(my_vdex_copy_pairs))
-ifeq (,$(filter %_2ND_ARCH,$(my_boot_image_arch)))
- # Only install the vdex to the shared location for the primary architecture.
- my_copy_pairs += $(my_vdex_copy_shared_pairs)
-endif
-
-my_unstripped_copy_pairs := $(call prefix-copy-many-files-dest,$(DEXPREOPT_IMAGE_UNSTRIPPED_BUILT_INSTALLED_$(my_suffix)),$(my_boot_image_syms))
-
-# Generate the boot image module only if there is any file to install.
-ifneq (,$(strip $(my_copy_pairs)))
- my_first_pair := $(firstword $(my_copy_pairs))
- my_rest_pairs := $(wordlist 2,$(words $(my_copy_pairs)),$(my_copy_pairs))
-
- my_first_src := $(call word-colon,1,$(my_first_pair))
- my_first_dest := $(call word-colon,2,$(my_first_pair))
-
- my_installed := $(call copy-many-files,$(my_copy_pairs))
- my_unstripped_installed := $(call copy-many-files,$(my_unstripped_copy_pairs))
-
- my_symlinks := $(call symlink-vdex-files,$(my_vdex_copy_pairs))
-
- # We don't have a LOCAL_PATH for the auto-generated modules, so let it be the $(BUILD_SYSTEM).
- LOCAL_PATH := $(BUILD_SYSTEM)
- # Hack to let these pseudo-modules wrapped around Soong modules use LOCAL_SOONG_INSTALLED_MODULE.
- LOCAL_MODULE_MAKEFILE := $(SOONG_ANDROID_MK)
-
- include $(CLEAR_VARS)
- LOCAL_MODULE := dexpreopt_bootjar.$(my_suffix)
- LOCAL_PREBUILT_MODULE_FILE := $(my_first_src)
- LOCAL_MODULE_PATH := $(dir $(my_first_dest))
- LOCAL_MODULE_STEM := $(notdir $(my_first_dest))
- LOCAL_SOONG_INSTALL_PAIRS := $(my_copy_pairs)
- LOCAL_SOONG_INSTALL_SYMLINKS := $(my_symlinks)
- LOCAL_SOONG_INSTALLED_MODULE := $(my_first_dest)
- LOCAL_SOONG_LICENSE_METADATA := $(DEXPREOPT_IMAGE_LICENSE_METADATA_$(my_suffix))
- ifneq (,$(strip $(filter HOST_%,$(my_boot_image_arch))))
- LOCAL_IS_HOST_MODULE := true
- endif
- LOCAL_MODULE_CLASS := ETC
- include $(BUILD_PREBUILT)
- $(LOCAL_BUILT_MODULE): | $(my_unstripped_installed)
- # Installing boot.art causes all boot image bits to be installed.
- # Keep this old behavior in case anyone still needs it.
- $(LOCAL_INSTALLED_MODULE): $(wordlist 2,$(words $(my_installed)),$(my_installed)) $(my_symlinks)
- $(my_all_targets): $(my_installed) $(my_symlinks)
-
- my_boot_image_module := $(LOCAL_MODULE)
-endif # my_copy_pairs != empty
diff --git a/core/dumpconfig.mk b/core/dumpconfig.mk
index 640fe10..eb4c822 100644
--- a/core/dumpconfig.mk
+++ b/core/dumpconfig.mk
@@ -56,7 +56,7 @@
# Escape quotation marks for CSV, and wraps in quotation marks.
define escape-for-csv
-"$(subst ","",$1)"
+"$(subst ","",$(subst $(newline), ,$1))"
endef
# Args:
@@ -68,7 +68,7 @@
# Args:
# $(1): include stack
define dump-import-done
-$(eval $(file >> $(DUMPCONFIG_FILE),imported,$(strip $(1))))
+$(eval $(file >> $(DUMPCONFIG_FILE),imported,$(strip $(1)),$(filter-out $(1),$(MAKEFILE_LIST))))
endef
# Args:
diff --git a/core/envsetup.mk b/core/envsetup.mk
index c063f60..f82e861 100644
--- a/core/envsetup.mk
+++ b/core/envsetup.mk
@@ -417,6 +417,7 @@
HOST_OUT_NATIVE_TESTS := $(HOST_OUT)/nativetest64
HOST_OUT_COVERAGE := $(HOST_OUT)/coverage
HOST_OUT_TESTCASES := $(HOST_OUT)/testcases
+HOST_OUT_ETC := $(HOST_OUT)/etc
.KATI_READONLY := \
HOST_OUT_EXECUTABLES \
HOST_OUT_SHARED_LIBRARIES \
@@ -425,7 +426,8 @@
HOST_OUT_SDK_ADDON \
HOST_OUT_NATIVE_TESTS \
HOST_OUT_COVERAGE \
- HOST_OUT_TESTCASES
+ HOST_OUT_TESTCASES \
+ HOST_OUT_ETC
HOST_CROSS_OUT_EXECUTABLES := $(HOST_CROSS_OUT)/bin
HOST_CROSS_OUT_SHARED_LIBRARIES := $(HOST_CROSS_OUT)/lib
diff --git a/core/main.mk b/core/main.mk
index 27ba526..5bbe1b1 100644
--- a/core/main.mk
+++ b/core/main.mk
@@ -31,8 +31,7 @@
.KATI_READONLY := $(foreach n,$(SOONG_CONFIG_NAMESPACES),SOONG_CONFIG_$(n))
.KATI_READONLY := $(foreach n,$(SOONG_CONFIG_NAMESPACES),$(foreach k,$(SOONG_CONFIG_$(n)),SOONG_CONFIG_$(n)_$(k)))
-include $(SOONG_MAKEVARS_MK)
-
+include $(SOONG_OUT_DIR)/make_vars-$(TARGET_PRODUCT)$(COVERAGE_SUFFIX).mk
YACC :=$= $(BISON) -d
include $(BUILD_SYSTEM)/clang/config.mk
@@ -276,17 +275,23 @@
# Include all of the makefiles in the system
#
-subdir_makefiles := $(SOONG_OUT_DIR)/installs-$(TARGET_PRODUCT).mk $(SOONG_ANDROID_MK)
+subdir_makefiles := $(SOONG_OUT_DIR)/installs-$(TARGET_PRODUCT)$(COVERAGE_SUFFIX).mk $(SOONG_ANDROID_MK)
+
# Android.mk files are only used on Linux builds, Mac only supports Android.bp
ifeq ($(HOST_OS),linux)
subdir_makefiles += $(file <$(OUT_DIR)/.module_paths/Android.mk.list)
endif
-subdir_makefiles += $(SOONG_OUT_DIR)/late-$(TARGET_PRODUCT).mk
+
+subdir_makefiles += $(SOONG_OUT_DIR)/late-$(TARGET_PRODUCT)$(COVERAGE_SUFFIX).mk
+
subdir_makefiles_total := $(words int $(subdir_makefiles) post finish)
.KATI_READONLY := subdir_makefiles_total
$(foreach mk,$(subdir_makefiles),$(info [$(call inc_and_print,subdir_makefiles_inc)/$(subdir_makefiles_total)] including $(mk) ...)$(eval include $(mk)))
+# Build bootloader.img/radio.img, and unpack the partitions.
+include $(BUILD_SYSTEM)/tasks/tools/update_bootloader_radio_image.mk
+
# For an unbundled image, we can skip blueprint_tools because unbundled image
# aims to remove a large number framework projects from the manifest, the
# sources or dependencies for these tools may be missing from the tree.
@@ -295,6 +300,9 @@
checkbuild: blueprint_tests
endif
+# Create necessary directories and symlinks in the root filesystem
+include system/core/rootdir/create_root_structure.mk
+
endif # dont_bother
ifndef subdir_makefiles_total
@@ -679,12 +687,12 @@
# Scan all modules in general-tests, device-tests and other selected suites and
# flatten the shared library dependencies.
define update-host-shared-libs-deps-for-suites
-$(foreach suite,general-tests device-tests vts tvts art-host-tests host-unit-tests,\
+$(foreach suite,general-tests device-tests vts tvts art-host-tests host-unit-tests camera-hal-tests,\
$(foreach m,$(COMPATIBILITY.$(suite).MODULES),\
$(eval my_deps := $(call get-all-shared-libs-deps,$(m)))\
$(foreach dep,$(my_deps),\
$(foreach f,$(ALL_MODULES.$(dep).HOST_SHARED_LIBRARY_FILES),\
- $(if $(filter $(suite),device-tests general-tests art-host-tests host-unit-tests),\
+ $(if $(filter $(suite),device-tests general-tests art-host-tests host-unit-tests camera-hal-tests),\
$(eval my_testcases := $(HOST_OUT_TESTCASES)),\
$(eval my_testcases := $$(COMPATIBILITY_TESTCASES_OUT_$(suite))))\
$(eval target := $(my_testcases)/$(lastword $(subst /, ,$(dir $(f))))/$(notdir $(f)))\
@@ -1858,6 +1866,11 @@
filter_out_files += $(PRODUCT_OUT)/recovery/%
endif
+# userdata.img
+ifndef BUILDING_USERDATA_IMAGE
+filter_out_files += $(PRODUCT_OUT)/data/%
+endif
+
installed_files := $(sort $(filter-out $(filter_out_files),$(filter $(PRODUCT_OUT)/%,$(modules_to_install))))
else
installed_files := $(apps_only_installed_files)
@@ -1886,7 +1899,6 @@
$(eval _kernel_module_copy_files := $(sort $(filter %$(_path_on_device),$(KERNEL_MODULE_COPY_FILES)))) \
$(eval _is_build_prop := $(call is-build-prop,$f)) \
$(eval _is_notice_file := $(call is-notice-file,$f)) \
- $(eval _is_dexpreopt_image_profile := $(if $(filter %:/$(_path_on_device),$(DEXPREOPT_IMAGE_PROFILE_BUILT_INSTALLED)),Y)) \
$(eval _is_product_system_other_avbkey := $(if $(findstring $f,$(INSTALLED_PRODUCT_SYSTEM_OTHER_AVBKEY_TARGET)),Y)) \
$(eval _is_event_log_tags_file := $(if $(findstring $f,$(event_log_tags_file)),Y)) \
$(eval _is_system_other_odex_marker := $(if $(findstring $f,$(INSTALLED_SYSTEM_OTHER_ODEX_MARKER)),Y)) \
@@ -1896,7 +1908,7 @@
$(eval _is_partition_compat_symlink := $(if $(findstring $f,$(PARTITION_COMPAT_SYMLINKS)),Y)) \
$(eval _is_flags_file := $(if $(findstring $f, $(ALL_FLAGS_FILES)),Y)) \
$(eval _is_rootdir_symlink := $(if $(findstring $f, $(ALL_ROOTDIR_SYMLINKS)),Y)) \
- $(eval _is_platform_generated := $(_is_build_prop)$(_is_notice_file)$(_is_dexpreopt_image_profile)$(_is_product_system_other_avbkey)$(_is_event_log_tags_file)$(_is_system_other_odex_marker)$(_is_kernel_modules_blocklist)$(_is_fsverity_build_manifest_apk)$(_is_linker_config)$(_is_partition_compat_symlink)$(_is_flags_file)$(_is_rootdir_symlink)) \
+ $(eval _is_platform_generated := $(_is_build_prop)$(_is_notice_file)$(_is_product_system_other_avbkey)$(_is_event_log_tags_file)$(_is_system_other_odex_marker)$(_is_kernel_modules_blocklist)$(_is_fsverity_build_manifest_apk)$(_is_linker_config)$(_is_partition_compat_symlink)$(_is_flags_file)$(_is_rootdir_symlink)) \
$(eval _static_libs := $(if $(_is_soong_module),,$(ALL_INSTALLED_FILES.$f.STATIC_LIBRARIES))) \
$(eval _whole_static_libs := $(if $(_is_soong_module),,$(ALL_INSTALLED_FILES.$f.WHOLE_STATIC_LIBRARIES))) \
$(eval _license_text := $(if $(filter $(_build_output_path),$(ALL_NON_MODULES)),$(ALL_NON_MODULES.$(_build_output_path).NOTICES))) \
diff --git a/core/os_licensing.mk b/core/os_licensing.mk
index 1e1b7df..d15a3d0 100644
--- a/core/os_licensing.mk
+++ b/core/os_licensing.mk
@@ -17,13 +17,17 @@
$(eval $(call text-notice-rule,$(target_notice_file_txt),"System image",$(system_notice_file_message),$(SYSTEM_NOTICE_DEPS),$(SYSTEM_NOTICE_DEPS)))
+ifneq ($(USE_SOONG_DEFINED_SYSTEM_IMAGE),true)
$(installed_notice_html_or_xml_gz): $(target_notice_file_xml_gz)
$(copy-file-to-target)
endif
+endif
$(call declare-1p-target,$(target_notice_file_xml_gz))
+ifneq ($(USE_SOONG_DEFINED_SYSTEM_IMAGE),true)
$(call declare-1p-target,$(installed_notice_html_or_xml_gz))
endif
+endif
.PHONY: vendorlicense
vendorlicense: $(call corresponding-license-metadata, $(VENDOR_NOTICE_DEPS)) reportmissinglicenses
diff --git a/core/packaging/flags.mk b/core/packaging/flags.mk
index a77956b..ccb502c 100644
--- a/core/packaging/flags.mk
+++ b/core/packaging/flags.mk
@@ -18,7 +18,7 @@
#
# TODO: Should we do all of the images in $(IMAGES_TO_BUILD)?
-_FLAG_PARTITIONS := product system system_ext vendor
+_FLAG_PARTITIONS := product system vendor
# -----------------------------------------------------------------
@@ -28,7 +28,6 @@
# $(1): built aconfig flags file (out)
# $(2): installed aconfig flags file (out)
# $(3): the partition (in)
-# $(4): input aconfig files for the partition (in)
define generate-partition-aconfig-flag-file
$(eval $(strip $(1)): PRIVATE_OUT := $(strip $(1)))
$(eval $(strip $(1)): PRIVATE_IN := $(strip $(4)))
@@ -36,12 +35,14 @@
mkdir -p $$(dir $$(PRIVATE_OUT))
$$(if $$(PRIVATE_IN), \
$$(ACONFIG) dump --dedup --format protobuf --out $$(PRIVATE_OUT) \
- --filter container:$$(strip $(3)) $$(addprefix --cache ,$$(PRIVATE_IN)), \
+ --filter container:$(strip $(3)) \
+ $$(addprefix --cache ,$$(PRIVATE_IN)), \
echo -n > $$(PRIVATE_OUT) \
)
$(call copy-one-file, $(1), $(2))
endef
+
# Create a summary file of build flags for each partition
# $(1): built aconfig flags file (out)
# $(2): installed aconfig flags file (out)
@@ -59,16 +60,22 @@
$(call copy-one-file, $(1), $(2))
endef
-
$(foreach partition, $(_FLAG_PARTITIONS), \
$(eval aconfig_flag_summaries_protobuf.$(partition) := $(PRODUCT_OUT)/$(partition)/etc/aconfig_flags.pb) \
$(eval $(call generate-partition-aconfig-flag-file, \
- $(TARGET_OUT_FLAGS)/$(partition)/aconfig_flags.pb, \
- $(aconfig_flag_summaries_protobuf.$(partition)), \
- $(partition), \
- $(sort $(foreach m,$(call register-names-for-partition, $(partition)), \
+ $(TARGET_OUT_FLAGS)/$(partition)/aconfig_flags.pb, \
+ $(aconfig_flag_summaries_protobuf.$(partition)), \
+ $(partition), \
+ $(sort \
+ $(foreach m, $(call register-names-for-partition, $(partition)), \
$(ALL_MODULES.$(m).ACONFIG_FILES) \
- )), \
+ ) \
+ $(if $(filter system, $(partition)), \
+ $(foreach m, $(call register-names-for-partition, system_ext), \
+ $(ALL_MODULES.$(m).ACONFIG_FILES) \
+ ) \
+ ) \
+ ) \
)) \
)
@@ -90,42 +97,54 @@
# $(1): built aconfig flags storage package map file (out)
# $(2): built aconfig flags storage flag map file (out)
# $(3): built aconfig flags storage flag val file (out)
-# $(4): installed aconfig flags storage package map file (out)
-# $(5): installed aconfig flags storage flag map file (out)
-# $(6): installed aconfig flags storage flag value file (out)
-# $(7): input aconfig files for the partition (in)
-# $(8): partition name
+# $(4): built aconfig flags storage flag info file (out)
+# $(5): installed aconfig flags storage package map file (out)
+# $(6): installed aconfig flags storage flag map file (out)
+# $(7): installed aconfig flags storage flag value file (out)
+# $(8): installed aconfig flags storage flag info file (out)
+# $(9): input aconfig files for the partition (in)
+# $(10): partition name
define generate-partition-aconfig-storage-file
$(eval $(strip $(1)): PRIVATE_OUT := $(strip $(1)))
-$(eval $(strip $(1)): PRIVATE_IN := $(strip $(7)))
-$(strip $(1)): $(ACONFIG) $(strip $(7))
+$(eval $(strip $(1)): PRIVATE_IN := $(strip $(9)))
+$(strip $(1)): $(ACONFIG) $(strip $(9))
mkdir -p $$(dir $$(PRIVATE_OUT))
$$(if $$(PRIVATE_IN), \
- $$(ACONFIG) create-storage --container $(8) --file package_map --out $$(PRIVATE_OUT) \
+ $$(ACONFIG) create-storage --container $(10) --file package_map --out $$(PRIVATE_OUT) \
$$(addprefix --cache ,$$(PRIVATE_IN)), \
)
touch $$(PRIVATE_OUT)
$(eval $(strip $(2)): PRIVATE_OUT := $(strip $(2)))
-$(eval $(strip $(2)): PRIVATE_IN := $(strip $(7)))
-$(strip $(2)): $(ACONFIG) $(strip $(7))
+$(eval $(strip $(2)): PRIVATE_IN := $(strip $(9)))
+$(strip $(2)): $(ACONFIG) $(strip $(9))
mkdir -p $$(dir $$(PRIVATE_OUT))
$$(if $$(PRIVATE_IN), \
- $$(ACONFIG) create-storage --container $(8) --file flag_map --out $$(PRIVATE_OUT) \
+ $$(ACONFIG) create-storage --container $(10) --file flag_map --out $$(PRIVATE_OUT) \
$$(addprefix --cache ,$$(PRIVATE_IN)), \
)
touch $$(PRIVATE_OUT)
$(eval $(strip $(3)): PRIVATE_OUT := $(strip $(3)))
-$(eval $(strip $(3)): PRIVATE_IN := $(strip $(7)))
-$(strip $(3)): $(ACONFIG) $(strip $(7))
+$(eval $(strip $(3)): PRIVATE_IN := $(strip $(9)))
+$(strip $(3)): $(ACONFIG) $(strip $(9))
mkdir -p $$(dir $$(PRIVATE_OUT))
$$(if $$(PRIVATE_IN), \
- $$(ACONFIG) create-storage --container $(8) --file flag_val --out $$(PRIVATE_OUT) \
+ $$(ACONFIG) create-storage --container $(10) --file flag_val --out $$(PRIVATE_OUT) \
$$(addprefix --cache ,$$(PRIVATE_IN)), \
)
touch $$(PRIVATE_OUT)
-$(call copy-one-file, $(strip $(1)), $(4))
-$(call copy-one-file, $(strip $(2)), $(5))
-$(call copy-one-file, $(strip $(3)), $(6))
+$(eval $(strip $(4)): PRIVATE_OUT := $(strip $(4)))
+$(eval $(strip $(4)): PRIVATE_IN := $(strip $(9)))
+$(strip $(4)): $(ACONFIG) $(strip $(9))
+ mkdir -p $$(dir $$(PRIVATE_OUT))
+ $$(if $$(PRIVATE_IN), \
+ $$(ACONFIG) create-storage --container $(10) --file flag_info --out $$(PRIVATE_OUT) \
+ $$(addprefix --cache ,$$(PRIVATE_IN)), \
+ )
+ touch $$(PRIVATE_OUT)
+$(call copy-one-file, $(strip $(1)), $(5))
+$(call copy-one-file, $(strip $(2)), $(6))
+$(call copy-one-file, $(strip $(3)), $(7))
+$(call copy-one-file, $(strip $(4)), $(8))
endef
ifeq ($(RELEASE_CREATE_ACONFIG_STORAGE_FILE),true)
@@ -133,13 +152,16 @@
$(eval aconfig_storage_package_map.$(partition) := $(PRODUCT_OUT)/$(partition)/etc/aconfig/package.map) \
$(eval aconfig_storage_flag_map.$(partition) := $(PRODUCT_OUT)/$(partition)/etc/aconfig/flag.map) \
$(eval aconfig_storage_flag_val.$(partition) := $(PRODUCT_OUT)/$(partition)/etc/aconfig/flag.val) \
+ $(eval aconfig_storage_flag_info.$(partition) := $(PRODUCT_OUT)/$(partition)/etc/aconfig/flag.info) \
$(eval $(call generate-partition-aconfig-storage-file, \
$(TARGET_OUT_FLAGS)/$(partition)/package.map, \
$(TARGET_OUT_FLAGS)/$(partition)/flag.map, \
$(TARGET_OUT_FLAGS)/$(partition)/flag.val, \
+ $(TARGET_OUT_FLAGS)/$(partition)/flag.info, \
$(aconfig_storage_package_map.$(partition)), \
$(aconfig_storage_flag_map.$(partition)), \
$(aconfig_storage_flag_val.$(partition)), \
+ $(aconfig_storage_flag_info.$(partition)), \
$(aconfig_flag_summaries_protobuf.$(partition)), \
$(partition), \
)) \
@@ -155,6 +177,7 @@
$(aconfig_storage_package_map.$(partition)) \
$(aconfig_storage_flag_map.$(partition)) \
$(aconfig_storage_flag_val.$(partition)) \
+ $(aconfig_storage_flag_info.$(partition)) \
))
ALL_DEFAULT_INSTALLED_MODULES += $(required_flags_files)
@@ -174,5 +197,5 @@
$(eval aconfig_storage_package_map.$(partition):=) \
$(eval aconfig_storage_flag_map.$(partition):=) \
$(eval aconfig_storage_flag_val.$(partition):=) \
+ $(eval aconfig_storage_flag_info.$(partition):=) \
)
-
diff --git a/core/product.mk b/core/product.mk
index 8d86d92..93a656d 100644
--- a/core/product.mk
+++ b/core/product.mk
@@ -390,20 +390,6 @@
# If set, Java module in product partition cannot use hidden APIs.
_product_single_value_vars += PRODUCT_ENFORCE_PRODUCT_PARTITION_INTERFACE
-# If set, only java_sdk_library can be used at inter-partition dependency.
-# Note: Build error if BOARD_VNDK_VERSION is not set while
-# PRODUCT_ENFORCE_INTER_PARTITION_JAVA_SDK_LIBRARY is true, because
-# PRODUCT_ENFORCE_INTER_PARTITION_JAVA_SDK_LIBRARY has no meaning if
-# BOARD_VNDK_VERSION is not set.
-# Note: When PRODUCT_ENFORCE_PRODUCT_PARTITION_INTERFACE is not set, there are
-# no restrictions at dependency between system and product partition.
-_product_single_value_vars += PRODUCT_ENFORCE_INTER_PARTITION_JAVA_SDK_LIBRARY
-
-# Allowlist for PRODUCT_ENFORCE_INTER_PARTITION_JAVA_SDK_LIBRARY option.
-# Listed modules are allowed at inter-partition dependency even if it isn't
-# a java_sdk_library module.
-_product_list_vars += PRODUCT_INTER_PARTITION_JAVA_LIBRARY_ALLOWLIST
-
# Install a copy of the debug policy to the system_ext partition, and allow
# init-second-stage to load debug policy from system_ext.
# This option is only meant to be set by compliance GSI targets.
@@ -493,9 +479,16 @@
# by this flag.
_product_single_value_vars += PRODUCT_NOT_DEBUGGABLE_IN_USERDEBUG
+# If set, the default value of the versionName of apps will include the build number.
+_product_single_value_vars += PRODUCT_BUILD_APPS_WITH_BUILD_NUMBER
+
# If set, build would generate system image from Soong-defined module.
_product_single_value_vars += PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE
+# List of stub libraries specific to the product that are already present in the system image and
+# should be included in the system_linker_config.
+_product_list_vars += PRODUCT_EXTRA_STUB_LIBRARIES
+
.KATI_READONLY := _product_single_value_vars _product_list_vars
_product_var_list :=$= $(_product_single_value_vars) $(_product_list_vars)
diff --git a/core/proguard.flags b/core/proguard.flags
index aa406b9..5148e56 100644
--- a/core/proguard.flags
+++ b/core/proguard.flags
@@ -38,6 +38,17 @@
@com.android.internal.annotations.KeepForWeakReference <fields>;
}
+# Needed to ensure callback field references are kept in their respective
+# owning classes when the downstream callback registrars only store weak refs.
+-if @com.android.internal.annotations.WeaklyReferencedCallback class *
+-keepclassmembers,allowaccessmodification class * {
+ <1> *;
+}
+-if class * extends @com.android.internal.annotations.WeaklyReferencedCallback **
+-keepclassmembers,allowaccessmodification class * {
+ <1> *;
+}
+
# Understand the common @Keep annotation from various Android packages:
# * android.support.annotation
# * androidx.annotation
diff --git a/core/soong_android_app_set.mk b/core/soong_android_app_set.mk
index ec3d8c8..d97980d 100644
--- a/core/soong_android_app_set.mk
+++ b/core/soong_android_app_set.mk
@@ -9,10 +9,6 @@
LOCAL_BUILT_MODULE_STEM := package.apk
LOCAL_INSTALLED_MODULE_STEM := $(notdir $(LOCAL_PREBUILT_MODULE_FILE))
-# Use the Soong output as the checkbuild target instead of LOCAL_BUILT_MODULE
-# to avoid checkbuilds making an extra copy of every module.
-LOCAL_CHECKED_MODULE := $(LOCAL_PREBUILT_MODULE_FILE)
-
#######################################
include $(BUILD_SYSTEM)/base_rules.mk
#######################################
diff --git a/core/soong_app_prebuilt.mk b/core/soong_app_prebuilt.mk
index 3aa244c..df1cf2d 100644
--- a/core/soong_app_prebuilt.mk
+++ b/core/soong_app_prebuilt.mk
@@ -29,16 +29,6 @@
full_classes_header_jar := $(intermediates.COMMON)/classes-header.jar
-# Use the Soong output as the checkbuild target instead of LOCAL_BUILT_MODULE
-# to avoid checkbuilds making an extra copy of every module.
-LOCAL_CHECKED_MODULE := $(LOCAL_PREBUILT_MODULE_FILE)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_CLASSES_JAR)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_HEADER_JAR)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_FULL_MANIFEST_FILE)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_DEXPREOPT_CONFIG)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_RESOURCE_EXPORT_PACKAGE)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_DEX_JAR)
-
#######################################
include $(BUILD_SYSTEM)/base_rules.mk
#######################################
diff --git a/core/soong_cc_rust_prebuilt.mk b/core/soong_cc_rust_prebuilt.mk
index a1c6478..da60832 100644
--- a/core/soong_cc_rust_prebuilt.mk
+++ b/core/soong_cc_rust_prebuilt.mk
@@ -38,10 +38,6 @@
endif
endif
-# Use the Soong output as the checkbuild target instead of LOCAL_BUILT_MODULE
-# to avoid checkbuilds making an extra copy of every module.
-LOCAL_CHECKED_MODULE := $(LOCAL_PREBUILT_MODULE_FILE)
-
my_check_same_vndk_variants :=
same_vndk_variants_stamp :=
ifeq ($(LOCAL_CHECK_SAME_VNDK_VARIANTS),true)
@@ -61,7 +57,7 @@
# Note that because `checkbuild` doesn't check LOCAL_BUILT_MODULE for soong-built modules adding
# the timestamp to LOCAL_BUILT_MODULE isn't enough. It is skipped when the vendor variant
# isn't used at all and it may break in the downstream trees.
- LOCAL_ADDITIONAL_CHECKED_MODULE := $(same_vndk_variants_stamp)
+ LOCAL_ADDITIONAL_CHECKED_MODULE += $(same_vndk_variants_stamp)
endif
#######################################
diff --git a/core/soong_config.mk b/core/soong_config.mk
index 5769c06..0421e19 100644
--- a/core/soong_config.mk
+++ b/core/soong_config.mk
@@ -1,5 +1,5 @@
-SOONG_MAKEVARS_MK := $(SOONG_OUT_DIR)/make_vars-$(TARGET_PRODUCT).mk
-SOONG_ANDROID_MK := $(SOONG_OUT_DIR)/Android-$(TARGET_PRODUCT).mk
+SOONG_MAKEVARS_MK := $(SOONG_OUT_DIR)/make_vars-$(TARGET_PRODUCT)$(COVERAGE_SUFFIX).mk
+SOONG_ANDROID_MK := $(SOONG_OUT_DIR)/Android-$(TARGET_PRODUCT)$(COVERAGE_SUFFIX).mk
include $(BUILD_SYSTEM)/art_config.mk
include $(BUILD_SYSTEM)/dex_preopt_config.mk
@@ -26,7 +26,7 @@
$(shell mkdir -p $(dir $(SOONG_VARIABLES)))
$(call json_start)
-$(call add_json_str, Make_suffix, -$(TARGET_PRODUCT))
+$(call add_json_str, Make_suffix, -$(TARGET_PRODUCT)$(COVERAGE_SUFFIX))
$(call add_json_str, BuildId, $(BUILD_ID))
$(call add_json_str, BuildFingerprintFile, build_fingerprint.txt)
@@ -236,6 +236,12 @@
$(call add_json_list, TargetFSConfigGen, $(TARGET_FS_CONFIG_GEN))
+# Although USE_SOONG_DEFINED_SYSTEM_IMAGE determines whether to use the system image specified by
+# PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE, PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE is still used to compare
+# installed files between make and soong, regardless of the USE_SOONG_DEFINED_SYSTEM_IMAGE setting.
+$(call add_json_bool, UseSoongSystemImage, $(filter true,$(USE_SOONG_DEFINED_SYSTEM_IMAGE)))
+$(call add_json_str, ProductSoongDefinedSystemImage, $(PRODUCT_SOONG_DEFINED_SYSTEM_IMAGE))
+
$(call add_json_map, VendorVars)
$(foreach namespace,$(sort $(SOONG_CONFIG_NAMESPACES)),\
$(call add_json_map, $(namespace))\
@@ -259,9 +265,6 @@
$(call add_json_bool, EnforceProductPartitionInterface, $(filter true,$(PRODUCT_ENFORCE_PRODUCT_PARTITION_INTERFACE)))
$(call add_json_str, DeviceCurrentApiLevelForVendorModules, $(BOARD_CURRENT_API_LEVEL_FOR_VENDOR_MODULES))
-$(call add_json_bool, EnforceInterPartitionJavaSdkLibrary, $(filter true,$(PRODUCT_ENFORCE_INTER_PARTITION_JAVA_SDK_LIBRARY)))
-$(call add_json_list, InterPartitionJavaLibraryAllowList, $(PRODUCT_INTER_PARTITION_JAVA_LIBRARY_ALLOWLIST))
-
$(call add_json_bool, CompressedApex, $(filter true,$(PRODUCT_COMPRESSED_APEX)))
ifndef APEX_BUILD_FOR_PRE_S_DEVICES
@@ -286,7 +289,6 @@
$(call add_json_bool, GenruleSandboxing, $(if $(GENRULE_SANDBOXING),$(filter true,$(GENRULE_SANDBOXING)),$(if $(filter true,$(BUILD_BROKEN_GENRULE_SANDBOXING)),,true)))
$(call add_json_bool, BuildBrokenEnforceSyspropOwner, $(filter true,$(BUILD_BROKEN_ENFORCE_SYSPROP_OWNER)))
$(call add_json_bool, BuildBrokenTrebleSyspropNeverallow, $(filter true,$(BUILD_BROKEN_TREBLE_SYSPROP_NEVERALLOW)))
-$(call add_json_bool, BuildBrokenUsesSoongPython2Modules, $(filter true,$(BUILD_BROKEN_USES_SOONG_PYTHON2_MODULES)))
$(call add_json_bool, BuildBrokenVendorPropertyNamespace, $(filter true,$(BUILD_BROKEN_VENDOR_PROPERTY_NAMESPACE)))
$(call add_json_bool, BuildBrokenIncorrectPartitionImages, $(filter true,$(BUILD_BROKEN_INCORRECT_PARTITION_IMAGES)))
$(call add_json_list, BuildBrokenInputDirModules, $(BUILD_BROKEN_INPUT_DIR_MODULES))
@@ -344,6 +346,7 @@
$(call add_json_list, SystemPropFiles, $(TARGET_SYSTEM_PROP))
$(call add_json_list, SystemExtPropFiles, $(TARGET_SYSTEM_EXT_PROP))
$(call add_json_list, ProductPropFiles, $(TARGET_PRODUCT_PROP))
+$(call add_json_list, OdmPropFiles, $(TARGET_ODM_PROP))
# Do not set ArtTargetIncludeDebugBuild into any value if PRODUCT_ART_TARGET_INCLUDE_DEBUG_BUILD is not set,
# to have the same behavior from runtime_libart.mk.
@@ -361,6 +364,62 @@
$(call add_json_list, BoardAvbSystemAddHashtreeFooterArgs, $(BOARD_AVB_SYSTEM_ADD_HASHTREE_FOOTER_ARGS))
$(call add_json_bool, BoardAvbEnable, $(filter true,$(BOARD_AVB_ENABLE)))
+$(call add_json_map, PartitionVarsForSoongMigrationOnlyDoNotUse)
+ $(call add_json_str, ProductDirectory, $(dir $(INTERNAL_PRODUCT)))
+
+ $(call add_json_map,PartitionQualifiedVariables)
+ $(foreach image_type,SYSTEM VENDOR CACHE USERDATA PRODUCT SYSTEM_EXT OEM ODM VENDOR_DLKM ODM_DLKM SYSTEM_DLKM, \
+ $(call add_json_map,$(call to-lower,$(image_type))) \
+ $(call add_json_bool, BuildingImage, $(filter true,$(BUILDING_$(image_type)_IMAGE))) \
+ $(call add_json_str, BoardErofsCompressor, $(BOARD_$(image_type)IMAGE_EROFS_COMPRESSOR)) \
+ $(call add_json_str, BoardErofsCompressHints, $(BOARD_$(image_type)IMAGE_EROFS_COMPRESS_HINTS)) \
+ $(call add_json_str, BoardErofsPclusterSize, $(BOARD_$(image_type)IMAGE_EROFS_PCLUSTER_SIZE)) \
+ $(call add_json_str, BoardExtfsInodeCount, $(BOARD_$(image_type)IMAGE_EXTFS_INODE_COUNT)) \
+ $(call add_json_str, BoardExtfsRsvPct, $(BOARD_$(image_type)IMAGE_EXTFS_RSV_PCT)) \
+ $(call add_json_str, BoardF2fsSloadCompressFlags, $(BOARD_$(image_type)IMAGE_F2FS_SLOAD_COMPRESS_FLAGS)) \
+ $(call add_json_str, BoardFileSystemCompress, $(BOARD_$(image_type)IMAGE_FILE_SYSTEM_COMPRESS)) \
+ $(call add_json_str, BoardFileSystemType, $(BOARD_$(image_type)IMAGE_FILE_SYSTEM_TYPE)) \
+ $(call add_json_str, BoardJournalSize, $(BOARD_$(image_type)IMAGE_JOURNAL_SIZE)) \
+ $(call add_json_str, BoardPartitionReservedSize, $(BOARD_$(image_type)IMAGE_PARTITION_RESERVED_SIZE)) \
+ $(call add_json_str, BoardPartitionSize, $(BOARD_$(image_type)IMAGE_PARTITION_SIZE)) \
+ $(call add_json_str, BoardSquashfsBlockSize, $(BOARD_$(image_type)IMAGE_SQUASHFS_BLOCK_SIZE)) \
+ $(call add_json_str, BoardSquashfsCompressor, $(BOARD_$(image_type)IMAGE_SQUASHFS_COMPRESSOR)) \
+ $(call add_json_str, BoardSquashfsCompressorOpt, $(BOARD_$(image_type)IMAGE_SQUASHFS_COMPRESSOR_OPT)) \
+ $(call add_json_str, BoardSquashfsDisable4kAlign, $(BOARD_$(image_type)IMAGE_SQUASHFS_DISABLE_4K_ALIGN)) \
+ $(call add_json_str, ProductBaseFsPath, $(PRODUCT_$(image_type)_BASE_FS_PATH)) \
+ $(call add_json_str, ProductHeadroom, $(PRODUCT_$(image_type)_HEADROOM)) \
+ $(call add_json_str, ProductVerityPartition, $(PRODUCT_$(image_type)_VERITY_PARTITION)) \
+ $(call end_json_map) \
+ )
+ $(call end_json_map)
+
+ $(call add_json_bool, TargetUserimagesUseExt2, $(filter true,$(TARGET_USERIMAGES_USE_EXT2)))
+ $(call add_json_bool, TargetUserimagesUseExt3, $(filter true,$(TARGET_USERIMAGES_USE_EXT3)))
+ $(call add_json_bool, TargetUserimagesUseExt4, $(filter true,$(TARGET_USERIMAGES_USE_EXT4)))
+
+ $(call add_json_bool, TargetUserimagesSparseExtDisabled, $(filter true,$(TARGET_USERIMAGES_SPARSE_EXT_DISABLED)))
+ $(call add_json_bool, TargetUserimagesSparseErofsDisabled, $(filter true,$(TARGET_USERIMAGES_SPARSE_EROFS_DISABLED)))
+ $(call add_json_bool, TargetUserimagesSparseSquashfsDisabled, $(filter true,$(TARGET_USERIMAGES_SPARSE_SQUASHFS_DISABLED)))
+ $(call add_json_bool, TargetUserimagesSparseF2fsDisabled, $(filter true,$(TARGET_USERIMAGES_SPARSE_F2FS_DISABLED)))
+
+ $(call add_json_str, BoardErofsCompressor, $(BOARD_EROFS_COMPRESSOR))
+ $(call add_json_str, BoardErofsCompressorHints, $(BOARD_EROFS_COMPRESS_HINTS))
+ $(call add_json_str, BoardErofsPclusterSize, $(BOARD_EROFS_PCLUSTER_SIZE))
+ $(call add_json_str, BoardErofsShareDupBlocks, $(BOARD_EROFS_SHARE_DUP_BLOCKS))
+ $(call add_json_str, BoardErofsUseLegacyCompression, $(BOARD_EROFS_USE_LEGACY_COMPRESSION))
+ $(call add_json_str, BoardExt4ShareDupBlocks, $(BOARD_EXT4_SHARE_DUP_BLOCKS))
+ $(call add_json_str, BoardFlashLogicalBlockSize, $(BOARD_FLASH_LOGICAL_BLOCK_SIZE))
+ $(call add_json_str, BoardFlashEraseBlockSize, $(BOARD_FLASH_ERASE_BLOCK_SIZE))
+
+ $(call add_json_bool, BoardUsesRecoveryAsBoot, $(filter true,$(BOARD_USES_RECOVERY_AS_BOOT)))
+ $(call add_json_bool, ProductUseDynamicPartitionSize, $(filter true,$(PRODUCT_USE_DYNAMIC_PARTITION_SIZE)))
+ $(call add_json_bool, CopyImagesForTargetFilesZip, $(filter true,$(COPY_IMAGES_FOR_TARGET_FILES_ZIP)))
+
+ $(call add_json_list, ProductPackages, $(PRODUCT_PACKAGES))
+ $(call add_json_list, ProductPackagesDebug, $(PRODUCT_PACKAGES_DEBUG))
+
+$(call end_json_map)
+
$(call json_end)
$(file >$(SOONG_VARIABLES).tmp,$(json_contents))
diff --git a/core/soong_java_prebuilt.mk b/core/soong_java_prebuilt.mk
index f74bb6d..8c3882f 100644
--- a/core/soong_java_prebuilt.mk
+++ b/core/soong_java_prebuilt.mk
@@ -21,19 +21,6 @@
full_classes_header_jar := $(intermediates.COMMON)/classes-header.jar
common_javalib.jar := $(intermediates.COMMON)/javalib.jar
-ifdef LOCAL_SOONG_AAR
- LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_AAR)
-endif
-
-# Use the Soong output as the checkbuild target instead of LOCAL_BUILT_MODULE
-# to avoid checkbuilds making an extra copy of every module.
-LOCAL_CHECKED_MODULE := $(LOCAL_PREBUILT_MODULE_FILE)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_HEADER_JAR)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_FULL_MANIFEST_FILE)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_DEXPREOPT_CONFIG)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_RESOURCE_EXPORT_PACKAGE)
-LOCAL_ADDITIONAL_CHECKED_MODULE += $(LOCAL_SOONG_DEX_JAR)
-
#######################################
include $(BUILD_SYSTEM)/base_rules.mk
#######################################
diff --git a/core/tasks/device-tests.mk b/core/tasks/device-tests.mk
index 5850c4e..6164c2e 100644
--- a/core/tasks/device-tests.mk
+++ b/core/tasks/device-tests.mk
@@ -14,6 +14,7 @@
.PHONY: device-tests
+.PHONY: device-tests-host-shared-libs
device-tests-zip := $(PRODUCT_OUT)/device-tests.zip
# Create an artifact to include a list of test config files in device-tests.
@@ -23,37 +24,45 @@
my_host_shared_lib_for_device_tests := $(call copy-many-files,$(COMPATIBILITY.device-tests.HOST_SHARED_LIBRARY.FILES))
device_tests_host_shared_libs_zip := $(PRODUCT_OUT)/device-tests_host-shared-libs.zip
-$(device-tests-zip) : .KATI_IMPLICIT_OUTPUTS := $(device-tests-list-zip) $(device-tests-configs-zip) $(device_tests_host_shared_libs_zip)
+$(device-tests-zip) : .KATI_IMPLICIT_OUTPUTS := $(device-tests-list-zip) $(device-tests-configs-zip)
$(device-tests-zip) : PRIVATE_device_tests_list := $(PRODUCT_OUT)/device-tests_list
$(device-tests-zip) : PRIVATE_HOST_SHARED_LIBS := $(my_host_shared_lib_for_device_tests)
-$(device-tests-zip) : PRIVATE_device_host_shared_libs_zip := $(device_tests_host_shared_libs_zip)
$(device-tests-zip) : $(COMPATIBILITY.device-tests.FILES) $(COMPATIBILITY.device-tests.SOONG_INSTALLED_COMPATIBILITY_SUPPORT_FILES) $(my_host_shared_lib_for_device_tests) $(SOONG_ZIP)
- rm -f $@-shared-libs.list
echo $(sort $(COMPATIBILITY.device-tests.FILES) $(COMPATIBILITY.device-tests.SOONG_INSTALLED_COMPATIBILITY_SUPPORT_FILES)) | tr " " "\n" > $@.list
grep $(HOST_OUT_TESTCASES) $@.list > $@-host.list || true
grep -e .*\\.config$$ $@-host.list > $@-host-test-configs.list || true
$(hide) for shared_lib in $(PRIVATE_HOST_SHARED_LIBS); do \
echo $$shared_lib >> $@-host.list; \
- echo $$shared_lib >> $@-shared-libs.list; \
done
- grep $(HOST_OUT_TESTCASES) $@-shared-libs.list > $@-host-shared-libs.list || true
grep $(TARGET_OUT_TESTCASES) $@.list > $@-target.list || true
grep -e .*\\.config$$ $@-target.list > $@-target-test-configs.list || true
$(hide) $(SOONG_ZIP) -d -o $@ -P host -C $(HOST_OUT) -l $@-host.list -P target -C $(PRODUCT_OUT) -l $@-target.list -sha256
$(hide) $(SOONG_ZIP) -d -o $(device-tests-configs-zip) \
-P host -C $(HOST_OUT) -l $@-host-test-configs.list \
-P target -C $(PRODUCT_OUT) -l $@-target-test-configs.list
- $(SOONG_ZIP) -d -o $(PRIVATE_device_host_shared_libs_zip) \
- -P host -C $(HOST_OUT) -l $@-host-shared-libs.list
rm -f $(PRIVATE_device_tests_list)
$(hide) grep -e .*\\.config$$ $@-host.list | sed s%$(HOST_OUT)%host%g > $(PRIVATE_device_tests_list)
$(hide) grep -e .*\\.config$$ $@-target.list | sed s%$(PRODUCT_OUT)%target%g >> $(PRIVATE_device_tests_list)
$(hide) $(SOONG_ZIP) -d -o $(device-tests-list-zip) -C $(dir $@) -f $(PRIVATE_device_tests_list)
rm -f $@.list $@-host.list $@-target.list $@-host-test-configs.list $@-target-test-configs.list \
- $@-shared-libs.list $@-host-shared-libs.list $(PRIVATE_device_tests_list)
+ $(PRIVATE_device_tests_list)
+
+$(device_tests_host_shared_libs_zip) : PRIVATE_device_host_shared_libs_zip := $(device_tests_host_shared_libs_zip)
+$(device_tests_host_shared_libs_zip) : PRIVATE_HOST_SHARED_LIBS := $(my_host_shared_lib_for_device_tests)
+$(device_tests_host_shared_libs_zip) : $(my_host_shared_lib_for_device_tests) $(SOONG_ZIP)
+ rm -f $@-shared-libs.list
+ $(hide) for shared_lib in $(PRIVATE_HOST_SHARED_LIBS); do \
+ echo $$shared_lib >> $@-shared-libs.list; \
+ done
+ grep $(HOST_OUT_TESTCASES) $@-shared-libs.list > $@-host-shared-libs.list || true
+ $(SOONG_ZIP) -d -o $(PRIVATE_device_host_shared_libs_zip) \
+ -P host -C $(HOST_OUT) -l $@-host-shared-libs.list
device-tests: $(device-tests-zip)
+device-tests-host-shared-libs: $(device_tests_host_shared_libs_zip)
+
$(call dist-for-goals, device-tests, $(device-tests-zip) $(device-tests-list-zip) $(device-tests-configs-zip) $(device_tests_host_shared_libs_zip))
+$(call dist-for-goals, device-tests-host-shared-libs, $(device_tests_host_shared_libs_zip))
$(call declare-1p-container,$(device-tests-zip),)
$(call declare-container-license-deps,$(device-tests-zip),$(COMPATIBILITY.device-tests.FILES) $(my_host_shared_lib_for_device_tests),$(PRODUCT_OUT)/:/)
diff --git a/core/tasks/meta-lic.mk b/core/tasks/meta-lic.mk
index 24adfc8..620b1e2 100644
--- a/core/tasks/meta-lic.mk
+++ b/core/tasks/meta-lic.mk
@@ -225,3 +225,6 @@
# Moved here from hardware/libhardware_legacy/Android.mk
$(eval $(call declare-1p-copy-files,hardware/libhardware_legacy,))
+
+# Moved here from system/core/rootdir/Android.mk
+$(eval $(call declare-1p-copy-files,system/core/rootdir,))
diff --git a/core/tasks/mke2fs-dist.mk b/core/tasks/mke2fs-dist.mk
new file mode 100644
index 0000000..3540c1f
--- /dev/null
+++ b/core/tasks/mke2fs-dist.mk
@@ -0,0 +1,22 @@
+# Copyright (C) 2024 Google Inc.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+# TODO: After Soong's recovery partition variation can be set to selectable
+# and the meta_lic file duplication issue is resolved, move it to the
+# dist section of the corresponding module's Android.bp.
+my_dist_files := $(HOST_OUT_EXECUTABLES)/mke2fs
+my_dist_files += $(HOST_OUT_EXECUTABLES)/make_f2fs
+my_dist_files += $(HOST_OUT_EXECUTABLES)/make_f2fs_casefold
+$(call dist-for-goals,dist_files sdk,$(my_dist_files))
+my_dist_files :=
diff --git a/core/tasks/module-info.mk b/core/tasks/module-info.mk
index 7593668..0ca27d8 100644
--- a/core/tasks/module-info.mk
+++ b/core/tasks/module-info.mk
@@ -13,7 +13,7 @@
$(if $(strip $(2)),'$(COMMA)$(strip $(1)): "$(strip $(2))"')
endef
-SOONG_MODULE_INFO := $(SOONG_OUT_DIR)/module-info-$(TARGET_PRODUCT).json
+SOONG_MODULE_INFO := $(SOONG_OUT_DIR)/module-info-$(TARGET_PRODUCT)${COVERAGE_SUFFIX}.json
$(MODULE_INFO_JSON): PRIVATE_SOONG_MODULE_INFO := $(SOONG_MODULE_INFO)
$(MODULE_INFO_JSON): PRIVATE_MERGE_JSON_OBJECTS := $(HOST_OUT_EXECUTABLES)/merge_module_info_json
diff --git a/core/tasks/sts-sdk.mk b/core/tasks/sts-sdk.mk
index b8ce5bf..4abbc29 100644
--- a/core/tasks/sts-sdk.mk
+++ b/core/tasks/sts-sdk.mk
@@ -28,8 +28,7 @@
rm -f $@ $(STS_SDK_ZIP)_filtered
$(ZIP2ZIP) -i $(STS_SDK_ZIP) -o $(STS_SDK_ZIP)_filtered \
-x android-sts-sdk/tools/sts-tradefed-tests.jar \
- 'android-sts-sdk/tools/*:plugin/src/main/resources/sts-tradefed-tools/' \
- 'android-sts-sdk/jdk/**/*:plugin/src/main/resources/jdk/'
+ 'android-sts-sdk/tools/*:sts-sdk/src/main/resources/sts-tradefed-tools/'
$(MERGE_ZIPS) $@ $(STS_SDK_ZIP)_filtered $(STS_SDK_PLUGIN_SKEL)
rm -f $(STS_SDK_ZIP)_filtered
diff --git a/core/tasks/tools/compatibility.mk b/core/tasks/tools/compatibility.mk
index 86c23f8..9189c2d 100644
--- a/core/tasks/tools/compatibility.mk
+++ b/core/tasks/tools/compatibility.mk
@@ -29,7 +29,9 @@
special_mts_test_suites :=
special_mts_test_suites += mcts
special_mts_test_suites += $(mts_modules)
-ifneq ($(filter $(special_mts_test_suites),$(subst -, ,$(test_suite_name))),)
+ifneq ($(filter $(special_mts_test_suites),$(patsubst mcts-%,%,$(test_suite_name))),)
+ test_suite_subdir := android-mts
+else ifneq ($(filter $(special_mts_test_suites),$(patsubst mts-%,%,$(test_suite_name))),)
test_suite_subdir := android-mts
else
test_suite_subdir := android-$(test_suite_name)
diff --git a/core/tasks/tools/update_bootloader_radio_image.mk b/core/tasks/tools/update_bootloader_radio_image.mk
new file mode 100644
index 0000000..0ebf247
--- /dev/null
+++ b/core/tasks/tools/update_bootloader_radio_image.mk
@@ -0,0 +1,17 @@
+# Copyright (C) 2024 The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http:#www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+ifeq ($(USES_DEVICE_GOOGLE_ZUMA),true)
+ -include vendor/google_devices/zuma/prebuilts/misc_bins/update_bootloader_radio_image.mk
+endif
diff --git a/core/tasks/tradefed-tests-list.mk b/core/tasks/tradefed-tests-list.mk
index 61bf136..47c360d 100644
--- a/core/tasks/tradefed-tests-list.mk
+++ b/core/tasks/tradefed-tests-list.mk
@@ -15,6 +15,11 @@
# List all TradeFed tests from COMPATIBILITY.tradefed_tests_dir
.PHONY: tradefed-tests-list
+COMPATIBILITY.tradefed_tests_dir := \
+ $(COMPATIBILITY.tradefed_tests_dir) \
+ tools/tradefederation/core/res/config \
+ tools/tradefederation/core/javatests/res/config
+
tradefed_tests :=
$(foreach dir, $(COMPATIBILITY.tradefed_tests_dir), \
$(eval tradefed_tests += $(shell find $(dir) -type f -name "*.xml")))
diff --git a/envsetup.sh b/envsetup.sh
index 06dadd3..3fed5ae 100644
--- a/envsetup.sh
+++ b/envsetup.sh
@@ -442,6 +442,7 @@
function lunch()
{
local answer
+ setup_cog_env_if_needed
if [[ $# -gt 1 ]]; then
echo "usage: lunch [target]" >&2
@@ -1079,10 +1080,7 @@
done
done
- if [[ "${PWD}" == /google/cog/* ]]; then
- f="build/make/cogsetup.sh"
- echo "including $f"; . "$T/$f"
- fi
+ setup_cog_env_if_needed
}
function showcommands() {
diff --git a/shell_utils.sh b/shell_utils.sh
index 86f3f49..c4a6756 100644
--- a/shell_utils.sh
+++ b/shell_utils.sh
@@ -63,6 +63,70 @@
}
fi
+# This function sets up the build environment to be appropriate for Cog.
+function setup_cog_env_if_needed() {
+ local top=$(gettop)
+
+ # return early if not in a cog workspace
+ if [[ ! "$top" =~ ^/google/cog ]]; then
+ return 0
+ fi
+
+ setup_cog_symlink
+
+ export ANDROID_BUILD_ENVIRONMENT_CONFIG="googler-cog"
+
+ # Running repo command within Cog workspaces is not supported, so override
+ # it with this function. If the user is running repo within a Cog workspace,
+ # we'll fail with an error, otherwise, we run the original repo command with
+ # the given args.
+ if ! ORIG_REPO_PATH=`which repo`; then
+ return 0
+ fi
+ function repo {
+ if [[ "${PWD}" == /google/cog/* ]]; then
+ echo -e "\e[01;31mERROR:\e[0mrepo command is disallowed within Cog workspaces."
+ kill -INT $$ # exits the script without exiting the user's shell
+ fi
+ ${ORIG_REPO_PATH} "$@"
+ }
+}
+
+# creates a symlink for the out/ dir when inside a cog workspace.
+function setup_cog_symlink() {
+ local out_dir=$(getoutdir)
+ local top=$(gettop)
+
+ # return early if out dir is already a symlink
+ if [[ -L "$out_dir" ]]; then
+ return 0
+ fi
+
+ # return early if out dir is not in the workspace
+ if [[ ! "$out_dir" =~ ^$top/ ]]; then
+ return 0
+ fi
+
+ local link_destination="${HOME}/.cog/android-build-out"
+
+ # remove existing out/ dir if it exists
+ if [[ -d "$out_dir" ]]; then
+ echo "Detected existing out/ directory in the Cog workspace which is not supported. Repairing workspace by removing it and creating the symlink to ~/.cog/android-build-out"
+ if ! rm -rf "$out_dir"; then
+ echo "Failed to remove existing out/ directory: $out_dir" >&2
+ kill -INT $$ # exits the script without exiting the user's shell
+ fi
+ fi
+
+ # create symlink
+ echo "Creating symlink: $out_dir -> $link_destination"
+ mkdir -p ${link_destination}
+ if ! ln -s "$link_destination" "$out_dir"; then
+ echo "Failed to create cog symlink: $out_dir -> $link_destination" >&2
+ kill -INT $$ # exits the script without exiting the user's shell
+ fi
+}
+
function getoutdir
{
local top=$(gettop)
diff --git a/target/product/base_product.mk b/target/product/base_product.mk
index 92fc420..acfc653 100644
--- a/target/product/base_product.mk
+++ b/target/product/base_product.mk
@@ -26,3 +26,7 @@
product_manifest.xml \
selinux_policy_product \
product-build.prop \
+
+# Packages included only for eng or userdebug builds, previously debug tagged
+PRODUCT_PACKAGES_DEBUG += \
+ adb_keys \
diff --git a/target/product/base_system.mk b/target/product/base_system.mk
index 5b54051..563511f 100644
--- a/target/product/base_system.mk
+++ b/target/product/base_system.mk
@@ -83,7 +83,6 @@
CtsShimPrivPrebuilt \
debuggerd\
device_config \
- DeviceDiagnostics \
dmctl \
dnsmasq \
dmesgd \
@@ -346,6 +345,11 @@
com.android.webview.bootstrap
endif
+ifneq (,$(RELEASE_RANGING_STACK))
+ PRODUCT_PACKAGES += \
+ com.android.ranging
+endif
+
# VINTF data for system image
PRODUCT_PACKAGES += \
system_manifest.xml \
@@ -466,7 +470,6 @@
# Packages included only for eng or userdebug builds, previously debug tagged
PRODUCT_PACKAGES_DEBUG := \
- adb_keys \
adevice_fingerprint \
arping \
dmuserd \
@@ -534,3 +537,4 @@
$(call soong_config_set, bionic, large_system_property_node, $(RELEASE_LARGE_SYSTEM_PROPERTY_NODE))
$(call soong_config_set, Aconfig, read_from_new_storage, $(RELEASE_READ_FROM_NEW_STORAGE))
+$(call soong_config_set, SettingsLib, legacy_avatar_picker_app_enabled, $(if $(RELEASE_AVATAR_PICKER_APP),,true))
diff --git a/target/product/default_art_config.mk b/target/product/default_art_config.mk
index 1a3f2cf..668f054 100644
--- a/target/product/default_art_config.mk
+++ b/target/product/default_art_config.mk
@@ -76,6 +76,7 @@
com.android.mediaprovider:framework-mediaprovider \
com.android.mediaprovider:framework-pdf \
com.android.mediaprovider:framework-pdf-v \
+ com.android.mediaprovider:framework-photopicker \
com.android.ondevicepersonalization:framework-ondevicepersonalization \
com.android.os.statsd:framework-statsd \
com.android.permission:framework-permission \
@@ -113,6 +114,12 @@
endif
+ifneq (,$(RELEASE_RANGING_STACK))
+ PRODUCT_APEX_BOOT_JARS += \
+ com.android.uwb:framework-ranging \
+ $(call soong_config_set,bootclasspath,release_ranging_stack,true)
+endif
+
# List of system_server classpath jars delivered via apex.
# Keep the list sorted by module names and then library names.
# Note: For modules available in Q, DO NOT add new entries here.
@@ -168,6 +175,11 @@
endif
+ifneq (,$(RELEASE_RANGING_STACK))
+ PRODUCT_APEX_STANDALONE_SYSTEM_SERVER_JARS += \
+ com.android.uwb:service-ranging
+endif
+
# Overrides the (apex, jar) pairs above when determining the on-device location. The format is:
# <old_apex>:<old_jar>:<new_apex>:<new_jar>
PRODUCT_CONFIGURED_JAR_LOCATION_OVERRIDES := \
diff --git a/target/product/generic/Android.bp b/target/product/generic/Android.bp
new file mode 100644
index 0000000..bf82ff1
--- /dev/null
+++ b/target/product/generic/Android.bp
@@ -0,0 +1,822 @@
+android_rootdirs = [
+ "acct",
+ "apex",
+ "bootstrap-apex",
+ "config",
+ "data",
+ "data_mirror",
+ "debug_ramdisk",
+ "dev",
+ "linkerconfig",
+ "metadata",
+ "mnt",
+ "odm",
+ "odm_dlkm",
+ "oem",
+ "postinstall",
+ "proc",
+ "product",
+ "second_stage_resources",
+ "storage",
+ "sys",
+ "system",
+ "system_dlkm",
+ "system_ext",
+ "tmp",
+ "vendor",
+ "vendor_dlkm",
+]
+
+android_symlinks = [
+ {
+ target: "/system/bin/init",
+ name: "init",
+ },
+ {
+ target: "/system/etc",
+ name: "etc",
+ },
+ {
+ target: "/system/bin",
+ name: "bin",
+ },
+ {
+ target: "/product",
+ name: "system/product",
+ },
+ {
+ target: "/vendor",
+ name: "system/vendor",
+ },
+ {
+ target: "/system_ext",
+ name: "system/system_ext",
+ },
+ {
+ target: "/system_dlkm/lib/modules",
+ name: "system/lib/modules",
+ },
+ {
+ target: "/data/user_de/0/com.android.shell/files/bugreports",
+ name: "bugreports",
+ },
+ {
+ target: "/data/cache",
+ name: "cache",
+ },
+ {
+ target: "/sys/kernel/debug",
+ name: "d",
+ },
+ {
+ target: "/storage/self/primary",
+ name: "sdcard",
+ },
+]
+
+filegroup {
+ name: "generic_system_sign_key",
+ srcs: [":avb_testkey_rsa4096"],
+}
+
+phony {
+ name: "generic_system_fonts",
+ required: [
+ "AndroidClock.ttf",
+ "CarroisGothicSC-Regular.ttf",
+ "ComingSoon.ttf",
+ "CutiveMono.ttf",
+ "DancingScript-Regular.ttf",
+ "DroidSansMono.ttf",
+ "NotoColorEmoji.ttf",
+ "NotoColorEmojiFlags.ttf",
+ "NotoNaskhArabic-Bold.ttf",
+ "NotoNaskhArabic-Regular.ttf",
+ "NotoNaskhArabicUI-Bold.ttf",
+ "NotoNaskhArabicUI-Regular.ttf",
+ "NotoSansAdlam-VF.ttf",
+ "NotoSansAhom-Regular.otf",
+ "NotoSansAnatolianHieroglyphs-Regular.otf",
+ "NotoSansArmenian-VF.ttf",
+ "NotoSansAvestan-Regular.ttf",
+ "NotoSansBalinese-Regular.ttf",
+ "NotoSansBamum-Regular.ttf",
+ "NotoSansBassaVah-Regular.otf",
+ "NotoSansBatak-Regular.ttf",
+ "NotoSansBengali-VF.ttf",
+ "NotoSansBengaliUI-VF.ttf",
+ "NotoSansBhaiksuki-Regular.otf",
+ "NotoSansBrahmi-Regular.ttf",
+ "NotoSansBuginese-Regular.ttf",
+ "NotoSansBuhid-Regular.ttf",
+ "NotoSansCJK-Regular.ttc",
+ "NotoSansCanadianAboriginal-Regular.ttf",
+ "NotoSansCarian-Regular.ttf",
+ "NotoSansChakma-Regular.otf",
+ "NotoSansCham-Bold.ttf",
+ "NotoSansCham-Regular.ttf",
+ "NotoSansCherokee-Regular.ttf",
+ "NotoSansCoptic-Regular.ttf",
+ "NotoSansCuneiform-Regular.ttf",
+ "NotoSansCypriot-Regular.ttf",
+ "NotoSansDeseret-Regular.ttf",
+ "NotoSansDevanagari-VF.ttf",
+ "NotoSansDevanagariUI-VF.ttf",
+ "NotoSansEgyptianHieroglyphs-Regular.ttf",
+ "NotoSansElbasan-Regular.otf",
+ "NotoSansEthiopic-VF.ttf",
+ "NotoSansGeorgian-VF.ttf",
+ "NotoSansGlagolitic-Regular.ttf",
+ "NotoSansGothic-Regular.ttf",
+ "NotoSansGrantha-Regular.ttf",
+ "NotoSansGujarati-Bold.ttf",
+ "NotoSansGujarati-Regular.ttf",
+ "NotoSansGujaratiUI-Bold.ttf",
+ "NotoSansGujaratiUI-Regular.ttf",
+ "NotoSansGunjalaGondi-Regular.otf",
+ "NotoSansGurmukhi-VF.ttf",
+ "NotoSansGurmukhiUI-VF.ttf",
+ "NotoSansHanifiRohingya-Regular.otf",
+ "NotoSansHanunoo-Regular.ttf",
+ "NotoSansHatran-Regular.otf",
+ "NotoSansHebrew-Bold.ttf",
+ "NotoSansHebrew-Regular.ttf",
+ "NotoSansImperialAramaic-Regular.ttf",
+ "NotoSansInscriptionalPahlavi-Regular.ttf",
+ "NotoSansInscriptionalParthian-Regular.ttf",
+ "NotoSansJavanese-Regular.otf",
+ "NotoSansKaithi-Regular.ttf",
+ "NotoSansKannada-VF.ttf",
+ "NotoSansKannadaUI-VF.ttf",
+ "NotoSansKayahLi-Regular.ttf",
+ "NotoSansKharoshthi-Regular.ttf",
+ "NotoSansKhmer-VF.ttf",
+ "NotoSansKhmerUI-Bold.ttf",
+ "NotoSansKhmerUI-Regular.ttf",
+ "NotoSansKhojki-Regular.otf",
+ "NotoSansLao-Bold.ttf",
+ "NotoSansLao-Regular.ttf",
+ "NotoSansLaoUI-Bold.ttf",
+ "NotoSansLaoUI-Regular.ttf",
+ "NotoSansLepcha-Regular.ttf",
+ "NotoSansLimbu-Regular.ttf",
+ "NotoSansLinearA-Regular.otf",
+ "NotoSansLinearB-Regular.ttf",
+ "NotoSansLisu-Regular.ttf",
+ "NotoSansLycian-Regular.ttf",
+ "NotoSansLydian-Regular.ttf",
+ "NotoSansMalayalam-VF.ttf",
+ "NotoSansMalayalamUI-VF.ttf",
+ "NotoSansMandaic-Regular.ttf",
+ "NotoSansManichaean-Regular.otf",
+ "NotoSansMarchen-Regular.otf",
+ "NotoSansMasaramGondi-Regular.otf",
+ "NotoSansMedefaidrin-VF.ttf",
+ "NotoSansMeeteiMayek-Regular.ttf",
+ "NotoSansMeroitic-Regular.otf",
+ "NotoSansMiao-Regular.otf",
+ "NotoSansModi-Regular.ttf",
+ "NotoSansMongolian-Regular.ttf",
+ "NotoSansMro-Regular.otf",
+ "NotoSansMultani-Regular.otf",
+ "NotoSansMyanmar-Bold.otf",
+ "NotoSansMyanmar-Medium.otf",
+ "NotoSansMyanmar-Regular.otf",
+ "NotoSansMyanmarUI-Bold.otf",
+ "NotoSansMyanmarUI-Medium.otf",
+ "NotoSansMyanmarUI-Regular.otf",
+ "NotoSansNKo-Regular.ttf",
+ "NotoSansNabataean-Regular.otf",
+ "NotoSansNewTaiLue-Regular.ttf",
+ "NotoSansNewa-Regular.otf",
+ "NotoSansOgham-Regular.ttf",
+ "NotoSansOlChiki-Regular.ttf",
+ "NotoSansOldItalic-Regular.ttf",
+ "NotoSansOldNorthArabian-Regular.otf",
+ "NotoSansOldPermic-Regular.otf",
+ "NotoSansOldPersian-Regular.ttf",
+ "NotoSansOldSouthArabian-Regular.ttf",
+ "NotoSansOldTurkic-Regular.ttf",
+ "NotoSansOriya-Bold.ttf",
+ "NotoSansOriya-Regular.ttf",
+ "NotoSansOriyaUI-Bold.ttf",
+ "NotoSansOriyaUI-Regular.ttf",
+ "NotoSansOsage-Regular.ttf",
+ "NotoSansOsmanya-Regular.ttf",
+ "NotoSansPahawhHmong-Regular.otf",
+ "NotoSansPalmyrene-Regular.otf",
+ "NotoSansPauCinHau-Regular.otf",
+ "NotoSansPhagsPa-Regular.ttf",
+ "NotoSansPhoenician-Regular.ttf",
+ "NotoSansRejang-Regular.ttf",
+ "NotoSansRunic-Regular.ttf",
+ "NotoSansSamaritan-Regular.ttf",
+ "NotoSansSaurashtra-Regular.ttf",
+ "NotoSansSharada-Regular.otf",
+ "NotoSansShavian-Regular.ttf",
+ "NotoSansSinhala-VF.ttf",
+ "NotoSansSinhalaUI-VF.ttf",
+ "NotoSansSoraSompeng-Regular.otf",
+ "NotoSansSoyombo-VF.ttf",
+ "NotoSansSundanese-Regular.ttf",
+ "NotoSansSylotiNagri-Regular.ttf",
+ "NotoSansSymbols-Regular-Subsetted.ttf",
+ "NotoSansSymbols-Regular-Subsetted2.ttf",
+ "NotoSansSyriacEastern-Regular.ttf",
+ "NotoSansSyriacEstrangela-Regular.ttf",
+ "NotoSansSyriacWestern-Regular.ttf",
+ "NotoSansTagalog-Regular.ttf",
+ "NotoSansTagbanwa-Regular.ttf",
+ "NotoSansTaiLe-Regular.ttf",
+ "NotoSansTaiTham-Regular.ttf",
+ "NotoSansTaiViet-Regular.ttf",
+ "NotoSansTakri-VF.ttf",
+ "NotoSansTamil-VF.ttf",
+ "NotoSansTamilUI-VF.ttf",
+ "NotoSansTelugu-VF.ttf",
+ "NotoSansTeluguUI-VF.ttf",
+ "NotoSansThaana-Bold.ttf",
+ "NotoSansThaana-Regular.ttf",
+ "NotoSansThai-Bold.ttf",
+ "NotoSansThai-Regular.ttf",
+ "NotoSansThaiUI-Bold.ttf",
+ "NotoSansThaiUI-Regular.ttf",
+ "NotoSansTifinagh-Regular.otf",
+ "NotoSansUgaritic-Regular.ttf",
+ "NotoSansVai-Regular.ttf",
+ "NotoSansWancho-Regular.otf",
+ "NotoSansWarangCiti-Regular.otf",
+ "NotoSansYi-Regular.ttf",
+ "NotoSerif-Bold.ttf",
+ "NotoSerif-BoldItalic.ttf",
+ "NotoSerif-Italic.ttf",
+ "NotoSerif-Regular.ttf",
+ "NotoSerifArmenian-VF.ttf",
+ "NotoSerifBengali-VF.ttf",
+ "NotoSerifCJK-Regular.ttc",
+ "NotoSerifDevanagari-VF.ttf",
+ "NotoSerifDogra-Regular.ttf",
+ "NotoSerifEthiopic-VF.ttf",
+ "NotoSerifGeorgian-VF.ttf",
+ "NotoSerifGujarati-VF.ttf",
+ "NotoSerifGurmukhi-VF.ttf",
+ "NotoSerifHebrew-Bold.ttf",
+ "NotoSerifHebrew-Regular.ttf",
+ "NotoSerifHentaigana.ttf",
+ "NotoSerifKannada-VF.ttf",
+ "NotoSerifKhmer-Bold.otf",
+ "NotoSerifKhmer-Regular.otf",
+ "NotoSerifLao-Bold.ttf",
+ "NotoSerifLao-Regular.ttf",
+ "NotoSerifMalayalam-VF.ttf",
+ "NotoSerifMyanmar-Bold.otf",
+ "NotoSerifMyanmar-Regular.otf",
+ "NotoSerifNyiakengPuachueHmong-VF.ttf",
+ "NotoSerifSinhala-VF.ttf",
+ "NotoSerifTamil-VF.ttf",
+ "NotoSerifTelugu-VF.ttf",
+ "NotoSerifThai-Bold.ttf",
+ "NotoSerifThai-Regular.ttf",
+ "NotoSerifTibetan-VF.ttf",
+ "NotoSerifYezidi-VF.ttf",
+ "Roboto-Regular.ttf",
+ "RobotoFlex-Regular.ttf",
+ "RobotoStatic-Regular.ttf",
+ "SourceSansPro-Bold.ttf",
+ "SourceSansPro-BoldItalic.ttf",
+ "SourceSansPro-Italic.ttf",
+ "SourceSansPro-Regular.ttf",
+ "SourceSansPro-SemiBold.ttf",
+ "SourceSansPro-SemiBoldItalic.ttf",
+ "font_fallback.xml",
+ "fonts.xml",
+ ],
+}
+
+android_system_image {
+ name: "generic_system_image",
+
+ partition_name: "system",
+ base_dir: "system",
+ dirs: android_rootdirs,
+ symlinks: android_symlinks,
+ file_contexts: ":plat_file_contexts",
+ linker_config_src: ":system_linker_config_json_file",
+ fsverity: {
+ inputs: [
+ "etc/boot-image.prof",
+ "etc/classpaths/*.pb",
+ "etc/dirty-image-objects",
+ "etc/preloaded-classes",
+ "framework/*",
+ "framework/*/*", // framework/{arch}
+ "framework/oat/*/*", // framework/oat/{arch}
+ ],
+ libs: [":framework-res{.export-package.apk}"],
+ },
+ build_logtags: true,
+ gen_aconfig_flags_pb: true,
+
+ compile_multilib: "both",
+
+ use_avb: true,
+ avb_private_key: ":generic_system_sign_key",
+ avb_algorithm: "SHA256_RSA4096",
+ avb_hash_algorithm: "sha256",
+
+ deps: [
+ "abx",
+ "aconfigd",
+ "aflags",
+ "am",
+ "android.software.credentials.prebuilt.xml", // generic_system
+ "android.software.webview.prebuilt.xml", // media_system
+ "android.software.window_magnification.prebuilt.xml", // handheld_system
+ "android.system.suspend-service",
+ "prebuilt_vintf_manifest",
+ "apexd",
+ "appops",
+ "approved-ogki-builds.xml", // base_system
+ "appwidget",
+ "atrace",
+ "audioserver",
+ "bcc",
+ "blank_screen",
+ "blkid",
+ "bmgr",
+ "bootanimation",
+ "bootstat",
+ "bpfloader",
+ "bu",
+ "bugreport",
+ "bugreportz",
+ "cameraserver",
+ "cgroups.json",
+ "cmd",
+ "content",
+ "cppreopts.sh", // generic_system
+ "credstore",
+ "debuggerd",
+ "device_config",
+ "dirty-image-objects",
+ "dmctl",
+ "dmesgd",
+ "dnsmasq",
+ "dpm",
+ "dump.erofs",
+ "dumpstate",
+ "dumpsys",
+ "e2fsck",
+ "enhanced-confirmation.xml", // base_system
+ "etc_hosts",
+ "flags_health_check",
+ "framework-audio_effects.xml", // for handheld // handheld_system
+ "framework-sysconfig.xml",
+ "fs_config_dirs_system",
+ "fs_config_files_system",
+ "fsck.erofs",
+ "fsck.f2fs", // for media_system
+ "fsck_msdos",
+ "fsverity-release-cert-der",
+ "gatekeeperd",
+ "gpu_counter_producer",
+ "gpuservice",
+ "group_system",
+ "gsi_tool",
+ "gsid",
+ "heapprofd",
+ "hid",
+ "hiddenapi-package-whitelist.xml", // from runtime_libart
+ "idc_data",
+ "idmap2",
+ "idmap2d",
+ "ime",
+ "incident",
+ "incident-helper-cmd",
+ "incident_helper",
+ "incidentd",
+ "init.environ.rc-soong",
+ "init.usb.configfs.rc",
+ "init.usb.rc",
+ "init.zygote32.rc",
+ "init.zygote64.rc",
+ "init.zygote64_32.rc",
+ "init_first_stage", // for boot partition
+ "initial-package-stopped-states.xml",
+ "input",
+ "installd",
+ "ip", // base_system
+ "iptables",
+ "kcmdlinectrl",
+ "kernel-lifetimes.xml", // base_system
+ "keychars_data",
+ "keylayout_data",
+ "keystore2",
+ "ld.mc",
+ "llkd", // base_system
+ "lmkd", // base_system
+ "local_time.default", // handheld_vendo
+ "locksettings", // base_system
+ "logcat", // base_system
+ "logd", // base_system
+ "logpersist.start",
+ "lpdump", // base_system
+ "lshal", // base_system
+ "make_f2fs", // media_system
+ "mdnsd", // base_system
+ "media_profiles_V1_0.dtd", // base_system
+ "mediacodec.policy", // base_system
+ "mediaextractor", // base_system
+ "mediametrics", // base_system
+ "misctrl", // from base_system
+ "mke2fs", // base_system
+ "mkfs.erofs", // base_system
+ "monkey", // base_system
+ "mtectrl", // base_system
+ "ndc", // base_system
+ "netd", // base_system
+ "netutils-wrapper-1.0", // full_base
+ "notice_xml_system",
+ "odsign", // base_system
+ "otapreopt_script", // generic_system
+ "package-shareduid-allowlist.xml", // base_system
+ "passwd_system", // base_system
+ "perfetto", // base_system
+ "ping", // base_system
+ "ping6", // base_system
+ "pintool", // base_system
+ "platform.xml", // base_system
+ "pm", // base_system
+ "preinstalled-packages-asl-files.xml", // base_system
+ "preinstalled-packages-platform-generic-system.xml", // generic_system
+ "preinstalled-packages-platform-handheld-system.xml", // handheld_system
+ "preinstalled-packages-platform.xml", // base_system
+ "preinstalled-packages-strict-signature.xml", // base_system
+ "preloaded-classes", // ok
+ "printflags", // base_system
+ "privapp-permissions-platform.xml", // base_system
+ "prng_seeder", // base_system
+ "public.libraries.android.txt",
+ "recovery-persist", // base_system
+ "recovery-refresh", // generic_system
+ "requestsync", // media_system
+ "resize2fs", // base_system
+ "rss_hwm_reset", // base_system
+ "run-as", // base_system
+ "schedtest", // base_system
+ "screencap", // base_system
+ "screenrecord", // handheld_system
+ "sdcard", // base_system
+ "secdiscard", // base_system
+ "sensorservice", // base_system
+ "service", // base_system
+ "servicemanager", // base_system
+ "settings", // base_system
+ "sfdo", // base_system
+ "sgdisk", // base_system
+ "sm", // base_system
+ "snapshotctl", // base_system
+ "snapuserd", // base_system
+ "snapuserd_ramdisk", // ramdisk
+ "storaged", // base_system
+ "surfaceflinger", // base_system
+ "svc", // base_system
+ "task_profiles.json", // base_system
+ "tc", // base_system
+ "telecom", // base_system
+ "tombstoned", // base_system
+ "traced", // base_system
+ "traced_probes", // base_system
+ "tune2fs", // base_system
+ "uiautomator", // base_system
+ "uinput", // base_system
+ "uncrypt", // base_system
+ "update_engine", // generic_system
+ "update_engine_sideload", // recovery
+ "update_verifier", // generic_system
+ "usbd", // base_system
+ "vdc", // base_system
+ "virtual_camera", // handheld_system // release_package_virtual_camera
+ "vold", // base_system
+ "vr", // handheld_system
+ "watchdogd", // base_system
+ "wifi.rc", // base_system
+ "wificond", // base_system
+ "wm", // base_system
+ ] + select(release_flag("RELEASE_PLATFORM_VERSION_CODENAME"), {
+ "REL": [],
+ default: [
+ "android.software.preview_sdk.prebuilt.xml", // media_system
+ ],
+ }) + select(soong_config_variable("ANDROID", "release_package_profiling_module"), {
+ "true": [
+ "trace_redactor", // base_system (RELEASE_PACKAGE_PROFILING_MODULE)
+ ],
+ default: [],
+ }) + select(product_variable("debuggable"), {
+ true: [
+ "adevice_fingerprint",
+ "arping",
+ "avbctl",
+ "bootctl",
+ "dmuserd",
+ "evemu-record",
+ "idlcli",
+ "init-debug.rc",
+ "iotop",
+ "iperf3",
+ "iw",
+ "layertracegenerator",
+ "logtagd.rc",
+ "ot-cli-ftd",
+ "ot-ctl",
+ "procrank",
+ "profcollectctl",
+ "profcollectd",
+ "record_binder",
+ "sanitizer-status",
+ "servicedispatcher",
+ "showmap",
+ "sqlite3",
+ "ss",
+ "start_with_lockagent",
+ "strace",
+ "su",
+ "tinycap",
+ "tinyhostless",
+ "tinymix",
+ "tinypcminfo",
+ "tinyplay", // host
+ "tracepath",
+ "tracepath6",
+ "traceroute6",
+ "unwind_info",
+ "unwind_reg_info",
+ "unwind_symbols",
+ "update_engine_client",
+ ],
+ default: [],
+ }),
+ multilib: {
+ common: {
+ deps: [
+ "BackupRestoreConfirmation", // base_system
+ "BasicDreams", // handheld_system
+ "BlockedNumberProvider", // handheld_system
+ "BluetoothMidiService", // handheld_system
+ "BookmarkProvider", // handheld_system
+ "BuiltInPrintService", // handheld_system
+ "CalendarProvider", // handheld_system
+ "CallLogBackup", // telephony_system
+ "CameraExtensionsProxy", // handheld_system
+ "CaptivePortalLogin", // handheld_system
+ "CarrierDefaultApp", // telephony_system
+ "CellBroadcastLegacyApp", // telephony_system
+ "CertInstaller", // handheld_system
+ "CompanionDeviceManager", // media_system
+ "ContactsProvider", // base_system
+ "CredentialManager", // handheld_system
+ "DeviceAsWebcam", // handheld_system
+ "DocumentsUI", // handheld_system
+ "DownloadProvider", // base_system
+ "DownloadProviderUi", // handheld_system
+ "DynamicSystemInstallationService", // base_system
+ "E2eeContactKeysProvider", // base_system
+ "EasterEgg", // handheld_system
+ "ExtShared", // base_system
+ "ExternalStorageProvider", // handheld_system
+ "FusedLocation", // handheld_system
+ "HTMLViewer", // media_system
+ "InputDevices", // handheld_system
+ "IntentResolver", // base_system
+ "KeyChain", // handheld_system
+ "LiveWallpapersPicker", // generic_system, full_base
+ "LocalTransport", // base_system
+ "ManagedProvisioning", // handheld_system
+ "MediaProviderLegacy", // base_system
+ "MmsService", // handheld_system
+ "MtpService", // handheld_system
+ "MusicFX", // handheld_system
+ "NetworkStack", // base_system
+ "ONS", // telephony_system
+ "PacProcessor", // handheld_system
+ "PackageInstaller", // base_system
+ "PartnerBookmarksProvider", // generic_system
+ "PhotoTable", // full_base
+ "PrintRecommendationService", // handheld_system
+ "PrintSpooler", // handheld_system
+ "ProxyHandler", // handheld_system
+ "SecureElement", // handheld_system
+ "SettingsProvider", // base_system
+ "SharedStorageBackup", // handheld_system
+ "Shell", // base_system
+ "SimAppDialog", // handheld_system
+ "SoundPicker", // not installed by anyone
+ "StatementService", // media_system
+ "Stk", // generic_system
+ "Tag", // generic_system
+ "TeleService", // handheld_system
+ "Telecom", // handheld_system
+ "TelephonyProvider", // handheld_system
+ "Traceur", // handheld_system
+ "UserDictionaryProvider", // handheld_system
+ "VpnDialogs", // handheld_system
+ "WallpaperBackup", // base_system
+ "adbd_system_api", // base_system
+ "android.hidl.base-V1.0-java", // base_system
+ "android.hidl.manager-V1.0-java", // base_system
+ "android.test.base", // from runtime_libart
+ "android.test.mock", // base_system
+ "android.test.runner", // base_system
+ "aosp_mainline_modules", // ok
+ "build_flag_system", // base_system
+ "charger_res_images", // generic_system
+ "com.android.apex.cts.shim.v1_prebuilt", // ok
+ "com.android.cellbroadcast", // telephony_system
+ "com.android.future.usb.accessory", // media_system
+ "com.android.location.provider", // base_system
+ "com.android.media.remotedisplay", // media_system
+ "com.android.media.remotedisplay.xml", // media_system
+ "com.android.mediadrm.signer", // media_system
+ "com.android.nfc_extras", // ok
+ "com.android.nfcservices", // base_system (RELEASE_PACKAGE_NFC_STACK != NfcNci)
+ "com.android.runtime", // ok
+ "dex_bootjars",
+ "ext", // from runtime_libart
+ "framework-graphics", // base_system
+ "framework-location", // base_system
+ "framework-minus-apex-install-dependencies", // base_system
+ "framework_compatibility_matrix.device.xml",
+ "generic_system_fonts", // ok
+ "hwservicemanager_compat_symlink_module", // base_system
+ "hyph-data",
+ "ims-common", // base_system
+ "init_system", // base_system
+ "javax.obex", // base_system
+ "llndk.libraries.txt", //ok
+ "org.apache.http.legacy", // base_system
+ "perfetto-extras", // system
+ "sanitizer.libraries.txt", // base_system
+ "selinux_policy_system_soong", // ok
+ "services", // base_system
+ "shell_and_utilities_system", // ok
+ "system-build.prop",
+ "system_compatibility_matrix.xml", //base_system
+ "telephony-common", // libs from TeleService
+ "voip-common", // base_system
+ ] + select(soong_config_variable("ANDROID", "release_crashrecovery_module"), {
+ "true": [
+ "com.android.crashrecovery", // base_system (RELEASE_CRASHRECOVERY_MODULE)
+ ],
+ default: [],
+ }) + select(soong_config_variable("ANDROID", "release_package_profiling_module"), {
+ "true": [
+ "com.android.profiling", // base_system (RELEASE_PACKAGE_PROFILING_MODULE)
+ ],
+ default: [],
+ }) + select(release_flag("RELEASE_AVATAR_PICKER_APP"), {
+ true: [
+ "AvatarPicker", // generic_system (RELEASE_AVATAR_PICKER_APP)
+ ],
+ default: [],
+ }),
+ },
+ prefer32: {
+ deps: [
+ "drmserver", // media_system
+ "mediaserver", // base_system
+ ],
+ },
+ lib64: {
+ deps: [
+ "android.system.virtualizationcommon-ndk",
+ "android.system.virtualizationservice-ndk",
+ "libgsi",
+ "servicemanager",
+ ],
+ },
+ both: {
+ deps: [
+ "android.hardware.biometrics.fingerprint@2.1", // generic_system
+ "android.hardware.radio.config@1.0", // generic_system
+ "android.hardware.radio.deprecated@1.0", // generic_system
+ "android.hardware.radio@1.0", // generic_system
+ "android.hardware.radio@1.1", // generic_system
+ "android.hardware.radio@1.2", // generic_system
+ "android.hardware.radio@1.3", // generic_system
+ "android.hardware.radio@1.4", // generic_system
+ "android.hardware.secure_element@1.0", // generic_system
+ "app_process", // base_system
+ "boringssl_self_test", // base_system
+ "heapprofd_client", // base_system
+ "libEGL", // base_system
+ "libEGL_angle", // base_system
+ "libETC1", // base_system
+ "libFFTEm", // base_system
+ "libGLESv1_CM", // base_system
+ "libGLESv1_CM_angle", // base_system
+ "libGLESv2", // base_system
+ "libGLESv2_angle", // base_system
+ "libGLESv3", // base_system
+ "libOpenMAXAL", // base_system
+ "libOpenSLES", // base_system
+ "libaaudio", // base_system
+ "libalarm_jni", // base_system
+ "libamidi", // base_system
+ "libandroid",
+ "libandroid_runtime",
+ "libandroid_servers",
+ "libandroidfw",
+ "libartpalette-system",
+ "libaudio-resampler", // generic-system
+ "libaudioeffect_jni",
+ "libaudiohal", // generic-system
+ "libaudiopolicyengineconfigurable", // generic-system
+ "libbinder",
+ "libbinder_ndk",
+ "libbinder_rpc_unstable",
+ "libcamera2ndk",
+ "libclang_rt.asan",
+ "libcompiler_rt",
+ "libcutils", // used by many libs
+ "libdmabufheap", // used by many libs
+ "libdrm", // used by many libs // generic_system
+ "libdrmframework", // base_system
+ "libdrmframework_jni", // base_system
+ "libfdtrack", // base_system
+ "libfilterfw", // base_system
+ "libfilterpack_imageproc", // media_system
+ "libfwdlockengine", // generic_system
+ "libgatekeeper", // base_system
+ "libgui", // base_system
+ "libhardware", // base_system
+ "libhardware_legacy", // base_system
+ "libhidltransport", // generic_system
+ "libhwbinder", // generic_system
+ "libinput", // base_system
+ "libinputflinger", // base_system
+ "libiprouteutil", // base_system
+ "libjnigraphics", // base_system
+ "libjpeg", // base_system
+ "liblog", // base_system
+ "liblogwrap", // generic_system
+ "liblz4", // generic_system
+ "libmedia", // base_system
+ "libmedia_jni", // base_system
+ "libmediandk", // base_system
+ "libminui", // generic_system
+ "libmtp", // base_system
+ "libnetd_client", // base_system
+ "libnetlink", // base_system
+ "libnetutils", // base_system
+ "libneuralnetworks_packageinfo", // base_system
+ "libnl", // generic_system
+ "libpdfium", // base_system
+ "libpolicy-subsystem", // generic_system
+ "libpower", // base_system
+ "libpowermanager", // base_system
+ "libprotobuf-cpp-full", // generic_system
+ "libradio_metadata", // base_system
+ "librs_jni", // handheld_system
+ "librtp_jni", // base_system
+ "libsensorservice", // base_system
+ "libsfplugin_ccodec", // base_system
+ "libskia", // base_system
+ "libsonic", // base_system
+ "libsonivox", // base_system
+ "libsoundpool", // base_system
+ "libspeexresampler", // base_system
+ "libsqlite", // base_system
+ "libstagefright", // base_system
+ "libstagefright_foundation", // base_system
+ "libstagefright_omx", // base_system
+ "libstdc++", // base_system
+ "libsysutils", // base_system
+ "libui", // base_system
+ "libusbhost", // base_system
+ "libutils", // base_system
+ "libvendorsupport", // llndk library
+ "libvintf_jni", // base_system
+ "libvulkan", // base_system
+ "libwebviewchromium_loader", // media_system
+ "libwebviewchromium_plat_support", // media_system
+ "libwilhelm", // base_system
+ "linker", // base_system
+ ] + select(soong_config_variable("ANDROID", "TARGET_DYNAMIC_64_32_DRMSERVER"), {
+ "true": ["drmserver"],
+ default: [],
+ }) + select(soong_config_variable("ANDROID", "TARGET_DYNAMIC_64_32_MEDIASERVER"), {
+ "true": ["mediaserver"],
+ default: [],
+ }),
+ },
+ },
+}
+
+prebuilt_etc {
+ name: "prebuilt_vintf_manifest",
+ src: "manifest.xml",
+ filename: "manifest.xml",
+ relative_install_path: "vintf",
+ no_full_install: true,
+}
diff --git a/target/product/generic/OWNERS b/target/product/generic/OWNERS
new file mode 100644
index 0000000..6d1446f
--- /dev/null
+++ b/target/product/generic/OWNERS
@@ -0,0 +1,6 @@
+# Bug component: 1322713
+inseob@google.com
+jeongik@google.com
+jiyong@google.com
+justinyun@google.com
+kiyoungkim@google.com
diff --git a/target/product/generic/manifest.xml b/target/product/generic/manifest.xml
new file mode 100644
index 0000000..1df2c0d
--- /dev/null
+++ b/target/product/generic/manifest.xml
@@ -0,0 +1,54 @@
+<!--
+ Input:
+ system/libhidl/vintfdata/manifest.xml
+-->
+<manifest version="8.0" type="framework">
+ <hal format="hidl" max-level="6">
+ <name>android.frameworks.displayservice</name>
+ <transport>hwbinder</transport>
+ <fqname>@1.0::IDisplayService/default</fqname>
+ </hal>
+ <hal format="hidl" max-level="5">
+ <name>android.frameworks.schedulerservice</name>
+ <transport>hwbinder</transport>
+ <fqname>@1.0::ISchedulingPolicyService/default</fqname>
+ </hal>
+ <hal format="aidl">
+ <name>android.frameworks.sensorservice</name>
+ <fqname>ISensorManager/default</fqname>
+ </hal>
+ <hal format="hidl" max-level="8">
+ <name>android.frameworks.sensorservice</name>
+ <transport>hwbinder</transport>
+ <fqname>@1.0::ISensorManager/default</fqname>
+ </hal>
+ <hal format="hidl" max-level="8">
+ <name>android.hidl.memory</name>
+ <transport arch="32+64">passthrough</transport>
+ <fqname>@1.0::IMapper/ashmem</fqname>
+ </hal>
+ <hal format="hidl" max-level="7">
+ <name>android.system.net.netd</name>
+ <transport>hwbinder</transport>
+ <fqname>@1.1::INetd/default</fqname>
+ </hal>
+ <hal format="hidl" max-level="7">
+ <name>android.system.wifi.keystore</name>
+ <transport>hwbinder</transport>
+ <fqname>@1.0::IKeystore/default</fqname>
+ </hal>
+ <hal format="native">
+ <name>netutils-wrapper</name>
+ <version>1.0</version>
+ </hal>
+ <system-sdk>
+ <version>29</version>
+ <version>30</version>
+ <version>31</version>
+ <version>32</version>
+ <version>33</version>
+ <version>34</version>
+ <version>35</version>
+ <version>VanillaIceCream</version>
+ </system-sdk>
+</manifest>
diff --git a/target/product/go_defaults.mk b/target/product/go_defaults.mk
index 4627fde..c928530 100644
--- a/target/product/go_defaults.mk
+++ b/target/product/go_defaults.mk
@@ -17,7 +17,8 @@
# Inherit common Android Go defaults.
$(call inherit-product, build/make/target/product/go_defaults_common.mk)
-PRODUCT_RELEASE_CONFIG_MAPS += $(wildcard vendor/google_shared/build/release/go_devices/release_config_map.textproto)
+# Product config map to toggle between sources and prebuilts of required mainline modules
+PRODUCT_RELEASE_CONFIG_MAPS += $(wildcard vendor/google_shared/build/release/gms_mainline_go/required/release_config_map.textproto)
# Add the system properties.
TARGET_SYSTEM_PROP += \
diff --git a/target/product/go_defaults_common.mk b/target/product/go_defaults_common.mk
index fd4047a..0fcf16b 100644
--- a/target/product/go_defaults_common.mk
+++ b/target/product/go_defaults_common.mk
@@ -24,11 +24,6 @@
# Speed profile services and wifi-service to reduce RAM and storage.
PRODUCT_SYSTEM_SERVER_COMPILER_FILTER := speed-profile
-# Use a profile based boot image for this device. Note that this is currently a
-# generic profile and not Android Go optimized.
-PRODUCT_USE_PROFILE_FOR_BOOT_IMAGE := true
-PRODUCT_DEX_PREOPT_BOOT_IMAGE_PROFILE_LOCATION := frameworks/base/config/boot-image-profile.txt
-
# Do not generate libartd.
PRODUCT_ART_TARGET_INCLUDE_DEBUG_BUILD := false
diff --git a/target/product/handheld_system.mk b/target/product/handheld_system.mk
index 3f3bd01..546bbe7 100644
--- a/target/product/handheld_system.mk
+++ b/target/product/handheld_system.mk
@@ -46,6 +46,7 @@
CertInstaller \
CredentialManager \
DeviceAsWebcam \
+ DeviceDiagnostics \
DocumentsUI \
DownloadProviderUi \
EasterEgg \
diff --git a/target/product/media_system_ext.mk b/target/product/media_system_ext.mk
index 2e20af3..30dd2e2 100644
--- a/target/product/media_system_ext.mk
+++ b/target/product/media_system_ext.mk
@@ -23,3 +23,6 @@
# /system_ext packages
PRODUCT_PACKAGES += \
vndk_apex_snapshot_package \
+
+# Window Extensions
+$(call inherit-product, $(SRC_TARGET_DIR)/product/window_extensions_base.mk)
diff --git a/target/product/security/Android.mk b/target/product/security/Android.mk
index 91b272c..138e5bb 100644
--- a/target/product/security/Android.mk
+++ b/target/product/security/Android.mk
@@ -10,7 +10,7 @@
LOCAL_LICENSE_CONDITIONS := notice
LOCAL_NOTICE_FILE := build/soong/licenses/LICENSE
LOCAL_MODULE_CLASS := ETC
- LOCAL_MODULE_PATH := $(TARGET_ROOT_OUT)
+ LOCAL_MODULE_PATH := $(TARGET_OUT_PRODUCT_ETC)/security
LOCAL_PREBUILT_MODULE_FILE := $(PRODUCT_ADB_KEYS)
include $(BUILD_PREBUILT)
endif
diff --git a/target/product/window_extensions.mk b/target/product/window_extensions.mk
index 5f5431f..d27a613 100644
--- a/target/product/window_extensions.mk
+++ b/target/product/window_extensions.mk
@@ -14,11 +14,14 @@
# limitations under the License.
#
-# /system_ext packages
-PRODUCT_PACKAGES += \
- androidx.window.extensions \
- androidx.window.sidecar
-
-# properties
+# Extension of window_extensions_base.mk to enable the activity embedding
+# feature for all apps by default. All large screen devices must inherit
+# this in build. Optional for other form factors.
+#
+# Indicated whether the Activity Embedding feature should be guarded by
+# Android 15 to avoid app compat impact.
+# If true (or not set), the feature is only enabled for apps with target
+# SDK of Android 15 or above.
+# If false, the feature is enabled for all apps.
PRODUCT_PRODUCT_PROPERTIES += \
- persist.wm.extensions.enabled=true
+ persist.wm.extensions.activity_embedding_guard_with_android_15=false
diff --git a/target/product/window_extensions_base.mk b/target/product/window_extensions_base.mk
new file mode 100644
index 0000000..ee0e5e7
--- /dev/null
+++ b/target/product/window_extensions_base.mk
@@ -0,0 +1,33 @@
+#
+# Copyright (C) 2024 The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+#
+
+# The base version of window_extensions.mk to be included on all non-wearable
+# devices. Devices that don't support multi-window can choose to drop this.
+#
+# Note: by default the Activity Embedding feature is guarded by app's
+# targetSDK on Android 15 to avoid app compat impact.
+#
+# Large screen devices must inherit window_extensions.mk to enable the
+# Activity Embedding feature for all apps.
+
+# /system_ext packages
+PRODUCT_PACKAGES += \
+ androidx.window.extensions \
+ androidx.window.sidecar
+
+# properties
+PRODUCT_PRODUCT_PROPERTIES += \
+ persist.wm.extensions.enabled=true
diff --git a/teams/Android.bp b/teams/Android.bp
index 94585fc..96d241b 100644
--- a/teams/Android.bp
+++ b/teams/Android.bp
@@ -4440,3 +4440,17 @@
// go/trendy/manage/engineers/5097003746426880
trendy_team_id: "5097003746426880",
}
+
+team {
+ name: "trendy_team_desktop_firmware",
+
+ // go/trendy/manage/engineers/5787938454863872
+ trendy_team_id: "5787938454863872",
+}
+
+team {
+ name: "trendy_team_art_cloud",
+
+ // go/trendy/manage/engineers/5121440647577600
+ trendy_team_id: "5121440647577600",
+}
diff --git a/tools/aconfig/aconfig/src/codegen/cpp.rs b/tools/aconfig/aconfig/src/codegen/cpp.rs
index 2c569da..7a9c382 100644
--- a/tools/aconfig/aconfig/src/codegen/cpp.rs
+++ b/tools/aconfig/aconfig/src/codegen/cpp.rs
@@ -283,39 +283,23 @@
virtual ~flag_provider_interface() = default;
virtual bool disabled_ro() = 0;
-
- virtual void disabled_ro(bool val) = 0;
-
virtual bool disabled_rw() = 0;
-
- virtual void disabled_rw(bool val) = 0;
-
virtual bool disabled_rw_exported() = 0;
-
- virtual void disabled_rw_exported(bool val) = 0;
-
virtual bool disabled_rw_in_other_namespace() = 0;
-
- virtual void disabled_rw_in_other_namespace(bool val) = 0;
-
virtual bool enabled_fixed_ro() = 0;
-
- virtual void enabled_fixed_ro(bool val) = 0;
-
virtual bool enabled_fixed_ro_exported() = 0;
-
- virtual void enabled_fixed_ro_exported(bool val) = 0;
-
virtual bool enabled_ro() = 0;
-
- virtual void enabled_ro(bool val) = 0;
-
virtual bool enabled_ro_exported() = 0;
-
- virtual void enabled_ro_exported(bool val) = 0;
-
virtual bool enabled_rw() = 0;
+ virtual void disabled_ro(bool val) = 0;
+ virtual void disabled_rw(bool val) = 0;
+ virtual void disabled_rw_exported(bool val) = 0;
+ virtual void disabled_rw_in_other_namespace(bool val) = 0;
+ virtual void enabled_fixed_ro(bool val) = 0;
+ virtual void enabled_fixed_ro_exported(bool val) = 0;
+ virtual void enabled_ro(bool val) = 0;
+ virtual void enabled_ro_exported(bool val) = 0;
virtual void enabled_rw(bool val) = 0;
virtual void reset_flags() {}
diff --git a/tools/aconfig/aconfig/src/codegen/java.rs b/tools/aconfig/aconfig/src/codegen/java.rs
index dbc4ab5..a34166d 100644
--- a/tools/aconfig/aconfig/src/codegen/java.rs
+++ b/tools/aconfig/aconfig/src/codegen/java.rs
@@ -137,6 +137,7 @@
pub default_value: bool,
pub device_config_namespace: String,
pub device_config_flag: String,
+ pub flag_name: String,
pub flag_name_constant_suffix: String,
pub flag_offset: u16,
pub is_read_write: bool,
@@ -156,6 +157,7 @@
default_value: pf.state() == ProtoFlagState::ENABLED,
device_config_namespace: pf.namespace().to_string(),
device_config_flag,
+ flag_name: pf.name().to_string(),
flag_name_constant_suffix: pf.name().to_ascii_uppercase(),
flag_offset: *flag_offsets.get(pf.name()).expect("didnt find package offset :("),
is_read_write: pf.permission() == ProtoFlagPermission::READ_WRITE,
@@ -507,97 +509,39 @@
private static FeatureFlags FEATURE_FLAGS = new FeatureFlagsImpl();
}"#;
- let expected_featureflagsmpl_content_0 = r#"
+ let expected_featureflagsmpl_content = r#"
package com.android.aconfig.test;
// TODO(b/303773055): Remove the annotation after access issue is resolved.
import android.compat.annotation.UnsupportedAppUsage;
import android.provider.DeviceConfig;
import android.provider.DeviceConfig.Properties;
- "#;
+ import android.aconfig.storage.StorageInternalReader;
+ import java.nio.file.Files;
+ import java.nio.file.Paths;
- let expected_featureflagsmpl_content_1 = r#"
/** @hide */
public final class FeatureFlagsImpl implements FeatureFlags {
+ private static final boolean isReadFromNew = Files.exists(Paths.get("/metadata/aconfig/boot/enable_only_new_storage"));
+ private static volatile boolean isCached = false;
private static volatile boolean aconfig_test_is_cached = false;
private static volatile boolean other_namespace_is_cached = false;
private static boolean disabledRw = false;
private static boolean disabledRwExported = false;
private static boolean disabledRwInOtherNamespace = false;
private static boolean enabledRw = true;
- "#;
- let expected_featureflagsmpl_content_2 = r#"
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean disabledRo() {
- return false;
- }
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean disabledRw() {
- if (!aconfig_test_is_cached) {
- load_overrides_aconfig_test();
+ private void init() {
+ StorageInternalReader reader = null;
+ try {
+ reader = new StorageInternalReader("system", "com.android.aconfig.test");
+ disabledRw = reader.getBooleanFlagValue(1);
+ disabledRwExported = reader.getBooleanFlagValue(2);
+ enabledRw = reader.getBooleanFlagValue(8);
+ disabledRwInOtherNamespace = reader.getBooleanFlagValue(3);
+ } catch (Exception e) {
+ throw new RuntimeException("Cannot read flag in codegen", e);
}
- return disabledRw;
+ isCached = true;
}
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean disabledRwExported() {
- if (!aconfig_test_is_cached) {
- load_overrides_aconfig_test();
- }
- return disabledRwExported;
- }
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean disabledRwInOtherNamespace() {
- if (!other_namespace_is_cached) {
- load_overrides_other_namespace();
- }
- return disabledRwInOtherNamespace;
- }
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean enabledFixedRo() {
- return true;
- }
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean enabledFixedRoExported() {
- return true;
- }
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean enabledRo() {
- return true;
- }
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean enabledRoExported() {
- return true;
- }
- @Override
- @com.android.aconfig.annotations.AconfigFlagAccessor
- @UnsupportedAppUsage
- public boolean enabledRw() {
- if (!aconfig_test_is_cached) {
- load_overrides_aconfig_test();
- }
- return enabledRw;
- }
- }
- "#;
-
- let expect_featureflagsimpl_content_old = expected_featureflagsmpl_content_0.to_owned()
- + expected_featureflagsmpl_content_1
- + r#"
private void load_overrides_aconfig_test() {
try {
Properties properties = DeviceConfig.getProperties("aconfig_test");
@@ -636,196 +580,104 @@
);
}
other_namespace_is_cached = true;
- }"#
- + expected_featureflagsmpl_content_2;
+ }
+
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean disabledRo() {
+ return false;
+ }
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean disabledRw() {
+ if (isReadFromNew) {
+ if (!isCached) {
+ init();
+ }
+ } else {
+ if (!aconfig_test_is_cached) {
+ load_overrides_aconfig_test();
+ }
+ }
+ return disabledRw;
+ }
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean disabledRwExported() {
+ if (isReadFromNew) {
+ if (!isCached) {
+ init();
+ }
+ } else {
+ if (!aconfig_test_is_cached) {
+ load_overrides_aconfig_test();
+ }
+ }
+ return disabledRwExported;
+ }
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean disabledRwInOtherNamespace() {
+ if (isReadFromNew) {
+ if (!isCached) {
+ init();
+ }
+ } else {
+ if (!other_namespace_is_cached) {
+ load_overrides_other_namespace();
+ }
+ }
+ return disabledRwInOtherNamespace;
+ }
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean enabledFixedRo() {
+ return true;
+ }
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean enabledFixedRoExported() {
+ return true;
+ }
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean enabledRo() {
+ return true;
+ }
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean enabledRoExported() {
+ return true;
+ }
+ @Override
+ @com.android.aconfig.annotations.AconfigFlagAccessor
+ @UnsupportedAppUsage
+ public boolean enabledRw() {
+ if (isReadFromNew) {
+ if (!isCached) {
+ init();
+ }
+ } else {
+ if (!aconfig_test_is_cached) {
+ load_overrides_aconfig_test();
+ }
+ }
+ return enabledRw;
+ }
+ }
+ "#;
let mut file_set = HashMap::from([
("com/android/aconfig/test/Flags.java", expect_flags_content.as_str()),
- (
- "com/android/aconfig/test/FeatureFlagsImpl.java",
- &expect_featureflagsimpl_content_old,
- ),
- ("com/android/aconfig/test/FeatureFlags.java", EXPECTED_FEATUREFLAGS_COMMON_CONTENT),
- (
- "com/android/aconfig/test/CustomFeatureFlags.java",
- EXPECTED_CUSTOMFEATUREFLAGS_CONTENT,
- ),
- (
- "com/android/aconfig/test/FakeFeatureFlagsImpl.java",
- EXPECTED_FAKEFEATUREFLAGSIMPL_CONTENT,
- ),
- ]);
-
- for file in generated_files {
- let file_path = file.path.to_str().unwrap();
- assert!(file_set.contains_key(file_path), "Cannot find {}", file_path);
- assert_eq!(
- None,
- crate::test::first_significant_code_diff(
- file_set.get(file_path).unwrap(),
- &String::from_utf8(file.contents).unwrap()
- ),
- "File {} content is not correct",
- file_path
- );
- file_set.remove(file_path);
- }
-
- assert!(file_set.is_empty());
-
- let parsed_flags = crate::test::parse_test_flags();
- let mode = CodegenMode::Production;
- let modified_parsed_flags =
- crate::commands::modify_parsed_flags_based_on_mode(parsed_flags, mode).unwrap();
- let flag_ids =
- assign_flag_ids(crate::test::TEST_PACKAGE, modified_parsed_flags.iter()).unwrap();
- let generated_files = generate_java_code(
- crate::test::TEST_PACKAGE,
- modified_parsed_flags.into_iter(),
- mode,
- flag_ids,
- true,
- )
- .unwrap();
-
- let expect_featureflagsimpl_content_new = expected_featureflagsmpl_content_0.to_owned()
- + r#"
- import android.aconfig.storage.StorageInternalReader;
- import android.util.Log;
- "#
- + expected_featureflagsmpl_content_1
- + r#"
- StorageInternalReader reader;
- boolean readFromNewStorage;
-
- boolean useNewStorageValueAndDiscardOld = false;
-
- private final static String TAG = "AconfigJavaCodegen";
- private final static String SUCCESS_LOG = "success: %s value matches";
- private final static String MISMATCH_LOG = "error: %s value mismatch, new storage value is %s, old storage value is %s";
- private final static String ERROR_LOG = "error: failed to read flag value";
-
- private void init() {
- if (reader != null) return;
- if (DeviceConfig.getBoolean("core_experiments_team_internal", "com.android.providers.settings.storage_test_mission_1", false)) {
- readFromNewStorage = true;
- try {
- reader = new StorageInternalReader("system", "com.android.aconfig.test");
- } catch (Exception e) {
- reader = null;
- }
- }
-
- useNewStorageValueAndDiscardOld =
- DeviceConfig.getBoolean("core_experiments_team_internal", "com.android.providers.settings.use_new_storage_value", false);
- }
-
- private void load_overrides_aconfig_test() {
- try {
- Properties properties = DeviceConfig.getProperties("aconfig_test");
- disabledRw =
- properties.getBoolean(Flags.FLAG_DISABLED_RW, false);
- disabledRwExported =
- properties.getBoolean(Flags.FLAG_DISABLED_RW_EXPORTED, false);
- enabledRw =
- properties.getBoolean(Flags.FLAG_ENABLED_RW, true);
- } catch (NullPointerException e) {
- throw new RuntimeException(
- "Cannot read value from namespace aconfig_test "
- + "from DeviceConfig. It could be that the code using flag "
- + "executed before SettingsProvider initialization. Please use "
- + "fixed read-only flag by adding is_fixed_read_only: true in "
- + "flag declaration.",
- e
- );
- }
- aconfig_test_is_cached = true;
- init();
- if (readFromNewStorage && reader != null) {
- boolean val;
- try {
- val = reader.getBooleanFlagValue(1);
- if (val == disabledRw) {
- Log.i(TAG, String.format(SUCCESS_LOG, "disabledRw"));
- } else {
- Log.i(TAG, String.format(MISMATCH_LOG, "disabledRw", val, disabledRw));
- }
-
- if (useNewStorageValueAndDiscardOld) {
- disabledRw = val;
- }
-
- val = reader.getBooleanFlagValue(2);
- if (val == disabledRwExported) {
- Log.i(TAG, String.format(SUCCESS_LOG, "disabledRwExported"));
- } else {
- Log.i(TAG, String.format(MISMATCH_LOG, "disabledRwExported", val, disabledRwExported));
- }
-
- if (useNewStorageValueAndDiscardOld) {
- disabledRwExported = val;
- }
-
- val = reader.getBooleanFlagValue(8);
- if (val == enabledRw) {
- Log.i(TAG, String.format(SUCCESS_LOG, "enabledRw"));
- } else {
- Log.i(TAG, String.format(MISMATCH_LOG, "enabledRw", val, enabledRw));
- }
-
- if (useNewStorageValueAndDiscardOld) {
- enabledRw = val;
- }
-
- } catch (Exception e) {
- Log.e(TAG, ERROR_LOG, e);
- }
- }
- }
-
- private void load_overrides_other_namespace() {
- try {
- Properties properties = DeviceConfig.getProperties("other_namespace");
- disabledRwInOtherNamespace =
- properties.getBoolean(Flags.FLAG_DISABLED_RW_IN_OTHER_NAMESPACE, false);
- } catch (NullPointerException e) {
- throw new RuntimeException(
- "Cannot read value from namespace other_namespace "
- + "from DeviceConfig. It could be that the code using flag "
- + "executed before SettingsProvider initialization. Please use "
- + "fixed read-only flag by adding is_fixed_read_only: true in "
- + "flag declaration.",
- e
- );
- }
- other_namespace_is_cached = true;
- init();
- if (readFromNewStorage && reader != null) {
- boolean val;
- try {
- val = reader.getBooleanFlagValue(3);
- if (val == disabledRwInOtherNamespace) {
- Log.i(TAG, String.format(SUCCESS_LOG, "disabledRwInOtherNamespace"));
- } else {
- Log.i(TAG, String.format(MISMATCH_LOG, "disabledRwInOtherNamespace", val, disabledRwInOtherNamespace));
- }
-
- if (useNewStorageValueAndDiscardOld) {
- disabledRwInOtherNamespace = val;
- }
-
- } catch (Exception e) {
- Log.e(TAG, ERROR_LOG, e);
- }
- }
- }"# + expected_featureflagsmpl_content_2;
-
- let mut file_set = HashMap::from([
- ("com/android/aconfig/test/Flags.java", expect_flags_content.as_str()),
- (
- "com/android/aconfig/test/FeatureFlagsImpl.java",
- &expect_featureflagsimpl_content_new,
- ),
+ ("com/android/aconfig/test/FeatureFlagsImpl.java", expected_featureflagsmpl_content),
("com/android/aconfig/test/FeatureFlags.java", EXPECTED_FEATUREFLAGS_COMMON_CONTENT),
(
"com/android/aconfig/test/CustomFeatureFlags.java",
@@ -916,7 +768,6 @@
private static boolean enabledFixedRoExported = false;
private static boolean enabledRoExported = false;
-
private void load_overrides_aconfig_test() {
try {
Properties properties = DeviceConfig.getProperties("aconfig_test");
@@ -941,21 +792,21 @@
@Override
public boolean disabledRwExported() {
if (!aconfig_test_is_cached) {
- load_overrides_aconfig_test();
+ load_overrides_aconfig_test();
}
return disabledRwExported;
}
@Override
public boolean enabledFixedRoExported() {
if (!aconfig_test_is_cached) {
- load_overrides_aconfig_test();
+ load_overrides_aconfig_test();
}
return enabledFixedRoExported;
}
@Override
public boolean enabledRoExported() {
if (!aconfig_test_is_cached) {
- load_overrides_aconfig_test();
+ load_overrides_aconfig_test();
}
return enabledRoExported;
}
diff --git a/tools/aconfig/aconfig/src/codegen/rust.rs b/tools/aconfig/aconfig/src/codegen/rust.rs
index 1292e0a..7bc34d6 100644
--- a/tools/aconfig/aconfig/src/codegen/rust.rs
+++ b/tools/aconfig/aconfig/src/codegen/rust.rs
@@ -113,41 +113,35 @@
use aconfig_storage_read_api::{Mmap, AconfigStorageError, StorageFileType, PackageReadContext, get_mapped_storage_file, get_boolean_flag_value, get_package_read_context};
use std::path::Path;
use std::io::Write;
+use std::sync::LazyLock;
use log::{log, LevelFilter, Level};
-static STORAGE_MIGRATION_MARKER_FILE: &str =
- "/metadata/aconfig_test_missions/mission_1";
-static MIGRATION_LOG_TAG: &str = "AconfigTestMission1";
-
/// flag provider
pub struct FlagProvider;
-lazy_static::lazy_static! {
/// flag value cache for disabled_rw
- static ref CACHED_disabled_rw: bool = flags_rust::GetServerConfigurableFlag(
+ static CACHED_disabled_rw: LazyLock<bool> = LazyLock::new(|| flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.disabled_rw",
- "false") == "true";
+ "false") == "true");
/// flag value cache for disabled_rw_exported
- static ref CACHED_disabled_rw_exported: bool = flags_rust::GetServerConfigurableFlag(
+ static CACHED_disabled_rw_exported: LazyLock<bool> = LazyLock::new(|| flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.disabled_rw_exported",
- "false") == "true";
+ "false") == "true");
/// flag value cache for disabled_rw_in_other_namespace
- static ref CACHED_disabled_rw_in_other_namespace: bool = flags_rust::GetServerConfigurableFlag(
+ static CACHED_disabled_rw_in_other_namespace: LazyLock<bool> = LazyLock::new(|| flags_rust::GetServerConfigurableFlag(
"aconfig_flags.other_namespace",
"com.android.aconfig.test.disabled_rw_in_other_namespace",
- "false") == "true";
+ "false") == "true");
/// flag value cache for enabled_rw
- static ref CACHED_enabled_rw: bool = flags_rust::GetServerConfigurableFlag(
+ static CACHED_enabled_rw: LazyLock<bool> = LazyLock::new(|| flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.enabled_rw",
- "true") == "true";
-
-}
+ "true") == "true");
impl FlagProvider {
/// query flag disabled_ro
@@ -259,291 +253,202 @@
use aconfig_storage_read_api::{Mmap, AconfigStorageError, StorageFileType, PackageReadContext, get_mapped_storage_file, get_boolean_flag_value, get_package_read_context};
use std::path::Path;
use std::io::Write;
+use std::sync::LazyLock;
use log::{log, LevelFilter, Level};
-static STORAGE_MIGRATION_MARKER_FILE: &str =
- "/metadata/aconfig_test_missions/mission_1";
-static MIGRATION_LOG_TAG: &str = "AconfigTestMission1";
-
/// flag provider
pub struct FlagProvider;
-lazy_static::lazy_static! {
+static READ_FROM_NEW_STORAGE: LazyLock<bool> = LazyLock::new(|| unsafe {
+ Path::new("/metadata/aconfig/boot/enable_only_new_storage").exists()
+});
- static ref PACKAGE_OFFSET: Result<Option<u32>, AconfigStorageError> = unsafe {
- get_mapped_storage_file("system", StorageFileType::PackageMap)
- .and_then(|package_map| get_package_read_context(&package_map, "com.android.aconfig.test"))
- .map(|context| context.map(|c| c.boolean_start_index))
- };
+static PACKAGE_OFFSET: LazyLock<Result<Option<u32>, AconfigStorageError>> = LazyLock::new(|| unsafe {
+ get_mapped_storage_file("system", StorageFileType::PackageMap)
+ .and_then(|package_map| get_package_read_context(&package_map, "com.android.aconfig.test"))
+ .map(|context| context.map(|c| c.boolean_start_index))
+});
- static ref FLAG_VAL_MAP: Result<Mmap, AconfigStorageError> = unsafe {
- get_mapped_storage_file("system", StorageFileType::FlagVal)
- };
- /// flag value cache for disabled_rw
+static FLAG_VAL_MAP: LazyLock<Result<Mmap, AconfigStorageError>> = LazyLock::new(|| unsafe {
+ get_mapped_storage_file("system", StorageFileType::FlagVal)
+});
- static ref CACHED_disabled_rw: bool = {
- let result = flags_rust::GetServerConfigurableFlag(
+/// flag value cache for disabled_rw
+static CACHED_disabled_rw: LazyLock<bool> = LazyLock::new(|| {
+ if *READ_FROM_NEW_STORAGE {
+ // This will be called multiple times. Subsequent calls after the first are noops.
+ logger::init(
+ logger::Config::default()
+ .with_tag_on_device("aconfig_rust_codegen")
+ .with_max_level(LevelFilter::Info));
+
+ let flag_value_result = FLAG_VAL_MAP
+ .as_ref()
+ .map_err(|err| format!("failed to get flag val map: {err}"))
+ .and_then(|flag_val_map| {
+ PACKAGE_OFFSET
+ .as_ref()
+ .map_err(|err| format!("failed to get package read offset: {err}"))
+ .and_then(|package_offset| {
+ match package_offset {
+ Some(offset) => {
+ get_boolean_flag_value(&flag_val_map, offset + 1)
+ .map_err(|err| format!("failed to get flag: {err}"))
+ },
+ None => Err("no context found for package 'com.android.aconfig.test'".to_string())
+ }
+ })
+ });
+
+ match flag_value_result {
+ Ok(flag_value) => {
+ return flag_value;
+ },
+ Err(err) => {
+ log!(Level::Error, "aconfig_rust_codegen: error: {err}");
+ panic!("failed to read flag value: {err}");
+ }
+ }
+ } else {
+ flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.disabled_rw",
- "false") == "true";
+ "false") == "true"
+ }
+});
- let use_new_storage_value = flags_rust::GetServerConfigurableFlag(
- "aconfig_flags.core_experiments_team_internal",
- "com.android.providers.settings.use_new_storage_value",
- "false") == "true";
+/// flag value cache for disabled_rw_exported
+static CACHED_disabled_rw_exported: LazyLock<bool> = LazyLock::new(|| {
+ if *READ_FROM_NEW_STORAGE {
+ // This will be called multiple times. Subsequent calls after the first are noops.
+ logger::init(
+ logger::Config::default()
+ .with_tag_on_device("aconfig_rust_codegen")
+ .with_max_level(LevelFilter::Info));
- if Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- // This will be called multiple times. Subsequent calls after the first are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info));
+ let flag_value_result = FLAG_VAL_MAP
+ .as_ref()
+ .map_err(|err| format!("failed to get flag val map: {err}"))
+ .and_then(|flag_val_map| {
+ PACKAGE_OFFSET
+ .as_ref()
+ .map_err(|err| format!("failed to get package read offset: {err}"))
+ .and_then(|package_offset| {
+ match package_offset {
+ Some(offset) => {
+ get_boolean_flag_value(&flag_val_map, offset + 2)
+ .map_err(|err| format!("failed to get flag: {err}"))
+ },
+ None => Err("no context found for package 'com.android.aconfig.test'".to_string())
+ }
+ })
+ });
- let aconfig_storage_result = FLAG_VAL_MAP
- .as_ref()
- .map_err(|err| format!("failed to get flag val map: {err}"))
- .and_then(|flag_val_map| {
- PACKAGE_OFFSET
- .as_ref()
- .map_err(|err| format!("failed to get package read offset: {err}"))
- .and_then(|package_offset| {
- match package_offset {
- Some(offset) => {
- get_boolean_flag_value(&flag_val_map, offset + 1)
- .map_err(|err| format!("failed to get flag: {err}"))
- },
- None => Err("no context found for package 'com.android.aconfig.test'".to_string())
- }
- })
- });
-
- match aconfig_storage_result {
- Ok(storage_result) if storage_result == result => {
- if use_new_storage_value {
- return storage_result;
- } else {
- return result;
- }
- },
- Ok(storage_result) => {
- log!(Level::Error, "AconfigTestMission1: error: mismatch for flag 'disabled_rw'. Legacy storage was {result}, new storage was {storage_result}");
- if use_new_storage_value {
- return storage_result;
- } else {
- return result;
- }
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: {err}");
- if use_new_storage_value {
- panic!("failed to read flag value: {err}");
- }
- }
+ match flag_value_result {
+ Ok(flag_value) => {
+ return flag_value;
+ },
+ Err(err) => {
+ log!(Level::Error, "aconfig_rust_codegen: error: {err}");
+ panic!("failed to read flag value: {err}");
}
}
-
- result
- };
-
- /// flag value cache for disabled_rw_exported
-
- static ref CACHED_disabled_rw_exported: bool = {
- let result = flags_rust::GetServerConfigurableFlag(
+ } else {
+ flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.disabled_rw_exported",
- "false") == "true";
+ "false") == "true"
+ }
+});
- let use_new_storage_value = flags_rust::GetServerConfigurableFlag(
- "aconfig_flags.core_experiments_team_internal",
- "com.android.providers.settings.use_new_storage_value",
- "false") == "true";
+/// flag value cache for disabled_rw_in_other_namespace
+static CACHED_disabled_rw_in_other_namespace: LazyLock<bool> = LazyLock::new(|| {
+ if *READ_FROM_NEW_STORAGE {
+ // This will be called multiple times. Subsequent calls after the first are noops.
+ logger::init(
+ logger::Config::default()
+ .with_tag_on_device("aconfig_rust_codegen")
+ .with_max_level(LevelFilter::Info));
- if Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- // This will be called multiple times. Subsequent calls after the first are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info));
+ let flag_value_result = FLAG_VAL_MAP
+ .as_ref()
+ .map_err(|err| format!("failed to get flag val map: {err}"))
+ .and_then(|flag_val_map| {
+ PACKAGE_OFFSET
+ .as_ref()
+ .map_err(|err| format!("failed to get package read offset: {err}"))
+ .and_then(|package_offset| {
+ match package_offset {
+ Some(offset) => {
+ get_boolean_flag_value(&flag_val_map, offset + 3)
+ .map_err(|err| format!("failed to get flag: {err}"))
+ },
+ None => Err("no context found for package 'com.android.aconfig.test'".to_string())
+ }
+ })
+ });
- let aconfig_storage_result = FLAG_VAL_MAP
- .as_ref()
- .map_err(|err| format!("failed to get flag val map: {err}"))
- .and_then(|flag_val_map| {
- PACKAGE_OFFSET
- .as_ref()
- .map_err(|err| format!("failed to get package read offset: {err}"))
- .and_then(|package_offset| {
- match package_offset {
- Some(offset) => {
- get_boolean_flag_value(&flag_val_map, offset + 2)
- .map_err(|err| format!("failed to get flag: {err}"))
- },
- None => Err("no context found for package 'com.android.aconfig.test'".to_string())
- }
- })
- });
-
- match aconfig_storage_result {
- Ok(storage_result) if storage_result == result => {
- if use_new_storage_value {
- return storage_result;
- } else {
- return result;
- }
- },
- Ok(storage_result) => {
- log!(Level::Error, "AconfigTestMission1: error: mismatch for flag 'disabled_rw_exported'. Legacy storage was {result}, new storage was {storage_result}");
- if use_new_storage_value {
- return storage_result;
- } else {
- return result;
- }
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: {err}");
- if use_new_storage_value {
- panic!("failed to read flag value: {err}");
- }
- }
+ match flag_value_result {
+ Ok(flag_value) => {
+ return flag_value;
+ },
+ Err(err) => {
+ log!(Level::Error, "aconfig_rust_codegen: error: {err}");
+ panic!("failed to read flag value: {err}");
}
}
-
- result
- };
-
- /// flag value cache for disabled_rw_in_other_namespace
-
- static ref CACHED_disabled_rw_in_other_namespace: bool = {
- let result = flags_rust::GetServerConfigurableFlag(
+ } else {
+ flags_rust::GetServerConfigurableFlag(
"aconfig_flags.other_namespace",
"com.android.aconfig.test.disabled_rw_in_other_namespace",
- "false") == "true";
+ "false") == "true"
+ }
+});
- let use_new_storage_value = flags_rust::GetServerConfigurableFlag(
- "aconfig_flags.core_experiments_team_internal",
- "com.android.providers.settings.use_new_storage_value",
- "false") == "true";
- if Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- // This will be called multiple times. Subsequent calls after the first are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info));
+/// flag value cache for enabled_rw
+static CACHED_enabled_rw: LazyLock<bool> = LazyLock::new(|| {
+ if *READ_FROM_NEW_STORAGE {
+ // This will be called multiple times. Subsequent calls after the first are noops.
+ logger::init(
+ logger::Config::default()
+ .with_tag_on_device("aconfig_rust_codegen")
+ .with_max_level(LevelFilter::Info));
- let aconfig_storage_result = FLAG_VAL_MAP
- .as_ref()
- .map_err(|err| format!("failed to get flag val map: {err}"))
- .and_then(|flag_val_map| {
- PACKAGE_OFFSET
- .as_ref()
- .map_err(|err| format!("failed to get package read offset: {err}"))
- .and_then(|package_offset| {
- match package_offset {
- Some(offset) => {
- get_boolean_flag_value(&flag_val_map, offset + 3)
- .map_err(|err| format!("failed to get flag: {err}"))
- },
- None => Err("no context found for package 'com.android.aconfig.test'".to_string())
- }
- })
- });
+ let flag_value_result = FLAG_VAL_MAP
+ .as_ref()
+ .map_err(|err| format!("failed to get flag val map: {err}"))
+ .and_then(|flag_val_map| {
+ PACKAGE_OFFSET
+ .as_ref()
+ .map_err(|err| format!("failed to get package read offset: {err}"))
+ .and_then(|package_offset| {
+ match package_offset {
+ Some(offset) => {
+ get_boolean_flag_value(&flag_val_map, offset + 8)
+ .map_err(|err| format!("failed to get flag: {err}"))
+ },
+ None => Err("no context found for package 'com.android.aconfig.test'".to_string())
+ }
+ })
+ });
- match aconfig_storage_result {
- Ok(storage_result) if storage_result == result => {
- if use_new_storage_value {
- return storage_result;
- } else {
- return result;
- }
- },
- Ok(storage_result) => {
- log!(Level::Error, "AconfigTestMission1: error: mismatch for flag 'disabled_rw_in_other_namespace'. Legacy storage was {result}, new storage was {storage_result}");
- if use_new_storage_value {
- return storage_result;
- } else {
- return result;
- }
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: {err}");
- if use_new_storage_value {
- panic!("failed to read flag value: {err}");
- }
- }
+ match flag_value_result {
+ Ok(flag_value) => {
+ return flag_value;
+ },
+ Err(err) => {
+ log!(Level::Error, "aconfig_rust_codegen: error: {err}");
+ panic!("failed to read flag value: {err}");
}
}
-
- result
- };
-
- /// flag value cache for enabled_rw
-
- static ref CACHED_enabled_rw: bool = {
- let result = flags_rust::GetServerConfigurableFlag(
+ } else {
+ flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.enabled_rw",
- "true") == "true";
-
- let use_new_storage_value = flags_rust::GetServerConfigurableFlag(
- "aconfig_flags.core_experiments_team_internal",
- "com.android.providers.settings.use_new_storage_value",
- "false") == "true";
-
- if Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- // This will be called multiple times. Subsequent calls after the first are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info));
-
- let aconfig_storage_result = FLAG_VAL_MAP
- .as_ref()
- .map_err(|err| format!("failed to get flag val map: {err}"))
- .and_then(|flag_val_map| {
- PACKAGE_OFFSET
- .as_ref()
- .map_err(|err| format!("failed to get package read offset: {err}"))
- .and_then(|package_offset| {
- match package_offset {
- Some(offset) => {
- get_boolean_flag_value(&flag_val_map, offset + 8)
- .map_err(|err| format!("failed to get flag: {err}"))
- },
- None => Err("no context found for package 'com.android.aconfig.test'".to_string())
- }
- })
- });
-
- match aconfig_storage_result {
- Ok(storage_result) if storage_result == result => {
- if use_new_storage_value {
- return storage_result;
- } else {
- return result;
- }
- },
- Ok(storage_result) => {
- log!(Level::Error, "AconfigTestMission1: error: mismatch for flag 'enabled_rw'. Legacy storage was {result}, new storage was {storage_result}");
- if use_new_storage_value {
- return storage_result;
- } else {
- return result;
- }
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: {err}");
- if use_new_storage_value {
- panic!("failed to read flag value: {err}");
- }
- }
- }
- }
-
- result
- };
-
-}
+ "true") == "true"
+ }
+});
impl FlagProvider {
@@ -603,65 +508,7 @@
/// query flag disabled_ro
#[inline(always)]
pub fn disabled_ro() -> bool {
-
-
- let result = false;
- if !Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- return result;
- }
-
- // This will be called multiple times. Subsequent calls after the first
- // are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info),
- );
-
- unsafe {
- let package_map = match get_mapped_storage_file("system", StorageFileType::PackageMap) {
- Ok(file) => file,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'disabled_ro': {err}");
- return result;
- }
- };
-
- let package_read_context = match get_package_read_context(&package_map, "com.android.aconfig.test") {
- Ok(Some(context)) => context,
- Ok(None) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'disabled_ro': did not get context");
- return result;
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'disabled_ro': {err}");
- return result;
- }
- };
- let flag_val_map = match get_mapped_storage_file("system", StorageFileType::FlagVal) {
- Ok(val_map) => val_map,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'disabled_ro': {err}");
- return result;
- }
- };
- let value = match get_boolean_flag_value(&flag_val_map, 0 + package_read_context.boolean_start_index) {
- Ok(val) => val,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'disabled_ro': {err}");
- return result;
- }
- };
-
- if result != value {
- log!(Level::Error, "AconfigTestMission1: error: flag mismatch for 'disabled_ro'. Legacy storage was {result}, new storage was {value}");
- } else {
- let default_value = false;
- }
- }
-
- result
-
+ false
}
/// query flag disabled_rw
@@ -685,257 +532,25 @@
/// query flag enabled_fixed_ro
#[inline(always)]
pub fn enabled_fixed_ro() -> bool {
-
-
- let result = true;
- if !Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- return result;
- }
-
- // This will be called multiple times. Subsequent calls after the first
- // are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info),
- );
-
- unsafe {
- let package_map = match get_mapped_storage_file("system", StorageFileType::PackageMap) {
- Ok(file) => file,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro': {err}");
- return result;
- }
- };
-
- let package_read_context = match get_package_read_context(&package_map, "com.android.aconfig.test") {
- Ok(Some(context)) => context,
- Ok(None) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro': did not get context");
- return result;
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro': {err}");
- return result;
- }
- };
- let flag_val_map = match get_mapped_storage_file("system", StorageFileType::FlagVal) {
- Ok(val_map) => val_map,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro': {err}");
- return result;
- }
- };
- let value = match get_boolean_flag_value(&flag_val_map, 4 + package_read_context.boolean_start_index) {
- Ok(val) => val,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro': {err}");
- return result;
- }
- };
-
- if result != value {
- log!(Level::Error, "AconfigTestMission1: error: flag mismatch for 'enabled_fixed_ro'. Legacy storage was {result}, new storage was {value}");
- } else {
- let default_value = true;
- }
- }
-
- result
-
+ true
}
/// query flag enabled_fixed_ro_exported
#[inline(always)]
pub fn enabled_fixed_ro_exported() -> bool {
-
-
- let result = true;
- if !Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- return result;
- }
-
- // This will be called multiple times. Subsequent calls after the first
- // are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info),
- );
-
- unsafe {
- let package_map = match get_mapped_storage_file("system", StorageFileType::PackageMap) {
- Ok(file) => file,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro_exported': {err}");
- return result;
- }
- };
-
- let package_read_context = match get_package_read_context(&package_map, "com.android.aconfig.test") {
- Ok(Some(context)) => context,
- Ok(None) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro_exported': did not get context");
- return result;
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro_exported': {err}");
- return result;
- }
- };
- let flag_val_map = match get_mapped_storage_file("system", StorageFileType::FlagVal) {
- Ok(val_map) => val_map,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro_exported': {err}");
- return result;
- }
- };
- let value = match get_boolean_flag_value(&flag_val_map, 5 + package_read_context.boolean_start_index) {
- Ok(val) => val,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_fixed_ro_exported': {err}");
- return result;
- }
- };
-
- if result != value {
- log!(Level::Error, "AconfigTestMission1: error: flag mismatch for 'enabled_fixed_ro_exported'. Legacy storage was {result}, new storage was {value}");
- } else {
- let default_value = true;
- }
- }
-
- result
-
+ true
}
/// query flag enabled_ro
#[inline(always)]
pub fn enabled_ro() -> bool {
-
-
- let result = true;
- if !Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- return result;
- }
-
- // This will be called multiple times. Subsequent calls after the first
- // are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info),
- );
-
- unsafe {
- let package_map = match get_mapped_storage_file("system", StorageFileType::PackageMap) {
- Ok(file) => file,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro': {err}");
- return result;
- }
- };
-
- let package_read_context = match get_package_read_context(&package_map, "com.android.aconfig.test") {
- Ok(Some(context)) => context,
- Ok(None) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro': did not get context");
- return result;
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro': {err}");
- return result;
- }
- };
- let flag_val_map = match get_mapped_storage_file("system", StorageFileType::FlagVal) {
- Ok(val_map) => val_map,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro': {err}");
- return result;
- }
- };
- let value = match get_boolean_flag_value(&flag_val_map, 6 + package_read_context.boolean_start_index) {
- Ok(val) => val,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro': {err}");
- return result;
- }
- };
-
- if result != value {
- log!(Level::Error, "AconfigTestMission1: error: flag mismatch for 'enabled_ro'. Legacy storage was {result}, new storage was {value}");
- } else {
- let default_value = true;
- }
- }
-
- result
-
+ true
}
/// query flag enabled_ro_exported
#[inline(always)]
pub fn enabled_ro_exported() -> bool {
-
-
- let result = true;
- if !Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() {
- return result;
- }
-
- // This will be called multiple times. Subsequent calls after the first
- // are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info),
- );
-
- unsafe {
- let package_map = match get_mapped_storage_file("system", StorageFileType::PackageMap) {
- Ok(file) => file,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro_exported': {err}");
- return result;
- }
- };
-
- let package_read_context = match get_package_read_context(&package_map, "com.android.aconfig.test") {
- Ok(Some(context)) => context,
- Ok(None) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro_exported': did not get context");
- return result;
- },
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro_exported': {err}");
- return result;
- }
- };
- let flag_val_map = match get_mapped_storage_file("system", StorageFileType::FlagVal) {
- Ok(val_map) => val_map,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro_exported': {err}");
- return result;
- }
- };
- let value = match get_boolean_flag_value(&flag_val_map, 7 + package_read_context.boolean_start_index) {
- Ok(val) => val,
- Err(err) => {
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag 'enabled_ro_exported': {err}");
- return result;
- }
- };
-
- if result != value {
- log!(Level::Error, "AconfigTestMission1: error: flag mismatch for 'enabled_ro_exported'. Legacy storage was {result}, new storage was {value}");
- } else {
- let default_value = true;
- }
- }
-
- result
-
+ true
}
/// query flag enabled_rw
@@ -1207,35 +822,29 @@
use aconfig_storage_read_api::{Mmap, AconfigStorageError, StorageFileType, PackageReadContext, get_mapped_storage_file, get_boolean_flag_value, get_package_read_context};
use std::path::Path;
use std::io::Write;
+use std::sync::LazyLock;
use log::{log, LevelFilter, Level};
-static STORAGE_MIGRATION_MARKER_FILE: &str =
- "/metadata/aconfig_test_missions/mission_1";
-static MIGRATION_LOG_TAG: &str = "AconfigTestMission1";
-
/// flag provider
pub struct FlagProvider;
-lazy_static::lazy_static! {
/// flag value cache for disabled_rw_exported
- static ref CACHED_disabled_rw_exported: bool = flags_rust::GetServerConfigurableFlag(
+ static CACHED_disabled_rw_exported: LazyLock<bool> = LazyLock::new(|| flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.disabled_rw_exported",
- "false") == "true";
+ "false") == "true");
/// flag value cache for enabled_fixed_ro_exported
- static ref CACHED_enabled_fixed_ro_exported: bool = flags_rust::GetServerConfigurableFlag(
+ static CACHED_enabled_fixed_ro_exported: LazyLock<bool> = LazyLock::new(|| flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.enabled_fixed_ro_exported",
- "false") == "true";
+ "false") == "true");
/// flag value cache for enabled_ro_exported
- static ref CACHED_enabled_ro_exported: bool = flags_rust::GetServerConfigurableFlag(
+ static CACHED_enabled_ro_exported: LazyLock<bool> = LazyLock::new(|| flags_rust::GetServerConfigurableFlag(
"aconfig_flags.aconfig_test",
"com.android.aconfig.test.enabled_ro_exported",
- "false") == "true";
-
-}
+ "false") == "true");
impl FlagProvider {
/// query flag disabled_rw_exported
@@ -1281,12 +890,9 @@
use aconfig_storage_read_api::{Mmap, AconfigStorageError, StorageFileType, PackageReadContext, get_mapped_storage_file, get_boolean_flag_value, get_package_read_context};
use std::path::Path;
use std::io::Write;
+use std::sync::LazyLock;
use log::{log, LevelFilter, Level};
-static STORAGE_MIGRATION_MARKER_FILE: &str =
- "/metadata/aconfig_test_missions/mission_1";
-static MIGRATION_LOG_TAG: &str = "AconfigTestMission1";
-
/// flag provider
pub struct FlagProvider;
diff --git a/tools/aconfig/aconfig/src/commands.rs b/tools/aconfig/aconfig/src/commands.rs
index 59f0662..797a893 100644
--- a/tools/aconfig/aconfig/src/commands.rs
+++ b/tools/aconfig/aconfig/src/commands.rs
@@ -17,7 +17,8 @@
use anyhow::{bail, ensure, Context, Result};
use itertools::Itertools;
use protobuf::Message;
-use std::collections::HashMap;
+use std::collections::{BTreeMap, HashMap};
+use std::hash::Hasher;
use std::io::Read;
use std::path::PathBuf;
@@ -31,6 +32,7 @@
ParsedFlagExt, ProtoFlagMetadata, ProtoFlagPermission, ProtoFlagState, ProtoParsedFlag,
ProtoParsedFlags, ProtoTracepoint,
};
+use aconfig_storage_file::sip_hasher13::SipHasher13;
use aconfig_storage_file::StorageFileType;
pub struct Input {
@@ -77,8 +79,18 @@
.read_to_string(&mut contents)
.with_context(|| format!("failed to read {}", input.source))?;
- let flag_declarations = aconfig_protos::flag_declarations::try_from_text_proto(&contents)
- .with_context(|| input.error_context())?;
+ let mut flag_declarations =
+ aconfig_protos::flag_declarations::try_from_text_proto(&contents)
+ .with_context(|| input.error_context())?;
+
+ // system_ext flags should be treated as system flags as we are combining /system_ext
+ // and /system as one container
+ // TODO: remove this logic when we start enforcing that system_ext cannot be set as
+ // container in aconfig declaration files.
+ if flag_declarations.container() == "system_ext" {
+ flag_declarations.set_container(String::from("system"));
+ }
+
ensure!(
package == flag_declarations.package(),
"failed to parse {}: expected package {}, got {}",
@@ -410,12 +422,43 @@
Ok(flag_ids)
}
+#[allow(dead_code)] // TODO: b/316357686 - Use fingerprint in codegen to
+ // protect hardcoded offset reads.
+pub fn compute_flag_offsets_fingerprint(flags_map: &HashMap<String, u16>) -> Result<u64> {
+ let mut hasher = SipHasher13::new();
+
+ // Need to sort to ensure the data is added to the hasher in the same order
+ // each run.
+ let sorted_map: BTreeMap<&String, &u16> = flags_map.iter().collect();
+
+ for (flag, offset) in sorted_map {
+ // See https://docs.rs/siphasher/latest/siphasher/#note for use of write
+ // over write_i16. Similarly, use to_be_bytes rather than to_ne_bytes to
+ // ensure consistency.
+ hasher.write(flag.as_bytes());
+ hasher.write(&offset.to_be_bytes());
+ }
+ Ok(hasher.finish())
+}
+
#[cfg(test)]
mod tests {
use super::*;
use aconfig_protos::ProtoFlagPurpose;
#[test]
+ fn test_offset_fingerprint() {
+ let parsed_flags = crate::test::parse_test_flags();
+ let package = find_unique_package(&parsed_flags.parsed_flag).unwrap().to_string();
+ let flag_ids = assign_flag_ids(&package, parsed_flags.parsed_flag.iter()).unwrap();
+ let expected_fingerprint = 10709892481002252132u64;
+
+ let hash_result = compute_flag_offsets_fingerprint(&flag_ids);
+
+ assert_eq!(hash_result.unwrap(), expected_fingerprint);
+ }
+
+ #[test]
fn test_parse_flags() {
let parsed_flags = crate::test::parse_test_flags(); // calls parse_flags
aconfig_protos::parsed_flags::verify_fields(&parsed_flags).unwrap();
diff --git a/tools/aconfig/aconfig/src/storage/flag_info.rs b/tools/aconfig/aconfig/src/storage/flag_info.rs
new file mode 100644
index 0000000..04e2b93
--- /dev/null
+++ b/tools/aconfig/aconfig/src/storage/flag_info.rs
@@ -0,0 +1,86 @@
+/*
+ * Copyright (C) 2024 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+use crate::commands::assign_flag_ids;
+use crate::storage::FlagPackage;
+use aconfig_protos::ProtoFlagPermission;
+use aconfig_storage_file::{
+ FlagInfoHeader, FlagInfoList, FlagInfoNode, StorageFileType, FILE_VERSION,
+};
+use anyhow::{anyhow, Result};
+
+fn new_header(container: &str, num_flags: u32) -> FlagInfoHeader {
+ FlagInfoHeader {
+ version: FILE_VERSION,
+ container: String::from(container),
+ file_type: StorageFileType::FlagInfo as u8,
+ file_size: 0,
+ num_flags,
+ boolean_flag_offset: 0,
+ }
+}
+
+pub fn create_flag_info(container: &str, packages: &[FlagPackage]) -> Result<FlagInfoList> {
+ // create list
+ let num_flags = packages.iter().map(|pkg| pkg.boolean_flags.len() as u32).sum();
+
+ let mut is_flag_rw = vec![false; num_flags as usize];
+ for pkg in packages.iter() {
+ let start_index = pkg.boolean_start_index as usize;
+ let flag_ids = assign_flag_ids(pkg.package_name, pkg.boolean_flags.iter().copied())?;
+ for pf in pkg.boolean_flags.iter() {
+ let fid = flag_ids
+ .get(pf.name())
+ .ok_or(anyhow!(format!("missing flag id for {}", pf.name())))?;
+ is_flag_rw[start_index + (*fid as usize)] =
+ pf.permission() == ProtoFlagPermission::READ_WRITE;
+ }
+ }
+
+ let mut list = FlagInfoList {
+ header: new_header(container, num_flags),
+ nodes: is_flag_rw.iter().map(|&rw| FlagInfoNode::create(rw)).collect(),
+ };
+
+ // initialize all header fields
+ list.header.boolean_flag_offset = list.header.into_bytes().len() as u32;
+ let bytes_per_node = FlagInfoNode::create(false).into_bytes().len() as u32;
+ list.header.file_size = list.header.boolean_flag_offset + num_flags * bytes_per_node;
+
+ Ok(list)
+}
+
+#[cfg(test)]
+mod tests {
+ use super::*;
+ use crate::storage::{group_flags_by_package, tests::parse_all_test_flags};
+
+ pub fn create_test_flag_info_list_from_source() -> Result<FlagInfoList> {
+ let caches = parse_all_test_flags();
+ let packages = group_flags_by_package(caches.iter());
+ create_flag_info("mockup", &packages)
+ }
+
+ #[test]
+ // this test point locks down the flag info creation and each field
+ fn test_list_contents() {
+ let flag_info_list = create_test_flag_info_list_from_source();
+ assert!(flag_info_list.is_ok());
+ let expected_flag_info_list =
+ aconfig_storage_file::test_utils::create_test_flag_info_list();
+ assert_eq!(flag_info_list.unwrap(), expected_flag_info_list);
+ }
+}
diff --git a/tools/aconfig/aconfig/src/storage/mod.rs b/tools/aconfig/aconfig/src/storage/mod.rs
index 73339f2..1d8dcfc 100644
--- a/tools/aconfig/aconfig/src/storage/mod.rs
+++ b/tools/aconfig/aconfig/src/storage/mod.rs
@@ -14,15 +14,16 @@
* limitations under the License.
*/
+pub mod flag_info;
pub mod flag_table;
pub mod flag_value;
pub mod package_table;
-use anyhow::{anyhow, Result};
+use anyhow::Result;
use std::collections::{HashMap, HashSet};
use crate::storage::{
- flag_table::create_flag_table, flag_value::create_flag_value,
+ flag_info::create_flag_info, flag_table::create_flag_table, flag_value::create_flag_value,
package_table::create_package_table,
};
use aconfig_protos::{ProtoParsedFlag, ProtoParsedFlags};
@@ -106,7 +107,10 @@
let flag_value = create_flag_value(container, &packages)?;
Ok(flag_value.into_bytes())
}
- _ => Err(anyhow!("aconfig does not support the creation of this storage file type")),
+ StorageFileType::FlagInfo => {
+ let flag_info = create_flag_info(container, &packages)?;
+ Ok(flag_info.into_bytes())
+ }
}
}
diff --git a/tools/aconfig/aconfig/templates/FeatureFlagsImpl.java.template b/tools/aconfig/aconfig/templates/FeatureFlagsImpl.java.template
index 9970b1f..d1cf191 100644
--- a/tools/aconfig/aconfig/templates/FeatureFlagsImpl.java.template
+++ b/tools/aconfig/aconfig/templates/FeatureFlagsImpl.java.template
@@ -9,56 +9,50 @@
import android.provider.DeviceConfig;
import android.provider.DeviceConfig.Properties;
-
{{ -if not library_exported }}
-{{ -if allow_instrumentation }}
import android.aconfig.storage.StorageInternalReader;
-import android.util.Log;
-{{ -endif }}
+import java.nio.file.Files;
+import java.nio.file.Paths;
{{ -endif }}
{{ -endif }}
/** @hide */
public final class FeatureFlagsImpl implements FeatureFlags \{
{{ -if runtime_lookup_required }}
+{{ -if not library_exported }}
+ private static final boolean isReadFromNew = Files.exists(Paths.get("/metadata/aconfig/boot/enable_only_new_storage"));
+ private static volatile boolean isCached = false;
+{{ -endif }}
{{ -for namespace_with_flags in namespace_flags }}
private static volatile boolean {namespace_with_flags.namespace}_is_cached = false;
{{ -endfor- }}
{{ for flag in flag_elements }}
-{{- if flag.is_read_write }}
+{{ -if flag.is_read_write }}
private static boolean {flag.method_name} = {flag.default_value};
{{ -endif }}
{{ -endfor }}
-{{ -if not library_exported }}
-{{ -if allow_instrumentation }}
- StorageInternalReader reader;
- boolean readFromNewStorage;
- boolean useNewStorageValueAndDiscardOld = false;
-
- private final static String TAG = "AconfigJavaCodegen";
- private final static String SUCCESS_LOG = "success: %s value matches";
- private final static String MISMATCH_LOG = "error: %s value mismatch, new storage value is %s, old storage value is %s";
- private final static String ERROR_LOG = "error: failed to read flag value";
-
+{{ if not library_exported }}
private void init() \{
- if (reader != null) return;
- if (DeviceConfig.getBoolean("core_experiments_team_internal", "com.android.providers.settings.storage_test_mission_1", false)) \{
- readFromNewStorage = true;
- try \{
- reader = new StorageInternalReader("{container}", "{package_name}");
- } catch (Exception e) \{
- reader = null;
- }
+ StorageInternalReader reader = null;
+ try \{
+ reader = new StorageInternalReader("{container}", "{package_name}");
+{{ for namespace_with_flags in namespace_flags }}
+{{ -for flag in namespace_with_flags.flags }}
+{{ if flag.is_read_write }}
+ {flag.method_name} = reader.getBooleanFlagValue({flag.flag_offset});
+{{ endif }}
+{{ -endfor }}
+{{ -endfor }}
+ } catch (Exception e) \{
+ throw new RuntimeException("Cannot read flag in codegen", e);
}
-
- useNewStorageValueAndDiscardOld =
- DeviceConfig.getBoolean("core_experiments_team_internal", "com.android.providers.settings.use_new_storage_value", false);
+ isCached = true;
}
+{{ endif }}
-{{ -endif }}
-{{ -endif }}
+
{{ for namespace_with_flags in namespace_flags }}
private void load_overrides_{namespace_with_flags.namespace}() \{
try \{
@@ -80,36 +74,9 @@
);
}
{namespace_with_flags.namespace}_is_cached = true;
-{{ -if not library_exported }}
-{{ -if allow_instrumentation }}
- init();
- if (readFromNewStorage && reader != null) \{
- boolean val;
- try \{
-{{ -for flag in namespace_with_flags.flags }}
-{{ -if flag.is_read_write }}
-
- val = reader.getBooleanFlagValue({flag.flag_offset});
- if (val == {flag.method_name}) \{
- Log.i(TAG, String.format(SUCCESS_LOG, "{flag.method_name}"));
- } else \{
- Log.i(TAG, String.format(MISMATCH_LOG, "{flag.method_name}", val, {flag.method_name}));
- }
-
- if (useNewStorageValueAndDiscardOld) \{
- {flag.method_name} = val;
- }
-
-{{ -endif }}
-{{ -endfor }}
- } catch (Exception e) \{
- Log.e(TAG, ERROR_LOG, e);
- }
- }
-{{ -endif }}
-{{ -endif }}
}
{{ endfor- }}
+
{{ -endif }}{#- end of runtime_lookup_required #}
{{ -for flag in flag_elements }}
@Override
@@ -118,19 +85,31 @@
@UnsupportedAppUsage
{{ -endif }}
public boolean {flag.method_name}() \{
+{{ -if not library_exported }}
{{ -if flag.is_read_write }}
- if (!{flag.device_config_namespace}_is_cached) \{
- load_overrides_{flag.device_config_namespace}();
+ if (isReadFromNew) \{
+ if (!isCached) \{
+ init();
+ }
+ } else \{
+ if (!{flag.device_config_namespace}_is_cached) \{
+ load_overrides_{flag.device_config_namespace}();
+ }
}
return {flag.method_name};
{{ -else }}
return {flag.default_value};
{{ -endif }}
+{{ else }}
+ if (!{flag.device_config_namespace}_is_cached) \{
+ load_overrides_{flag.device_config_namespace}();
+ }
+ return {flag.method_name};
+{{ -endif }}
}
{{ endfor }}
}
-{{ else }}
-{#- Generate only stub if in test mode #}
+{{ else }} {#- Generate only stub if in test mode #}
/** @hide */
public final class FeatureFlagsImpl implements FeatureFlags \{
{{ for flag in flag_elements }}
diff --git a/tools/aconfig/aconfig/templates/cpp_exported_header.template b/tools/aconfig/aconfig/templates/cpp_exported_header.template
index 0f7853e..4643c97 100644
--- a/tools/aconfig/aconfig/templates/cpp_exported_header.template
+++ b/tools/aconfig/aconfig/templates/cpp_exported_header.template
@@ -27,12 +27,13 @@
{{ -for item in class_elements}}
virtual bool {item.flag_name}() = 0;
- {{ -if is_test_mode }}
- virtual void {item.flag_name}(bool val) = 0;
- {{ -endif }}
{{ -endfor }}
{{ -if is_test_mode }}
+ {{ -for item in class_elements}}
+ virtual void {item.flag_name}(bool val) = 0;
+ {{ -endfor }}
+
virtual void reset_flags() \{}
{{ -endif }}
};
diff --git a/tools/aconfig/aconfig/templates/cpp_source_file.template b/tools/aconfig/aconfig/templates/cpp_source_file.template
index b6012e7..852b905 100644
--- a/tools/aconfig/aconfig/templates/cpp_source_file.template
+++ b/tools/aconfig/aconfig/templates/cpp_source_file.template
@@ -2,11 +2,11 @@
{{ if allow_instrumentation }}
{{ if readwrite- }}
-#include <sys/stat.h>
+#include <unistd.h>
#include "aconfig_storage/aconfig_storage_read_api.hpp"
#include <android/log.h>
#define LOG_TAG "aconfig_cpp_codegen"
-#define ALOGI(...) __android_log_print(ANDROID_LOG_INFO, LOG_TAG, __VA_ARGS__)
+#define ALOGE(...) __android_log_print(ANDROID_LOG_ERROR, LOG_TAG, __VA_ARGS__)
{{ -endif }}
{{ endif }}
@@ -76,13 +76,13 @@
: boolean_start_index_()
{{ -endif }}
, flag_value_file_(nullptr)
- , read_from_new_storage_(false)
- , use_new_storage_value(false) \{
+ , read_from_new_storage_(false) \{
- struct stat buffer;
- if (stat("/metadata/aconfig_test_missions/mission_1", &buffer) == 0) \{
+ if (access("/metadata/aconfig/boot/enable_only_new_storage", F_OK) == 0) \{
read_from_new_storage_ = true;
- } else \{
+ }
+
+ if (!read_from_new_storage_) \{
return;
}
@@ -90,15 +90,13 @@
"{container}",
aconfig_storage::StorageFileType::package_map);
if (!package_map_file.ok()) \{
- ALOGI("error: failed to get package map file: %s", package_map_file.error().c_str());
- return;
+ ALOGE("error: failed to get package map file: %s", package_map_file.error().c_str());
}
auto context = aconfig_storage::get_package_read_context(
**package_map_file, "{package}");
if (!context.ok()) \{
- ALOGI("error: failed to get package read context: %s", context.error().c_str());
- return;
+ ALOGE("error: failed to get package read context: %s", context.error().c_str());
}
// cache package boolean flag start index
@@ -111,18 +109,13 @@
"{container}",
aconfig_storage::StorageFileType::flag_val);
if (!flag_value_file.ok()) \{
- ALOGI("error: failed to get flag value file: %s", flag_value_file.error().c_str());
- return;
+ ALOGE("error: failed to get flag value file: %s", flag_value_file.error().c_str());
}
// cache flag value file
flag_value_file_ = std::unique_ptr<aconfig_storage::MappedStorageFile>(
*flag_value_file);
- use_new_storage_value = server_configurable_flags::GetServerConfigurableFlag(
- "aconfig_flags.core_experiments_team_internal",
- "com.android.providers.settings.use_new_storage_value",
- "false") == "true";
}
{{ -endif }}
{{ -endif }}
@@ -131,44 +124,30 @@
virtual bool {item.flag_name}() override \{
{{ -if item.readwrite }}
if (cache_[{item.readwrite_idx}] == -1) \{
+ {{ if allow_instrumentation- }}
+ if (read_from_new_storage_) \{
+ auto value = aconfig_storage::get_boolean_flag_value(
+ *flag_value_file_,
+ boolean_start_index_ + {item.flag_offset});
+
+ if (!value.ok()) \{
+ ALOGE("error: failed to read flag value: %s", value.error().c_str());
+ }
+
+ cache_[{item.readwrite_idx}] = *value;
+ } else \{
+ cache_[{item.readwrite_idx}] = server_configurable_flags::GetServerConfigurableFlag(
+ "aconfig_flags.{item.device_config_namespace}",
+ "{item.device_config_flag}",
+ "{item.default_value}") == "true";
+ }
+ {{ -else- }}
cache_[{item.readwrite_idx}] = server_configurable_flags::GetServerConfigurableFlag(
"aconfig_flags.{item.device_config_namespace}",
"{item.device_config_flag}",
"{item.default_value}") == "true";
- }
-
-
- {{ if allow_instrumentation- }}
- if (read_from_new_storage_) \{
- if (!flag_value_file_) \{
- ALOGI("error: failed to get flag {item.flag_name}: flag value file is null");
- return cache_[{item.readwrite_idx}];
- }
-
- auto value = aconfig_storage::get_boolean_flag_value(
- *flag_value_file_,
- boolean_start_index_ + {item.flag_offset});
-
- if (!value.ok()) \{
- ALOGI("error: failed to read flag value: %s", value.error().c_str());
- return cache_[{item.readwrite_idx}];
- }
-
- bool expected_value = cache_[{item.readwrite_idx}];
- if (*value != expected_value) \{
- ALOGI("error: {item.flag_name} value mismatch, new storage value is %s, old storage value is %s",
- *value ? "true" : "false", expected_value ? "true" : "false");
- }
-
- if (use_new_storage_value) \{
- return *value;
- } else \{
- return expected_value;
- }
- }
{{ -endif }}
-
-
+ }
return cache_[{item.readwrite_idx}];
{{ -else }}
{{ -if item.is_fixed_read_only }}
@@ -189,7 +168,6 @@
std::unique_ptr<aconfig_storage::MappedStorageFile> flag_value_file_;
bool read_from_new_storage_;
- bool use_new_storage_value;
{{ -endif }}
{{ -endif }}
diff --git a/tools/aconfig/aconfig/templates/rust.template b/tools/aconfig/aconfig/templates/rust.template
index 77a9984..c2f162f 100644
--- a/tools/aconfig/aconfig/templates/rust.template
+++ b/tools/aconfig/aconfig/templates/rust.template
@@ -2,105 +2,85 @@
use aconfig_storage_read_api::\{Mmap, AconfigStorageError, StorageFileType, PackageReadContext, get_mapped_storage_file, get_boolean_flag_value, get_package_read_context};
use std::path::Path;
use std::io::Write;
+use std::sync::LazyLock;
use log::\{log, LevelFilter, Level};
-static STORAGE_MIGRATION_MARKER_FILE: &str =
- "/metadata/aconfig_test_missions/mission_1";
-static MIGRATION_LOG_TAG: &str = "AconfigTestMission1";
-
/// flag provider
pub struct FlagProvider;
{{ if has_readwrite- }}
-lazy_static::lazy_static! \{
- {{ if allow_instrumentation }}
- static ref PACKAGE_OFFSET: Result<Option<u32>, AconfigStorageError> = unsafe \{
- get_mapped_storage_file("{container}", StorageFileType::PackageMap)
- .and_then(|package_map| get_package_read_context(&package_map, "{package}"))
- .map(|context| context.map(|c| c.boolean_start_index))
- };
+{{ if allow_instrumentation }}
+static READ_FROM_NEW_STORAGE: LazyLock<bool> = LazyLock::new(|| unsafe \{
+ Path::new("/metadata/aconfig/boot/enable_only_new_storage").exists()
+});
- static ref FLAG_VAL_MAP: Result<Mmap, AconfigStorageError> = unsafe \{
- get_mapped_storage_file("{container}", StorageFileType::FlagVal)
- };
- {{ -endif }}
+static PACKAGE_OFFSET: LazyLock<Result<Option<u32>, AconfigStorageError>> = LazyLock::new(|| unsafe \{
+ get_mapped_storage_file("{container}", StorageFileType::PackageMap)
+ .and_then(|package_map| get_package_read_context(&package_map, "{package}"))
+ .map(|context| context.map(|c| c.boolean_start_index))
+});
+static FLAG_VAL_MAP: LazyLock<Result<Mmap, AconfigStorageError>> = LazyLock::new(|| unsafe \{
+ get_mapped_storage_file("{container}", StorageFileType::FlagVal)
+});
+{{ -endif }}
{{ -for flag in template_flags }}
- {{ -if flag.readwrite }}
- /// flag value cache for {flag.name}
- {{ if allow_instrumentation }}
- static ref CACHED_{flag.name}: bool = \{
- let result = flags_rust::GetServerConfigurableFlag(
- "aconfig_flags.{flag.device_config_namespace}",
- "{flag.device_config_flag}",
- "{flag.default_value}") == "true";
- let use_new_storage_value = flags_rust::GetServerConfigurableFlag(
- "aconfig_flags.core_experiments_team_internal",
- "com.android.providers.settings.use_new_storage_value",
- "false") == "true";
+{{ -if flag.readwrite }}
+/// flag value cache for {flag.name}
+{{ if allow_instrumentation }}
+static CACHED_{flag.name}: LazyLock<bool> = LazyLock::new(|| \{
- if Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() \{
- // This will be called multiple times. Subsequent calls after the first are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info));
+ if *READ_FROM_NEW_STORAGE \{
+ // This will be called multiple times. Subsequent calls after the first are noops.
+ logger::init(
+ logger::Config::default()
+ .with_tag_on_device("aconfig_rust_codegen")
+ .with_max_level(LevelFilter::Info));
- let aconfig_storage_result = FLAG_VAL_MAP
- .as_ref()
- .map_err(|err| format!("failed to get flag val map: \{err}"))
- .and_then(|flag_val_map| \{
- PACKAGE_OFFSET
- .as_ref()
- .map_err(|err| format!("failed to get package read offset: \{err}"))
- .and_then(|package_offset| \{
- match package_offset \{
- Some(offset) => \{
- get_boolean_flag_value(&flag_val_map, offset + {flag.flag_offset})
- .map_err(|err| format!("failed to get flag: \{err}"))
- },
- None => Err("no context found for package '{package}'".to_string())
- }
- })
- });
+ let flag_value_result = FLAG_VAL_MAP
+ .as_ref()
+ .map_err(|err| format!("failed to get flag val map: \{err}"))
+ .and_then(|flag_val_map| \{
+ PACKAGE_OFFSET
+ .as_ref()
+ .map_err(|err| format!("failed to get package read offset: \{err}"))
+ .and_then(|package_offset| \{
+ match package_offset \{
+ Some(offset) => \{
+ get_boolean_flag_value(&flag_val_map, offset + {flag.flag_offset})
+ .map_err(|err| format!("failed to get flag: \{err}"))
+ },
+ None => Err("no context found for package '{package}'".to_string())
+ }
+ })
+ });
- match aconfig_storage_result \{
- Ok(storage_result) if storage_result == result => \{
- if use_new_storage_value \{
- return storage_result;
- } else \{
- return result;
- }
- },
- Ok(storage_result) => \{
- log!(Level::Error, "AconfigTestMission1: error: mismatch for flag '{flag.name}'. Legacy storage was \{result}, new storage was \{storage_result}");
- if use_new_storage_value \{
- return storage_result;
- } else \{
- return result;
- }
- },
- Err(err) => \{
- log!(Level::Error, "AconfigTestMission1: error: \{err}");
- if use_new_storage_value \{
- panic!("failed to read flag value: \{err}");
- }
- }
+ match flag_value_result \{
+ Ok(flag_value) => \{
+ return flag_value;
+ },
+ Err(err) => \{
+ log!(Level::Error, "aconfig_rust_codegen: error: \{err}");
+ panic!("failed to read flag value: \{err}");
}
}
+ } else \{
+ flags_rust::GetServerConfigurableFlag(
+ "aconfig_flags.{flag.device_config_namespace}",
+ "{flag.device_config_flag}",
+ "{flag.default_value}") == "true"
+ }
- result
- };
- {{ else }}
- static ref CACHED_{flag.name}: bool = flags_rust::GetServerConfigurableFlag(
- "aconfig_flags.{flag.device_config_namespace}",
- "{flag.device_config_flag}",
- "{flag.default_value}") == "true";
- {{ endif }}
- {{ -endif }}
+});
+{{ else }}
+static CACHED_{flag.name}: LazyLock<bool> = LazyLock::new(|| flags_rust::GetServerConfigurableFlag(
+ "aconfig_flags.{flag.device_config_namespace}",
+ "{flag.device_config_flag}",
+ "{flag.default_value}") == "true");
+{{ endif }}
+{{ -endif }}
{{ -endfor }}
-}
{{ -endif }}
impl FlagProvider \{
@@ -124,72 +104,11 @@
{{ for flag in template_flags }}
/// query flag {flag.name}
#[inline(always)]
-{{ -if flag.readwrite }}
pub fn {flag.name}() -> bool \{
+{{ -if flag.readwrite }}
PROVIDER.{flag.name}()
{{ -else }}
-pub fn {flag.name}() -> bool \{
- {{ if not allow_instrumentation }}
{flag.default_value}
- {{ else }}
-
- let result = {flag.default_value};
- if !Path::new(STORAGE_MIGRATION_MARKER_FILE).exists() \{
- return result;
- }
-
- // This will be called multiple times. Subsequent calls after the first
- // are noops.
- logger::init(
- logger::Config::default()
- .with_tag_on_device(MIGRATION_LOG_TAG)
- .with_max_level(LevelFilter::Info),
- );
-
- unsafe \{
- let package_map = match get_mapped_storage_file("{flag.container}", StorageFileType::PackageMap) \{
- Ok(file) => file,
- Err(err) => \{
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag '{flag.name}': \{err}");
- return result;
- }
- };
-
- let package_read_context = match get_package_read_context(&package_map, "{package}") \{
- Ok(Some(context)) => context,
- Ok(None) => \{
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag '{flag.name}': did not get context");
- return result;
- },
- Err(err) => \{
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag '{flag.name}': \{err}");
- return result;
- }
- };
- let flag_val_map = match get_mapped_storage_file("{flag.container}", StorageFileType::FlagVal) \{
- Ok(val_map) => val_map,
- Err(err) => \{
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag '{flag.name}': \{err}");
- return result;
- }
- };
- let value = match get_boolean_flag_value(&flag_val_map, {flag.flag_offset} + package_read_context.boolean_start_index) \{
- Ok(val) => val,
- Err(err) => \{
- log!(Level::Error, "AconfigTestMission1: error: failed to read flag '{flag.name}': \{err}");
- return result;
- }
- };
-
- if result != value \{
- log!(Level::Error, "AconfigTestMission1: error: flag mismatch for '{flag.name}'. Legacy storage was \{result}, new storage was \{value}");
- } else \{
- let default_value = {flag.default_value};
- }
- }
-
- result
- {{ endif }}
{{ -endif }}
}
{{ endfor }}
diff --git a/tools/aconfig/aconfig_device_paths/Android.bp b/tools/aconfig/aconfig_device_paths/Android.bp
index 95cecf4..dda7a55 100644
--- a/tools/aconfig/aconfig_device_paths/Android.bp
+++ b/tools/aconfig/aconfig_device_paths/Android.bp
@@ -51,4 +51,21 @@
static_libs: [
"libaconfig_java_proto_nano",
],
+ sdk_version: "core_platform",
+ apex_available: [
+ "//apex_available:platform",
+ ],
+}
+
+genrule {
+ name: "libaconfig_java_host_device_paths_src",
+ srcs: ["src/HostDeviceProtosTemplate.java"],
+ out: ["HostDeviceProtos.java"],
+ tool_files: ["partition_aconfig_flags_paths.txt"],
+ cmd: "sed -e '/TEMPLATE/{r$(location partition_aconfig_flags_paths.txt)' -e 'd}' $(in) > $(out)",
+}
+
+java_library_host {
+ name: "aconfig_host_device_paths_java",
+ srcs: [":libaconfig_java_host_device_paths_src"],
}
diff --git a/tools/aconfig/aconfig_device_paths/partition_aconfig_flags_paths.txt b/tools/aconfig/aconfig_device_paths/partition_aconfig_flags_paths.txt
index 140cd21..e997e3d 100644
--- a/tools/aconfig/aconfig_device_paths/partition_aconfig_flags_paths.txt
+++ b/tools/aconfig/aconfig_device_paths/partition_aconfig_flags_paths.txt
@@ -1,4 +1,3 @@
"/system/etc/aconfig_flags.pb",
-"/system_ext/etc/aconfig_flags.pb",
"/product/etc/aconfig_flags.pb",
"/vendor/etc/aconfig_flags.pb",
diff --git a/tools/aconfig/aconfig_device_paths/src/DeviceProtosTemplate.java b/tools/aconfig/aconfig_device_paths/src/DeviceProtosTemplate.java
index 58c58de..4d41199 100644
--- a/tools/aconfig/aconfig_device_paths/src/DeviceProtosTemplate.java
+++ b/tools/aconfig/aconfig_device_paths/src/DeviceProtosTemplate.java
@@ -29,7 +29,7 @@
* @hide
*/
public class DeviceProtos {
- static final String[] PATHS = {
+ public static final String[] PATHS = {
TEMPLATE
};
@@ -50,10 +50,11 @@
ArrayList<parsed_flag> result = new ArrayList();
for (String path : parsedFlagsProtoPaths()) {
- FileInputStream inputStream = new FileInputStream(path);
- parsed_flags parsedFlags = parsed_flags.parseFrom(inputStream.readAllBytes());
- for (parsed_flag flag : parsedFlags.parsedFlag) {
- result.add(flag);
+ try (FileInputStream inputStream = new FileInputStream(path)) {
+ parsed_flags parsedFlags = parsed_flags.parseFrom(inputStream.readAllBytes());
+ for (parsed_flag flag : parsedFlags.parsedFlag) {
+ result.add(flag);
+ }
}
}
@@ -64,7 +65,7 @@
* Returns the list of all on-device aconfig protos paths.
* @hide
*/
- private static List<String> parsedFlagsProtoPaths() {
+ public static List<String> parsedFlagsProtoPaths() {
ArrayList<String> paths = new ArrayList(Arrays.asList(PATHS));
File apexDirectory = new File(APEX_DIR);
diff --git a/tools/aconfig/aconfig_device_paths/src/HostDeviceProtosTemplate.java b/tools/aconfig/aconfig_device_paths/src/HostDeviceProtosTemplate.java
new file mode 100644
index 0000000..e2ad40a
--- /dev/null
+++ b/tools/aconfig/aconfig_device_paths/src/HostDeviceProtosTemplate.java
@@ -0,0 +1,84 @@
+/*
+ * Copyright (C) 2024 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package android.aconfig;
+
+import java.io.File;
+import java.util.ArrayList;
+import java.util.Arrays;
+import java.util.HashSet;
+import java.util.List;
+import java.util.Set;
+import java.util.stream.Collectors;
+
+/**
+ * A host lib that can read all aconfig proto file paths on a given device.
+ * This lib is only available on device with root access (userdebug/eng).
+ */
+public class HostDeviceProtos {
+ /**
+ * An interface that executes ADB command and return the result.
+ */
+ public static interface AdbCommandExecutor {
+ /** Executes the ADB command. */
+ String executeAdbCommand(String command);
+ }
+
+ static final String[] PATHS = {
+ TEMPLATE
+ };
+
+ private static final String APEX_DIR = "/apex";
+ private static final String RECURSIVELY_LIST_APEX_DIR_COMMAND =
+ "shell su 0 find /apex | grep aconfig_flags";
+ private static final String APEX_ACONFIG_PATH_SUFFIX = "/etc/aconfig_flags.pb";
+
+
+ /**
+ * Returns the list of all on-device aconfig proto paths from host side.
+ */
+ public static List<String> parsedFlagsProtoPaths(AdbCommandExecutor adbCommandExecutor) {
+ ArrayList<String> paths = new ArrayList(Arrays.asList(PATHS));
+
+ String adbCommandOutput = adbCommandExecutor.executeAdbCommand(
+ RECURSIVELY_LIST_APEX_DIR_COMMAND);
+
+ if (adbCommandOutput == null) {
+ return paths;
+ }
+
+ Set<String> allFiles = new HashSet<>(Arrays.asList(adbCommandOutput.split("\n")));
+
+ Set<String> subdirs = allFiles.stream().map(file -> {
+ String[] filePaths = file.split("/");
+ // The first element is "", the second element is "apex".
+ return filePaths.length > 2 ? filePaths[2] : "";
+ }).collect(Collectors.toSet());
+
+ for (String prefix : subdirs) {
+ // For each mainline modules, there are two directories, one <modulepackage>/,
+ // and one <modulepackage>@<versioncode>/. Just read the former.
+ if (prefix.contains("@")) {
+ continue;
+ }
+
+ String protoPath = APEX_DIR + "/" + prefix + APEX_ACONFIG_PATH_SUFFIX;
+ if (allFiles.contains(protoPath)) {
+ paths.add(protoPath);
+ }
+ }
+ return paths;
+ }
+}
diff --git a/tools/aconfig/aconfig_device_paths/src/lib.rs b/tools/aconfig/aconfig_device_paths/src/lib.rs
index 9ab9cea..8871b4f 100644
--- a/tools/aconfig/aconfig_device_paths/src/lib.rs
+++ b/tools/aconfig/aconfig_device_paths/src/lib.rs
@@ -62,13 +62,12 @@
#[test]
fn test_read_partition_paths() {
- assert_eq!(read_partition_paths().len(), 4);
+ assert_eq!(read_partition_paths().len(), 3);
assert_eq!(
read_partition_paths(),
vec![
PathBuf::from("/system/etc/aconfig_flags.pb"),
- PathBuf::from("/system_ext/etc/aconfig_flags.pb"),
PathBuf::from("/product/etc/aconfig_flags.pb"),
PathBuf::from("/vendor/etc/aconfig_flags.pb")
]
diff --git a/tools/aconfig/aconfig_flags/Android.bp b/tools/aconfig/aconfig_flags/Android.bp
new file mode 100644
index 0000000..e327ced
--- /dev/null
+++ b/tools/aconfig/aconfig_flags/Android.bp
@@ -0,0 +1,46 @@
+/*
+ * Copyright (C) 2024 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+rust_library {
+ name: "libaconfig_flags",
+ crate_name: "aconfig_flags",
+ srcs: [
+ "src/lib.rs",
+ ],
+ rustlibs: [
+ "libaconfig_flags_rust",
+ ],
+ host_supported: true,
+}
+
+aconfig_declarations {
+ name: "aconfig_flags",
+ package: "com.android.aconfig.flags",
+ container: "system",
+ srcs: ["flags.aconfig"],
+}
+
+rust_aconfig_library {
+ name: "libaconfig_flags_rust",
+ crate_name: "aconfig_flags_rust",
+ aconfig_declarations: "aconfig_flags",
+ host_supported: true,
+}
+
+cc_aconfig_library {
+ name: "libaconfig_flags_cc",
+ aconfig_declarations: "aconfig_flags",
+}
diff --git a/tools/aconfig/aconfig_flags/Cargo.toml b/tools/aconfig/aconfig_flags/Cargo.toml
new file mode 100644
index 0000000..6eb9f14
--- /dev/null
+++ b/tools/aconfig/aconfig_flags/Cargo.toml
@@ -0,0 +1,10 @@
+[package]
+name = "aconfig_flags"
+version = "0.1.0"
+edition = "2021"
+
+[features]
+default = ["cargo"]
+cargo = []
+
+[dependencies]
\ No newline at end of file
diff --git a/tools/aconfig/aconfig_flags/flags.aconfig b/tools/aconfig/aconfig_flags/flags.aconfig
new file mode 100644
index 0000000..0a004ca
--- /dev/null
+++ b/tools/aconfig/aconfig_flags/flags.aconfig
@@ -0,0 +1,16 @@
+package: "com.android.aconfig.flags"
+container: "system"
+
+flag {
+ name: "enable_only_new_storage"
+ namespace: "core_experiments_team_internal"
+ bug: "312235596"
+ description: "When enabled, aconfig flags are read from the new aconfig storage only."
+}
+
+flag {
+ name: "enable_aconfigd_from_mainline"
+ namespace: "core_experiments_team_internal"
+ bug: "369808805"
+ description: "When enabled, launch aconfigd from config infra module."
+}
diff --git a/tools/aconfig/aconfig_flags/src/lib.rs b/tools/aconfig/aconfig_flags/src/lib.rs
new file mode 100644
index 0000000..2e89127
--- /dev/null
+++ b/tools/aconfig/aconfig_flags/src/lib.rs
@@ -0,0 +1,58 @@
+/*
+ * Copyright (C) 2024 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+//! `aconfig_flags` is a crate for reading aconfig flags from Rust
+// When building with the Android tool-chain
+//
+// - the flag functions will read from aconfig_flags_inner
+// - the feature "cargo" will be disabled
+//
+// When building with cargo
+//
+// - the flag functions will all return some trivial value, like true
+// - the feature "cargo" will be enabled
+//
+// This module hides these differences from the rest of aconfig.
+
+/// Module used when building with the Android tool-chain
+#[cfg(not(feature = "cargo"))]
+pub mod auto_generated {
+ /// Returns the value for the enable_only_new_storage flag.
+ pub fn enable_only_new_storage() -> bool {
+ aconfig_flags_rust::enable_only_new_storage()
+ }
+
+ /// Returns the value for the enable_aconfigd_from_mainline flag.
+ pub fn enable_aconfigd_from_mainline() -> bool {
+ aconfig_flags_rust::enable_only_new_storage()
+ }
+}
+
+/// Module used when building with cargo
+#[cfg(feature = "cargo")]
+pub mod auto_generated {
+ /// Returns a placeholder value for the enable_only_new_storage flag.
+ pub fn enable_only_new_storage() -> bool {
+ // Used only to enable typechecking and testing with cargo
+ true
+ }
+
+ /// Returns a placeholder value for the enable_aconfigd_from_mainline flag.
+ pub fn enable_aconfigd_from_mainline() -> bool {
+ // Used only to enable typechecking and testing with cargo
+ true
+ }
+}
diff --git a/tools/aconfig/aconfig_storage_file/Android.bp b/tools/aconfig/aconfig_storage_file/Android.bp
index 40b4464..e875c7b 100644
--- a/tools/aconfig/aconfig_storage_file/Android.bp
+++ b/tools/aconfig/aconfig_storage_file/Android.bp
@@ -14,6 +14,7 @@
"libclap",
"libcxx",
"libaconfig_storage_protos",
+ "libserde",
],
}
@@ -36,7 +37,10 @@
name: "aconfig-storage",
defaults: ["aconfig_storage_file.defaults"],
srcs: ["src/main.rs"],
- rustlibs: ["libaconfig_storage_file"],
+ rustlibs: [
+ "libaconfig_storage_file",
+ "libserde_json",
+ ],
}
rust_test_host {
diff --git a/tools/aconfig/aconfig_storage_file/Cargo.toml b/tools/aconfig/aconfig_storage_file/Cargo.toml
index 192dfad..a405578 100644
--- a/tools/aconfig/aconfig_storage_file/Cargo.toml
+++ b/tools/aconfig/aconfig_storage_file/Cargo.toml
@@ -14,6 +14,8 @@
thiserror = "1.0.56"
clap = { version = "4.1.8", features = ["derive"] }
cxx = "1.0"
+serde = { version = "1.0.152", features = ["derive"] }
+serde_json = "1.0.93"
[[bin]]
name = "aconfig-storage"
diff --git a/tools/aconfig/aconfig_storage_file/src/flag_info.rs b/tools/aconfig/aconfig_storage_file/src/flag_info.rs
index beac38d..f090396 100644
--- a/tools/aconfig/aconfig_storage_file/src/flag_info.rs
+++ b/tools/aconfig/aconfig_storage_file/src/flag_info.rs
@@ -20,10 +20,11 @@
use crate::{read_str_from_bytes, read_u32_from_bytes, read_u8_from_bytes};
use crate::{AconfigStorageError, StorageFileType};
use anyhow::anyhow;
+use serde::{Deserialize, Serialize};
use std::fmt;
/// Flag info header struct
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct FlagInfoHeader {
pub version: u32,
pub container: String,
@@ -89,7 +90,7 @@
}
/// bit field for flag info
-#[derive(Clone, Debug, PartialEq, Eq)]
+#[derive(Clone, Debug, PartialEq, Eq, Serialize, Deserialize)]
pub enum FlagInfoBit {
HasServerOverride = 1 << 0,
IsReadWrite = 1 << 1,
@@ -97,7 +98,7 @@
}
/// Flag info node struct
-#[derive(PartialEq, Clone)]
+#[derive(PartialEq, Clone, Serialize, Deserialize)]
pub struct FlagInfoNode {
pub attributes: u8,
}
@@ -138,7 +139,7 @@
}
/// Flag info list struct
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct FlagInfoList {
pub header: FlagInfoHeader,
pub nodes: Vec<FlagInfoNode>,
diff --git a/tools/aconfig/aconfig_storage_file/src/flag_table.rs b/tools/aconfig/aconfig_storage_file/src/flag_table.rs
index 64b90ea..0588fe5 100644
--- a/tools/aconfig/aconfig_storage_file/src/flag_table.rs
+++ b/tools/aconfig/aconfig_storage_file/src/flag_table.rs
@@ -23,10 +23,11 @@
};
use crate::{AconfigStorageError, StorageFileType, StoredFlagType};
use anyhow::anyhow;
+use serde::{Deserialize, Serialize};
use std::fmt;
/// Flag table header struct
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct FlagTableHeader {
pub version: u32,
pub container: String,
@@ -95,7 +96,7 @@
}
/// Flag table node struct
-#[derive(PartialEq, Clone)]
+#[derive(PartialEq, Clone, Serialize, Deserialize)]
pub struct FlagTableNode {
pub package_id: u32,
pub flag_name: String,
@@ -150,11 +151,11 @@
/// Calculate node bucket index
pub fn find_bucket_index(package_id: u32, flag_name: &str, num_buckets: u32) -> u32 {
let full_flag_name = package_id.to_string() + "/" + flag_name;
- get_bucket_index(&full_flag_name, num_buckets)
+ get_bucket_index(full_flag_name.as_bytes(), num_buckets)
}
}
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct FlagTable {
pub header: FlagTableHeader,
pub buckets: Vec<Option<u32>>,
diff --git a/tools/aconfig/aconfig_storage_file/src/flag_value.rs b/tools/aconfig/aconfig_storage_file/src/flag_value.rs
index 506924b..b64c10e 100644
--- a/tools/aconfig/aconfig_storage_file/src/flag_value.rs
+++ b/tools/aconfig/aconfig_storage_file/src/flag_value.rs
@@ -20,10 +20,11 @@
use crate::{read_str_from_bytes, read_u32_from_bytes, read_u8_from_bytes};
use crate::{AconfigStorageError, StorageFileType};
use anyhow::anyhow;
+use serde::{Deserialize, Serialize};
use std::fmt;
/// Flag value header struct
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct FlagValueHeader {
pub version: u32,
pub container: String,
@@ -89,7 +90,7 @@
}
/// Flag value list struct
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct FlagValueList {
pub header: FlagValueHeader,
pub booleans: Vec<bool>,
diff --git a/tools/aconfig/aconfig_storage_file/src/lib.rs b/tools/aconfig/aconfig_storage_file/src/lib.rs
index 26e9c1a..cf52bc0 100644
--- a/tools/aconfig/aconfig_storage_file/src/lib.rs
+++ b/tools/aconfig/aconfig_storage_file/src/lib.rs
@@ -37,19 +37,21 @@
pub mod flag_value;
pub mod package_table;
pub mod protos;
+pub mod sip_hasher13;
pub mod test_utils;
use anyhow::anyhow;
+use serde::{Deserialize, Serialize};
use std::cmp::Ordering;
-use std::collections::hash_map::DefaultHasher;
use std::fs::File;
-use std::hash::{Hash, Hasher};
+use std::hash::Hasher;
use std::io::Read;
pub use crate::flag_info::{FlagInfoBit, FlagInfoHeader, FlagInfoList, FlagInfoNode};
pub use crate::flag_table::{FlagTable, FlagTableHeader, FlagTableNode};
pub use crate::flag_value::{FlagValueHeader, FlagValueList};
pub use crate::package_table::{PackageTable, PackageTableHeader, PackageTableNode};
+pub use crate::sip_hasher13::SipHasher13;
use crate::AconfigStorageError::{
BytesParseFail, HashTableSizeLimit, InvalidFlagValueType, InvalidStoredFlagType,
@@ -106,7 +108,7 @@
/// Flag type enum as stored by storage file
/// ONLY APPEND, NEVER REMOVE FOR BACKWARD COMPATIBILITY. THE MAX IS U16.
-#[derive(Clone, Copy, Debug, PartialEq, Eq)]
+#[derive(Clone, Copy, Debug, PartialEq, Eq, Serialize, Deserialize)]
pub enum StoredFlagType {
ReadWriteBoolean = 0,
ReadOnlyBoolean = 1,
@@ -211,10 +213,12 @@
}
/// Get the corresponding bucket index given the key and number of buckets
-pub(crate) fn get_bucket_index<T: Hash>(val: &T, num_buckets: u32) -> u32 {
- let mut s = DefaultHasher::new();
- val.hash(&mut s);
- (s.finish() % num_buckets as u64) as u32
+pub(crate) fn get_bucket_index(val: &[u8], num_buckets: u32) -> u32 {
+ let mut s = SipHasher13::new();
+ s.write(val);
+ s.write_u8(0xff);
+ let ret = (s.finish() % num_buckets as u64) as u32;
+ ret
}
/// Read and parse bytes as u8
diff --git a/tools/aconfig/aconfig_storage_file/src/main.rs b/tools/aconfig/aconfig_storage_file/src/main.rs
index 8b9e38d..a9cfd19 100644
--- a/tools/aconfig/aconfig_storage_file/src/main.rs
+++ b/tools/aconfig/aconfig_storage_file/src/main.rs
@@ -20,9 +20,29 @@
list_flags, list_flags_with_info, read_file_to_bytes, AconfigStorageError, FlagInfoList,
FlagTable, FlagValueList, PackageTable, StorageFileType,
};
-
use clap::{builder::ArgAction, Arg, Command};
+use serde::Serialize;
+use serde_json;
+use std::fmt;
+use std::fs;
+use std::fs::File;
+use std::io::Write;
+/**
+ * Usage Examples
+ *
+ * Print file:
+ * $ aconfig-storage print --file=path/to/flag.map --type=flag_map
+ *
+ * List flags:
+ * $ aconfig-storage list --flag-map=path/to/flag.map \
+ * --flag-val=path/to/flag.val --package-map=path/to/package.map
+ *
+ * Write binary file for testing:
+ * $ aconfig-storage print --file=path/to/flag.map --type=flag_map --format=json > flag_map.json
+ * $ vim flag_map.json // Manually make updates
+ * $ aconfig-storage write-bytes --input-file=flag_map.json --output-file=path/to/flag.map --type=flag_map
+ */
fn cli() -> Command {
Command::new("aconfig-storage")
.subcommand_required(true)
@@ -34,7 +54,8 @@
.long("type")
.required(true)
.value_parser(|s: &str| StorageFileType::try_from(s)),
- ),
+ )
+ .arg(Arg::new("format").long("format").required(false).action(ArgAction::Set)),
)
.subcommand(
Command::new("list")
@@ -50,41 +71,75 @@
Arg::new("flag-info").long("flag-info").required(false).action(ArgAction::Set),
),
)
+ .subcommand(
+ Command::new("write-bytes")
+ // Where to write the output bytes. Suggest to use the StorageFileType names (e.g. flag.map).
+ .arg(
+ Arg::new("output-file")
+ .long("output-file")
+ .required(true)
+ .action(ArgAction::Set),
+ )
+ // Input file should be json.
+ .arg(
+ Arg::new("input-file").long("input-file").required(true).action(ArgAction::Set),
+ )
+ .arg(
+ Arg::new("type")
+ .long("type")
+ .required(true)
+ .value_parser(|s: &str| StorageFileType::try_from(s)),
+ ),
+ )
}
fn print_storage_file(
file_path: &str,
file_type: &StorageFileType,
+ as_json: bool,
) -> Result<(), AconfigStorageError> {
let bytes = read_file_to_bytes(file_path)?;
match file_type {
StorageFileType::PackageMap => {
let package_table = PackageTable::from_bytes(&bytes)?;
- println!("{:?}", package_table);
+ println!("{}", to_print_format(package_table, as_json));
}
StorageFileType::FlagMap => {
let flag_table = FlagTable::from_bytes(&bytes)?;
- println!("{:?}", flag_table);
+ println!("{}", to_print_format(flag_table, as_json));
}
StorageFileType::FlagVal => {
let flag_value = FlagValueList::from_bytes(&bytes)?;
- println!("{:?}", flag_value);
+ println!("{}", to_print_format(flag_value, as_json));
}
StorageFileType::FlagInfo => {
let flag_info = FlagInfoList::from_bytes(&bytes)?;
- println!("{:?}", flag_info);
+ println!("{}", to_print_format(flag_info, as_json));
}
}
Ok(())
}
+fn to_print_format<T>(file_contents: T, as_json: bool) -> String
+where
+ T: Serialize + fmt::Debug,
+{
+ if as_json {
+ serde_json::to_string(&file_contents).unwrap()
+ } else {
+ format!("{:?}", file_contents)
+ }
+}
+
fn main() -> Result<(), AconfigStorageError> {
let matches = cli().get_matches();
match matches.subcommand() {
Some(("print", sub_matches)) => {
let file_path = sub_matches.get_one::<String>("file").unwrap();
let file_type = sub_matches.get_one::<StorageFileType>("type").unwrap();
- print_storage_file(file_path, file_type)?
+ let format = sub_matches.get_one::<String>("format");
+ let as_json: bool = format == Some(&"json".to_string());
+ print_storage_file(file_path, file_type, as_json)?
}
Some(("list", sub_matches)) => {
let package_map = sub_matches.get_one::<String>("package-map").unwrap();
@@ -96,10 +151,10 @@
let flags = list_flags_with_info(package_map, flag_map, flag_val, info_file)?;
for flag in flags.iter() {
println!(
- "{} {} {} {:?} IsReadWrite: {}, HasServerOverride: {}, HasLocalOverride: {}",
- flag.package_name, flag.flag_name, flag.flag_value, flag.value_type,
- flag.is_readwrite, flag.has_server_override, flag.has_local_override,
- );
+ "{} {} {} {:?} IsReadWrite: {}, HasServerOverride: {}, HasLocalOverride: {}",
+ flag.package_name, flag.flag_name, flag.flag_value, flag.value_type,
+ flag.is_readwrite, flag.has_server_override, flag.has_local_override,
+ );
}
}
None => {
@@ -113,6 +168,40 @@
}
}
}
+ // Converts JSON of the file into raw bytes (as is used on-device).
+ // Intended to generate/easily update these files for testing.
+ Some(("write-bytes", sub_matches)) => {
+ let input_file_path = sub_matches.get_one::<String>("input-file").unwrap();
+ let input_json = fs::read_to_string(input_file_path).unwrap();
+
+ let file_type = sub_matches.get_one::<StorageFileType>("type").unwrap();
+ let output_bytes: Vec<u8>;
+ match file_type {
+ StorageFileType::FlagVal => {
+ let list: FlagValueList = serde_json::from_str(&input_json).unwrap();
+ output_bytes = list.into_bytes();
+ }
+ StorageFileType::FlagInfo => {
+ let list: FlagInfoList = serde_json::from_str(&input_json).unwrap();
+ output_bytes = list.into_bytes();
+ }
+ StorageFileType::FlagMap => {
+ let table: FlagTable = serde_json::from_str(&input_json).unwrap();
+ output_bytes = table.into_bytes();
+ }
+ StorageFileType::PackageMap => {
+ let table: PackageTable = serde_json::from_str(&input_json).unwrap();
+ output_bytes = table.into_bytes();
+ }
+ }
+
+ let output_file_path = sub_matches.get_one::<String>("output-file").unwrap();
+ let file = File::create(output_file_path);
+ if file.is_err() {
+ panic!("can't make file");
+ }
+ let _ = file.unwrap().write_all(&output_bytes);
+ }
_ => unreachable!(),
}
Ok(())
diff --git a/tools/aconfig/aconfig_storage_file/src/package_table.rs b/tools/aconfig/aconfig_storage_file/src/package_table.rs
index b734972..a5bd9e6 100644
--- a/tools/aconfig/aconfig_storage_file/src/package_table.rs
+++ b/tools/aconfig/aconfig_storage_file/src/package_table.rs
@@ -20,10 +20,11 @@
use crate::{get_bucket_index, read_str_from_bytes, read_u32_from_bytes, read_u8_from_bytes};
use crate::{AconfigStorageError, StorageFileType};
use anyhow::anyhow;
+use serde::{Deserialize, Serialize};
use std::fmt;
/// Package table header struct
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct PackageTableHeader {
pub version: u32,
pub container: String,
@@ -92,7 +93,7 @@
}
/// Package table node struct
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct PackageTableNode {
pub package_name: String,
pub package_id: u32,
@@ -146,12 +147,12 @@
/// construction side (aconfig binary) and consumption side (flag read lib)
/// use the same method of hashing
pub fn find_bucket_index(package: &str, num_buckets: u32) -> u32 {
- get_bucket_index(&package, num_buckets)
+ get_bucket_index(package.as_bytes(), num_buckets)
}
}
/// Package table struct
-#[derive(PartialEq)]
+#[derive(PartialEq, Serialize, Deserialize)]
pub struct PackageTable {
pub header: PackageTableHeader,
pub buckets: Vec<Option<u32>>,
diff --git a/tools/aconfig/aconfig_storage_file/src/sip_hasher13.rs b/tools/aconfig/aconfig_storage_file/src/sip_hasher13.rs
new file mode 100644
index 0000000..9be3175
--- /dev/null
+++ b/tools/aconfig/aconfig_storage_file/src/sip_hasher13.rs
@@ -0,0 +1,327 @@
+/*
+ * Copyright (C) 2023 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+//! An implementation of SipHash13
+
+use std::cmp;
+use std::mem;
+use std::ptr;
+use std::slice;
+
+use std::hash::Hasher;
+
+/// An implementation of SipHash 2-4.
+///
+#[derive(Debug, Clone, Default)]
+pub struct SipHasher13 {
+ k0: u64,
+ k1: u64,
+ length: usize, // how many bytes we've processed
+ state: State, // hash State
+ tail: u64, // unprocessed bytes le
+ ntail: usize, // how many bytes in tail are valid
+}
+
+#[derive(Debug, Clone, Copy, Default)]
+#[repr(C)]
+struct State {
+ // v0, v2 and v1, v3 show up in pairs in the algorithm,
+ // and simd implementations of SipHash will use vectors
+ // of v02 and v13. By placing them in this order in the struct,
+ // the compiler can pick up on just a few simd optimizations by itself.
+ v0: u64,
+ v2: u64,
+ v1: u64,
+ v3: u64,
+}
+
+macro_rules! compress {
+ ($state:expr) => {{
+ compress!($state.v0, $state.v1, $state.v2, $state.v3)
+ }};
+ ($v0:expr, $v1:expr, $v2:expr, $v3:expr) => {{
+ $v0 = $v0.wrapping_add($v1);
+ $v1 = $v1.rotate_left(13);
+ $v1 ^= $v0;
+ $v0 = $v0.rotate_left(32);
+ $v2 = $v2.wrapping_add($v3);
+ $v3 = $v3.rotate_left(16);
+ $v3 ^= $v2;
+ $v0 = $v0.wrapping_add($v3);
+ $v3 = $v3.rotate_left(21);
+ $v3 ^= $v0;
+ $v2 = $v2.wrapping_add($v1);
+ $v1 = $v1.rotate_left(17);
+ $v1 ^= $v2;
+ $v2 = $v2.rotate_left(32);
+ }};
+}
+
+/// Load an integer of the desired type from a byte stream, in LE order. Uses
+/// `copy_nonoverlapping` to let the compiler generate the most efficient way
+/// to load it from a possibly unaligned address.
+///
+/// Unsafe because: unchecked indexing at i..i+size_of(int_ty)
+macro_rules! load_int_le {
+ ($buf:expr, $i:expr, $int_ty:ident) => {{
+ debug_assert!($i + mem::size_of::<$int_ty>() <= $buf.len());
+ let mut data = 0 as $int_ty;
+ ptr::copy_nonoverlapping(
+ $buf.get_unchecked($i),
+ &mut data as *mut _ as *mut u8,
+ mem::size_of::<$int_ty>(),
+ );
+ data.to_le()
+ }};
+}
+
+/// Load an u64 using up to 7 bytes of a byte slice.
+///
+/// Unsafe because: unchecked indexing at start..start+len
+#[inline]
+unsafe fn u8to64_le(buf: &[u8], start: usize, len: usize) -> u64 {
+ debug_assert!(len < 8);
+ let mut i = 0; // current byte index (from LSB) in the output u64
+ let mut out = 0;
+ if i + 3 < len {
+ out = load_int_le!(buf, start + i, u32) as u64;
+ i += 4;
+ }
+ if i + 1 < len {
+ out |= (load_int_le!(buf, start + i, u16) as u64) << (i * 8);
+ i += 2
+ }
+ if i < len {
+ out |= (*buf.get_unchecked(start + i) as u64) << (i * 8);
+ i += 1;
+ }
+ debug_assert_eq!(i, len);
+ out
+}
+
+impl SipHasher13 {
+ /// Creates a new `SipHasher13` with the two initial keys set to 0.
+ #[inline]
+ pub fn new() -> SipHasher13 {
+ SipHasher13::new_with_keys(0, 0)
+ }
+
+ /// Creates a `SipHasher13` that is keyed off the provided keys.
+ #[inline]
+ pub fn new_with_keys(key0: u64, key1: u64) -> SipHasher13 {
+ let mut sip_hasher = SipHasher13 {
+ k0: key0,
+ k1: key1,
+ length: 0,
+ state: State { v0: 0, v1: 0, v2: 0, v3: 0 },
+ tail: 0,
+ ntail: 0,
+ };
+ sip_hasher.reset();
+ sip_hasher
+ }
+
+ #[inline]
+ fn c_rounds(state: &mut State) {
+ compress!(state);
+ }
+
+ #[inline]
+ fn d_rounds(state: &mut State) {
+ compress!(state);
+ compress!(state);
+ compress!(state);
+ }
+
+ #[inline]
+ fn reset(&mut self) {
+ self.length = 0;
+ self.state.v0 = self.k0 ^ 0x736f6d6570736575;
+ self.state.v1 = self.k1 ^ 0x646f72616e646f6d;
+ self.state.v2 = self.k0 ^ 0x6c7967656e657261;
+ self.state.v3 = self.k1 ^ 0x7465646279746573;
+ self.ntail = 0;
+ }
+
+ // Specialized write function that is only valid for buffers with len <= 8.
+ // It's used to force inlining of write_u8 and write_usize, those would normally be inlined
+ // except for composite types (that includes slices and str hashing because of delimiter).
+ // Without this extra push the compiler is very reluctant to inline delimiter writes,
+ // degrading performance substantially for the most common use cases.
+ #[inline]
+ fn short_write(&mut self, msg: &[u8]) {
+ debug_assert!(msg.len() <= 8);
+ let length = msg.len();
+ self.length += length;
+
+ let needed = 8 - self.ntail;
+ let fill = cmp::min(length, needed);
+ if fill == 8 {
+ // safe to call since msg hasn't been loaded
+ self.tail = unsafe { load_int_le!(msg, 0, u64) };
+ } else {
+ // safe to call since msg hasn't been loaded, and fill <= msg.len()
+ self.tail |= unsafe { u8to64_le(msg, 0, fill) } << (8 * self.ntail);
+ if length < needed {
+ self.ntail += length;
+ return;
+ }
+ }
+ self.state.v3 ^= self.tail;
+ Self::c_rounds(&mut self.state);
+ self.state.v0 ^= self.tail;
+
+ // Buffered tail is now flushed, process new input.
+ self.ntail = length - needed;
+ // safe to call since number of `needed` bytes has been loaded
+ // and self.ntail + needed == msg.len()
+ self.tail = unsafe { u8to64_le(msg, needed, self.ntail) };
+ }
+}
+
+impl Hasher for SipHasher13 {
+ // see short_write comment for explanation
+ #[inline]
+ fn write_usize(&mut self, i: usize) {
+ // safe to call, since convert the pointer to u8
+ let bytes = unsafe {
+ slice::from_raw_parts(&i as *const usize as *const u8, mem::size_of::<usize>())
+ };
+ self.short_write(bytes);
+ }
+
+ // see short_write comment for explanation
+ #[inline]
+ fn write_u8(&mut self, i: u8) {
+ self.short_write(&[i]);
+ }
+
+ #[inline]
+ fn write(&mut self, msg: &[u8]) {
+ let length = msg.len();
+ self.length += length;
+
+ let mut needed = 0;
+
+ // loading unprocessed byte from last write
+ if self.ntail != 0 {
+ needed = 8 - self.ntail;
+ // safe to call, since msg hasn't been processed
+ // and cmp::min(length, needed) < 8
+ self.tail |= unsafe { u8to64_le(msg, 0, cmp::min(length, needed)) } << 8 * self.ntail;
+ if length < needed {
+ self.ntail += length;
+ return;
+ } else {
+ self.state.v3 ^= self.tail;
+ Self::c_rounds(&mut self.state);
+ self.state.v0 ^= self.tail;
+ self.ntail = 0;
+ }
+ }
+
+ // Buffered tail is now flushed, process new input.
+ let len = length - needed;
+ let left = len & 0x7;
+
+ let mut i = needed;
+ while i < len - left {
+ // safe to call since if i < len - left, it means msg has at least 1 byte to load
+ let mi = unsafe { load_int_le!(msg, i, u64) };
+
+ self.state.v3 ^= mi;
+ Self::c_rounds(&mut self.state);
+ self.state.v0 ^= mi;
+
+ i += 8;
+ }
+
+ // safe to call since if left == 0, since this call will load nothing
+ // if left > 0, it means there are number of `left` bytes in msg
+ self.tail = unsafe { u8to64_le(msg, i, left) };
+ self.ntail = left;
+ }
+
+ #[inline]
+ fn finish(&self) -> u64 {
+ let mut state = self.state;
+
+ let b: u64 = ((self.length as u64 & 0xff) << 56) | self.tail;
+
+ state.v3 ^= b;
+ Self::c_rounds(&mut state);
+ state.v0 ^= b;
+
+ state.v2 ^= 0xff;
+ Self::d_rounds(&mut state);
+
+ state.v0 ^ state.v1 ^ state.v2 ^ state.v3
+ }
+}
+
+#[cfg(test)]
+mod tests {
+ use super::*;
+
+ use std::hash::{Hash, Hasher};
+ use std::string::String;
+
+ #[test]
+ // this test point locks down the value list serialization
+ fn test_sip_hash13_string_hash() {
+ let mut sip_hash13 = SipHasher13::new();
+ let test_str1 = String::from("com.google.android.test");
+ test_str1.hash(&mut sip_hash13);
+ assert_eq!(17898838669067067585, sip_hash13.finish());
+
+ let test_str2 = String::from("adfadfadf adfafadadf 1231241241");
+ test_str2.hash(&mut sip_hash13);
+ assert_eq!(13543518987672889310, sip_hash13.finish());
+ }
+
+ #[test]
+ fn test_sip_hash13_write() {
+ let mut sip_hash13 = SipHasher13::new();
+ let test_str1 = String::from("com.google.android.test");
+ sip_hash13.write(test_str1.as_bytes());
+ sip_hash13.write_u8(0xff);
+ assert_eq!(17898838669067067585, sip_hash13.finish());
+
+ let mut sip_hash132 = SipHasher13::new();
+ let test_str1 = String::from("com.google.android.test");
+ sip_hash132.write(test_str1.as_bytes());
+ assert_eq!(9685440969685209025, sip_hash132.finish());
+ sip_hash132.write(test_str1.as_bytes());
+ assert_eq!(6719694176662736568, sip_hash132.finish());
+
+ let mut sip_hash133 = SipHasher13::new();
+ let test_str2 = String::from("abcdefg");
+ test_str2.hash(&mut sip_hash133);
+ assert_eq!(2492161047327640297, sip_hash133.finish());
+
+ let mut sip_hash134 = SipHasher13::new();
+ let test_str3 = String::from("abcdefgh");
+ test_str3.hash(&mut sip_hash134);
+ assert_eq!(6689927370435554326, sip_hash134.finish());
+ }
+
+ #[test]
+ fn test_sip_hash13_write_short() {
+ let mut sip_hash13 = SipHasher13::new();
+ sip_hash13.write_u8(0x61);
+ assert_eq!(4644417185603328019, sip_hash13.finish());
+ }
+}
diff --git a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/ByteBufferReader.java b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/ByteBufferReader.java
index 1c72364..4bea083 100644
--- a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/ByteBufferReader.java
+++ b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/ByteBufferReader.java
@@ -48,6 +48,10 @@
return new String(bytes, StandardCharsets.UTF_8);
}
+ public int readByte(int i) {
+ return Byte.toUnsignedInt(mByteBuffer.get(i));
+ }
+
public void position(int newPosition) {
mByteBuffer.position(newPosition);
}
diff --git a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/FlagTable.java b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/FlagTable.java
index e85fdee..757844a 100644
--- a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/FlagTable.java
+++ b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/FlagTable.java
@@ -16,41 +16,57 @@
package android.aconfig.storage;
+import static java.nio.charset.StandardCharsets.UTF_8;
+
import java.nio.ByteBuffer;
-import java.util.HashMap;
-import java.util.Map;
import java.util.Objects;
public class FlagTable {
private Header mHeader;
- private Map<String, Node> mNodeMap;
+ private ByteBufferReader mReader;
public static FlagTable fromBytes(ByteBuffer bytes) {
FlagTable flagTable = new FlagTable();
- ByteBufferReader reader = new ByteBufferReader(bytes);
- Header header = Header.fromBytes(reader);
- flagTable.mHeader = header;
- flagTable.mNodeMap = new HashMap(TableUtils.getTableSize(header.mNumFlags));
- reader.position(header.mNodeOffset);
- for (int i = 0; i < header.mNumFlags; i++) {
- Node node = Node.fromBytes(reader);
- flagTable.mNodeMap.put(makeKey(node.mPackageId, node.mFlagName), node);
- }
+ flagTable.mReader = new ByteBufferReader(bytes);
+ flagTable.mHeader = Header.fromBytes(flagTable.mReader);
+
return flagTable;
}
public Node get(int packageId, String flagName) {
- return mNodeMap.get(makeKey(packageId, flagName));
+ int numBuckets = (mHeader.mNodeOffset - mHeader.mBucketOffset) / 4;
+ int bucketIndex = TableUtils.getBucketIndex(makeKey(packageId, flagName), numBuckets);
+ int newPosition = mHeader.mBucketOffset + bucketIndex * 4;
+ if (newPosition >= mHeader.mNodeOffset) {
+ return null;
+ }
+
+ mReader.position(newPosition);
+ int nodeIndex = mReader.readInt();
+ if (nodeIndex < mHeader.mNodeOffset || nodeIndex >= mHeader.mFileSize) {
+ return null;
+ }
+
+ while (nodeIndex != -1) {
+ mReader.position(nodeIndex);
+ Node node = Node.fromBytes(mReader);
+ if (Objects.equals(flagName, node.mFlagName) && packageId == node.mPackageId) {
+ return node;
+ }
+ nodeIndex = node.mNextOffset;
+ }
+
+ return null;
}
public Header getHeader() {
return mHeader;
}
- private static String makeKey(int packageId, String flagName) {
+ private static byte[] makeKey(int packageId, String flagName) {
StringBuilder ret = new StringBuilder();
- return ret.append(packageId).append('/').append(flagName).toString();
+ return ret.append(packageId).append('/').append(flagName).toString().getBytes(UTF_8);
}
public static class Header {
diff --git a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/FlagValueList.java b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/FlagValueList.java
index 0ddc147..493436d 100644
--- a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/FlagValueList.java
+++ b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/FlagValueList.java
@@ -17,33 +17,21 @@
package android.aconfig.storage;
import java.nio.ByteBuffer;
-import java.util.ArrayList;
-import java.util.List;
public class FlagValueList {
private Header mHeader;
- private List<Boolean> mList;
-
- private int mSize;
+ private ByteBufferReader mReader;
public static FlagValueList fromBytes(ByteBuffer bytes) {
FlagValueList flagValueList = new FlagValueList();
- ByteBufferReader reader = new ByteBufferReader(bytes);
- Header header = Header.fromBytes(reader);
- flagValueList.mHeader = header;
- flagValueList.mList = new ArrayList(header.mNumFlags);
- reader.position(header.mBooleanValueOffset);
- for (int i = 0; i < header.mNumFlags; i++) {
- boolean val = reader.readByte() == 1;
- flagValueList.mList.add(val);
- }
- flagValueList.mSize = flagValueList.mList.size();
+ flagValueList.mReader = new ByteBufferReader(bytes);
+ flagValueList.mHeader = Header.fromBytes(flagValueList.mReader);
return flagValueList;
}
- public boolean get(int index) {
- return mList.get(index);
+ public boolean getBoolean(int index) {
+ return mReader.readByte(mHeader.mBooleanValueOffset + index) == 1;
}
public Header getHeader() {
@@ -51,7 +39,7 @@
}
public int size() {
- return mSize;
+ return mHeader.mNumFlags;
}
public static class Header {
diff --git a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/PackageTable.java b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/PackageTable.java
index d04e1ac..dc1c583 100644
--- a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/PackageTable.java
+++ b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/PackageTable.java
@@ -16,32 +16,48 @@
package android.aconfig.storage;
+import static java.nio.charset.StandardCharsets.UTF_8;
+
import java.nio.ByteBuffer;
-import java.util.HashMap;
-import java.util.Map;
import java.util.Objects;
public class PackageTable {
private Header mHeader;
- private Map<String, Node> mNodeMap;
+ private ByteBufferReader mReader;
public static PackageTable fromBytes(ByteBuffer bytes) {
PackageTable packageTable = new PackageTable();
- ByteBufferReader reader = new ByteBufferReader(bytes);
- Header header = Header.fromBytes(reader);
- packageTable.mHeader = header;
- packageTable.mNodeMap = new HashMap(TableUtils.getTableSize(header.mNumPackages));
- reader.position(header.mNodeOffset);
- for (int i = 0; i < header.mNumPackages; i++) {
- Node node = Node.fromBytes(reader);
- packageTable.mNodeMap.put(node.mPackageName, node);
- }
+ packageTable.mReader = new ByteBufferReader(bytes);
+ packageTable.mHeader = Header.fromBytes(packageTable.mReader);
+
return packageTable;
}
public Node get(String packageName) {
- return mNodeMap.get(packageName);
+ int numBuckets = (mHeader.mNodeOffset - mHeader.mBucketOffset) / 4;
+ int bucketIndex = TableUtils.getBucketIndex(packageName.getBytes(UTF_8), numBuckets);
+ int newPosition = mHeader.mBucketOffset + bucketIndex * 4;
+ if (newPosition >= mHeader.mNodeOffset) {
+ return null;
+ }
+ mReader.position(newPosition);
+ int nodeIndex = mReader.readInt();
+
+ if (nodeIndex < mHeader.mNodeOffset || nodeIndex >= mHeader.mFileSize) {
+ return null;
+ }
+
+ while (nodeIndex != -1) {
+ mReader.position(nodeIndex);
+ Node node = Node.fromBytes(mReader);
+ if (Objects.equals(packageName, node.mPackageName)) {
+ return node;
+ }
+ nodeIndex = node.mNextOffset;
+ }
+
+ return null;
}
public Header getHeader() {
diff --git a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/SipHasher13.java b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/SipHasher13.java
new file mode 100644
index 0000000..64714ee
--- /dev/null
+++ b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/SipHasher13.java
@@ -0,0 +1,115 @@
+/*
+ * Copyright (C) 2024 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package android.aconfig.storage;
+
+public class SipHasher13 {
+ static class State {
+ private long v0;
+ private long v2;
+ private long v1;
+ private long v3;
+
+ public State(long k0, long k1) {
+ v0 = k0 ^ 0x736f6d6570736575L;
+ v1 = k1 ^ 0x646f72616e646f6dL;
+ v2 = k0 ^ 0x6c7967656e657261L;
+ v3 = k1 ^ 0x7465646279746573L;
+ }
+
+ public void compress(long m) {
+ v3 ^= m;
+ cRounds();
+ v0 ^= m;
+ }
+
+ public long finish() {
+ v2 ^= 0xff;
+ dRounds();
+ return v0 ^ v1 ^ v2 ^ v3;
+ }
+
+ private void cRounds() {
+ v0 += v1;
+ v1 = Long.rotateLeft(v1, 13);
+ v1 ^= v0;
+ v0 = Long.rotateLeft(v0, 32);
+ v2 += v3;
+ v3 = Long.rotateLeft(v3, 16);
+ v3 ^= v2;
+ v0 += v3;
+ v3 = Long.rotateLeft(v3, 21);
+ v3 ^= v0;
+ v2 += v1;
+ v1 = Long.rotateLeft(v1, 17);
+ v1 ^= v2;
+ v2 = Long.rotateLeft(v2, 32);
+ }
+
+ private void dRounds() {
+ for (int i = 0; i < 3; i++) {
+ v0 += v1;
+ v1 = Long.rotateLeft(v1, 13);
+ v1 ^= v0;
+ v0 = Long.rotateLeft(v0, 32);
+ v2 += v3;
+ v3 = Long.rotateLeft(v3, 16);
+ v3 ^= v2;
+ v0 += v3;
+ v3 = Long.rotateLeft(v3, 21);
+ v3 ^= v0;
+ v2 += v1;
+ v1 = Long.rotateLeft(v1, 17);
+ v1 ^= v2;
+ v2 = Long.rotateLeft(v2, 32);
+ }
+ }
+ }
+
+ public static long hash(byte[] data) {
+ State state = new State(0, 0);
+ int len = data.length;
+ int left = len & 0x7;
+ int index = 0;
+
+ while (index < len - left) {
+ long mi = loadLe(data, index, 8);
+ index += 8;
+ state.compress(mi);
+ }
+
+ // padding the end with 0xff to be consistent with rust
+ long m = (0xffL << (left * 8)) | loadLe(data, index, left);
+ if (left == 0x7) {
+ // compress the m w-2
+ state.compress(m);
+ m = 0L;
+ }
+ // len adds 1 since padded 0xff
+ m |= (((len + 1) & 0xffL) << 56);
+ state.compress(m);
+
+ return state.finish();
+ }
+
+ private static long loadLe(byte[] data, int offset, int size) {
+ long m = 0;
+ for (int i = 0; i < size; i++) {
+ m |= (data[i + offset] & 0xffL) << (i * 8);
+ }
+ return m;
+ }
+}
diff --git a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/TableUtils.java b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/TableUtils.java
index 714b53b..81168f5 100644
--- a/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/TableUtils.java
+++ b/tools/aconfig/aconfig_storage_file/srcs/android/aconfig/storage/TableUtils.java
@@ -58,4 +58,9 @@
}
throw new AconfigStorageException("Number of items in a hash table exceeds limit");
}
+
+ public static int getBucketIndex(byte[] val, int numBuckets) {
+ long hashVal = SipHasher13.hash(val);
+ return (int) Long.remainderUnsigned(hashVal, numBuckets);
+ }
}
diff --git a/tools/aconfig/aconfig_storage_file/tests/Android.bp b/tools/aconfig/aconfig_storage_file/tests/Android.bp
index e2e225d..12e4aca 100644
--- a/tools/aconfig/aconfig_storage_file/tests/Android.bp
+++ b/tools/aconfig/aconfig_storage_file/tests/Android.bp
@@ -30,9 +30,10 @@
static_libs: [
"androidx.test.runner",
"junit",
+ "aconfig_storage_file_java",
],
test_config: "AndroidStorageJaveTest.xml",
- certificate: "platform",
+ sdk_version: "test_current",
data: [
"package.map",
"flag.map",
@@ -42,4 +43,5 @@
test_suites: [
"general-tests",
],
+ jarjar_rules: "jarjar.txt",
}
diff --git a/tools/aconfig/aconfig_storage_file/tests/jarjar.txt b/tools/aconfig/aconfig_storage_file/tests/jarjar.txt
new file mode 100644
index 0000000..a6c17fa
--- /dev/null
+++ b/tools/aconfig/aconfig_storage_file/tests/jarjar.txt
@@ -0,0 +1,15 @@
+rule android.aconfig.storage.AconfigStorageException android.aconfig.storage.test.AconfigStorageException
+rule android.aconfig.storage.FlagTable android.aconfig.storage.test.FlagTable
+rule android.aconfig.storage.PackageTable android.aconfig.storage.test.PackageTable
+rule android.aconfig.storage.ByteBufferReader android.aconfig.storage.test.ByteBufferReader
+rule android.aconfig.storage.FlagType android.aconfig.storage.test.FlagType
+rule android.aconfig.storage.SipHasher13 android.aconfig.storage.test.SipHasher13
+rule android.aconfig.storage.FileType android.aconfig.storage.test.FileType
+rule android.aconfig.storage.FlagValueList android.aconfig.storage.test.FlagValueList
+rule android.aconfig.storage.TableUtils android.aconfig.storage.test.TableUtils
+
+
+rule android.aconfig.storage.FlagTable$* android.aconfig.storage.test.FlagTable$@1
+rule android.aconfig.storage.PackageTable$* android.aconfig.storage.test.PackageTable$@1
+rule android.aconfig.storage.FlagValueList$* android.aconfig.storage.test.FlagValueList@1
+rule android.aconfig.storage.SipHasher13$* android.aconfig.storage.test.SipHasher13@1
diff --git a/tools/aconfig/aconfig_storage_file/tests/srcs/FlagValueListTest.java b/tools/aconfig/aconfig_storage_file/tests/srcs/FlagValueListTest.java
index c18590a..1b0de63 100644
--- a/tools/aconfig/aconfig_storage_file/tests/srcs/FlagValueListTest.java
+++ b/tools/aconfig/aconfig_storage_file/tests/srcs/FlagValueListTest.java
@@ -53,7 +53,7 @@
assertEquals(expected.length, flagValueList.size());
for (int i = 0; i < flagValueList.size(); i++) {
- assertEquals(expected[i], flagValueList.get(i));
+ assertEquals(expected[i], flagValueList.getBoolean(i));
}
}
@@ -68,10 +68,10 @@
PackageTable.Node pNode = packageTable.get("com.android.aconfig.storage.test_1");
FlagTable.Node fNode = flagTable.get(pNode.getPackageId(), "enabled_rw");
- assertTrue(flagValueList.get(pNode.getBooleanStartIndex() + fNode.getFlagIndex()));
+ assertTrue(flagValueList.getBoolean(pNode.getBooleanStartIndex() + fNode.getFlagIndex()));
pNode = packageTable.get("com.android.aconfig.storage.test_4");
fNode = flagTable.get(pNode.getPackageId(), "enabled_fixed_ro");
- assertTrue(flagValueList.get(pNode.getBooleanStartIndex() + fNode.getFlagIndex()));
+ assertTrue(flagValueList.getBoolean(pNode.getBooleanStartIndex() + fNode.getFlagIndex()));
}
}
diff --git a/tools/aconfig/aconfig_storage_file/tests/srcs/SipHasher13Test.java b/tools/aconfig/aconfig_storage_file/tests/srcs/SipHasher13Test.java
new file mode 100644
index 0000000..10620d2
--- /dev/null
+++ b/tools/aconfig/aconfig_storage_file/tests/srcs/SipHasher13Test.java
@@ -0,0 +1,44 @@
+/*
+ * Copyright (C) 2024 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package android.aconfig.storage.test;
+
+import static org.junit.Assert.assertEquals;
+import static java.nio.charset.StandardCharsets.UTF_8;
+
+import android.aconfig.storage.SipHasher13;
+
+import org.junit.Test;
+import org.junit.runner.RunWith;
+import org.junit.runners.JUnit4;
+
+@RunWith(JUnit4.class)
+public class SipHasher13Test {
+ @Test
+ public void testSipHash_hashString() throws Exception {
+ String testStr = "com.google.android.test";
+ long result = SipHasher13.hash(testStr.getBytes(UTF_8));
+ assertEquals(0xF86572EFF9C4A0C1L, result);
+
+ testStr = "abcdefg";
+ result = SipHasher13.hash(testStr.getBytes(UTF_8));
+ assertEquals(0x2295EF44BD078AE9L, result);
+
+ testStr = "abcdefgh";
+ result = SipHasher13.hash(testStr.getBytes(UTF_8));
+ assertEquals(0x5CD7657FA7F96C16L, result);
+ }
+}
diff --git a/tools/aconfig/aconfig_storage_read_api/Android.bp b/tools/aconfig/aconfig_storage_read_api/Android.bp
index 9e950a6..6ae34f3 100644
--- a/tools/aconfig/aconfig_storage_read_api/Android.bp
+++ b/tools/aconfig/aconfig_storage_read_api/Android.bp
@@ -107,31 +107,12 @@
afdo: true,
}
-soong_config_module_type {
- name: "aconfig_lib_cc_shared_link_defaults",
- module_type: "cc_defaults",
- config_namespace: "Aconfig",
- bool_variables: [
- "read_from_new_storage",
- ],
- properties: [
- "shared_libs",
- ],
-}
-
-soong_config_bool_variable {
- name: "read_from_new_storage",
-}
-
-aconfig_lib_cc_shared_link_defaults {
+cc_defaults {
name: "aconfig_lib_cc_shared_link.defaults",
- soong_config_variables: {
- read_from_new_storage: {
- shared_libs: [
- "libaconfig_storage_read_api_cc",
- ],
- },
- },
+ shared_libs: select(release_flag("RELEASE_READ_FROM_NEW_STORAGE"), {
+ true: ["libaconfig_storage_read_api_cc"],
+ default: [],
+ }),
}
cc_defaults {
@@ -147,6 +128,7 @@
crate_name: "aconfig_storage_read_api_rust_jni",
srcs: ["srcs/lib.rs"],
rustlibs: [
+ "libaconfig_storage_file",
"libaconfig_storage_read_api",
"libanyhow",
"libjni",
diff --git a/tools/aconfig/aconfig_storage_read_api/srcs/android/aconfig/storage/AconfigStorageReadAPI.java b/tools/aconfig/aconfig_storage_read_api/srcs/android/aconfig/storage/AconfigStorageReadAPI.java
index 406ff24..850c2b8 100644
--- a/tools/aconfig/aconfig_storage_read_api/srcs/android/aconfig/storage/AconfigStorageReadAPI.java
+++ b/tools/aconfig/aconfig_storage_read_api/srcs/android/aconfig/storage/AconfigStorageReadAPI.java
@@ -16,18 +16,14 @@
package android.aconfig.storage;
+import dalvik.annotation.optimization.FastNative;
+
import java.io.FileInputStream;
import java.io.IOException;
import java.nio.ByteBuffer;
import java.nio.ByteOrder;
import java.nio.MappedByteBuffer;
import java.nio.channels.FileChannel;
-import java.nio.channels.FileChannel.MapMode;
-
-import android.aconfig.storage.PackageReadContext;
-import android.aconfig.storage.FlagReadContext;
-
-import dalvik.annotation.optimization.FastNative;
public class AconfigStorageReadAPI {
@@ -50,9 +46,8 @@
}
// Map a storage file given container and file type
- public static MappedByteBuffer getMappedFile(
- String container,
- StorageFileType type) throws IOException{
+ public static MappedByteBuffer getMappedFile(String container, StorageFileType type)
+ throws IOException {
switch (type) {
case PACKAGE_MAP:
return mapStorageFile(STORAGEDIR + "/maps/" + container + ".package.map");
@@ -73,14 +68,14 @@
// @throws IOException if the passed in file is not a valid package map file
@FastNative
private static native ByteBuffer getPackageReadContextImpl(
- ByteBuffer mappedFile, String packageName) throws IOException;
+ ByteBuffer mappedFile, String packageName) throws IOException;
// API to get package read context
// @param mappedFile: memory mapped package map file
// @param packageName: package name
// @throws IOException if the passed in file is not a valid package map file
- static public PackageReadContext getPackageReadContext (
- ByteBuffer mappedFile, String packageName) throws IOException {
+ public static PackageReadContext getPackageReadContext(
+ ByteBuffer mappedFile, String packageName) throws IOException {
ByteBuffer buffer = getPackageReadContextImpl(mappedFile, packageName);
buffer.order(ByteOrder.LITTLE_ENDIAN);
return new PackageReadContext(buffer.getInt(), buffer.getInt(4));
@@ -94,7 +89,7 @@
// @throws IOException if the passed in file is not a valid flag map file
@FastNative
private static native ByteBuffer getFlagReadContextImpl(
- ByteBuffer mappedFile, int packageId, String flagName) throws IOException;
+ ByteBuffer mappedFile, int packageId, String flagName) throws IOException;
// API to get flag read context
// @param mappedFile: memory mapped flag map file
@@ -103,7 +98,7 @@
// @param flagName: flag name
// @throws IOException if the passed in file is not a valid flag map file
public static FlagReadContext getFlagReadContext(
- ByteBuffer mappedFile, int packageId, String flagName) throws IOException {
+ ByteBuffer mappedFile, int packageId, String flagName) throws IOException {
ByteBuffer buffer = getFlagReadContextImpl(mappedFile, packageId, flagName);
buffer.order(ByteOrder.LITTLE_ENDIAN);
return new FlagReadContext(buffer.getInt(), buffer.getInt(4));
@@ -115,8 +110,11 @@
// @throws IOException if the passed in file is not a valid flag value file or the
// flag index went over the file boundary.
@FastNative
- public static native boolean getBooleanFlagValue(
- ByteBuffer mappedFile, int flagIndex) throws IOException;
+ public static native boolean getBooleanFlagValue(ByteBuffer mappedFile, int flagIndex)
+ throws IOException;
+
+ @FastNative
+ public static native long hash(String packageName) throws IOException;
static {
System.loadLibrary("aconfig_storage_read_api_rust_jni");
diff --git a/tools/aconfig/aconfig_storage_read_api/srcs/android/aconfig/storage/StorageInternalReader.java b/tools/aconfig/aconfig_storage_read_api/srcs/android/aconfig/storage/StorageInternalReader.java
index 71c1c0d..6fbcdb3 100644
--- a/tools/aconfig/aconfig_storage_read_api/srcs/android/aconfig/storage/StorageInternalReader.java
+++ b/tools/aconfig/aconfig_storage_read_api/srcs/android/aconfig/storage/StorageInternalReader.java
@@ -53,10 +53,7 @@
@UnsupportedAppUsage
public boolean getBooleanFlagValue(int index) {
index += mPackageBooleanStartOffset;
- if (index >= mFlagValueList.size()) {
- throw new AconfigStorageException("Fail to get boolean flag value");
- }
- return mFlagValueList.get(index);
+ return mFlagValueList.getBoolean(index);
}
private int getPackageBooleanStartOffset(String packageName) {
diff --git a/tools/aconfig/aconfig_storage_read_api/srcs/lib.rs b/tools/aconfig/aconfig_storage_read_api/srcs/lib.rs
index 304a059..f5f12bb 100644
--- a/tools/aconfig/aconfig_storage_read_api/srcs/lib.rs
+++ b/tools/aconfig/aconfig_storage_read_api/srcs/lib.rs
@@ -1,5 +1,6 @@
//! aconfig storage read api java rust interlop
+use aconfig_storage_file::SipHasher13;
use aconfig_storage_read_api::flag_table_query::find_flag_read_context;
use aconfig_storage_read_api::flag_value_query::find_boolean_flag_value;
use aconfig_storage_read_api::package_table_query::find_package_read_context;
@@ -7,8 +8,9 @@
use anyhow::Result;
use jni::objects::{JByteBuffer, JClass, JString};
-use jni::sys::{jboolean, jint};
+use jni::sys::{jboolean, jint, jlong};
use jni::JNIEnv;
+use std::hash::Hasher;
/// Call rust find package read context
fn get_package_read_context_java(
@@ -158,3 +160,30 @@
}
}
}
+
+/// Get flag value JNI
+#[no_mangle]
+#[allow(unused)]
+pub extern "system" fn Java_android_aconfig_storage_AconfigStorageReadAPI_hash<'local>(
+ mut env: JNIEnv<'local>,
+ class: JClass<'local>,
+ package_name: JString<'local>,
+) -> jlong {
+ match siphasher13_hash(&mut env, package_name) {
+ Ok(value) => value as jlong,
+ Err(errmsg) => {
+ env.throw(("java/io/IOException", errmsg.to_string())).expect("failed to throw");
+ 0i64
+ }
+ }
+}
+
+fn siphasher13_hash(env: &mut JNIEnv, package_name: JString) -> Result<u64> {
+ // SAFETY:
+ // The safety here is ensured as the flag name is guaranteed to be a java string
+ let flag_name: String = unsafe { env.get_string_unchecked(&package_name)?.into() };
+ let mut s = SipHasher13::new();
+ s.write(flag_name.as_bytes());
+ s.write_u8(0xff);
+ Ok(s.finish())
+}
diff --git a/tools/aconfig/aconfig_storage_read_api/tests/java/AconfigStorageReadAPITest.java b/tools/aconfig/aconfig_storage_read_api/tests/java/AconfigStorageReadAPITest.java
index a26b257..191741e 100644
--- a/tools/aconfig/aconfig_storage_read_api/tests/java/AconfigStorageReadAPITest.java
+++ b/tools/aconfig/aconfig_storage_read_api/tests/java/AconfigStorageReadAPITest.java
@@ -16,28 +16,29 @@
package android.aconfig.storage.test;
-import java.io.IOException;
-import java.nio.MappedByteBuffer;
-import java.nio.ByteBuffer;
-import java.nio.ByteOrder;
-import java.util.ArrayList;
-import java.util.List;
-import java.util.Random;
-
import static org.junit.Assert.assertEquals;
-import static org.junit.Assert.assertFalse;
import static org.junit.Assert.assertTrue;
+
+import android.aconfig.DeviceProtos;
+import android.aconfig.nano.Aconfig.parsed_flag;
+import android.aconfig.storage.AconfigStorageReadAPI;
+import android.aconfig.storage.FlagReadContext;
+import android.aconfig.storage.FlagReadContext.StoredFlagType;
+import android.aconfig.storage.PackageReadContext;
+import android.aconfig.storage.SipHasher13;
+import android.aconfig.storage.StorageInternalReader;
+
import org.junit.Test;
import org.junit.runner.RunWith;
import org.junit.runners.JUnit4;
-import android.aconfig.storage.AconfigStorageReadAPI;
-import android.aconfig.storage.PackageReadContext;
-import android.aconfig.storage.FlagReadContext;
-import android.aconfig.storage.FlagReadContext.StoredFlagType;
+import java.io.IOException;
+import java.nio.MappedByteBuffer;
+import java.util.ArrayList;
+import java.util.List;
@RunWith(JUnit4.class)
-public class AconfigStorageReadAPITest{
+public class AconfigStorageReadAPITest {
private String mStorageDir = "/data/local/tmp/aconfig_java_api_test";
@@ -45,26 +46,29 @@
public void testPackageContextQuery() {
MappedByteBuffer packageMap = null;
try {
- packageMap = AconfigStorageReadAPI.mapStorageFile(
- mStorageDir + "/maps/mockup.package.map");
- } catch(IOException ex){
+ packageMap =
+ AconfigStorageReadAPI.mapStorageFile(mStorageDir + "/maps/mockup.package.map");
+ } catch (IOException ex) {
assertTrue(ex.toString(), false);
}
assertTrue(packageMap != null);
try {
- PackageReadContext context = AconfigStorageReadAPI.getPackageReadContext(
- packageMap, "com.android.aconfig.storage.test_1");
+ PackageReadContext context =
+ AconfigStorageReadAPI.getPackageReadContext(
+ packageMap, "com.android.aconfig.storage.test_1");
assertEquals(context.mPackageId, 0);
assertEquals(context.mBooleanStartIndex, 0);
- context = AconfigStorageReadAPI.getPackageReadContext(
- packageMap, "com.android.aconfig.storage.test_2");
+ context =
+ AconfigStorageReadAPI.getPackageReadContext(
+ packageMap, "com.android.aconfig.storage.test_2");
assertEquals(context.mPackageId, 1);
assertEquals(context.mBooleanStartIndex, 3);
- context = AconfigStorageReadAPI.getPackageReadContext(
- packageMap, "com.android.aconfig.storage.test_4");
+ context =
+ AconfigStorageReadAPI.getPackageReadContext(
+ packageMap, "com.android.aconfig.storage.test_4");
assertEquals(context.mPackageId, 2);
assertEquals(context.mBooleanStartIndex, 6);
} catch (IOException ex) {
@@ -76,19 +80,19 @@
public void testNonExistPackageContextQuery() {
MappedByteBuffer packageMap = null;
try {
- packageMap = AconfigStorageReadAPI.mapStorageFile(
- mStorageDir + "/maps/mockup.package.map");
- } catch(IOException ex){
+ packageMap =
+ AconfigStorageReadAPI.mapStorageFile(mStorageDir + "/maps/mockup.package.map");
+ } catch (IOException ex) {
assertTrue(ex.toString(), false);
}
assertTrue(packageMap != null);
try {
- PackageReadContext context = AconfigStorageReadAPI.getPackageReadContext(
- packageMap, "unknown");
+ PackageReadContext context =
+ AconfigStorageReadAPI.getPackageReadContext(packageMap, "unknown");
assertEquals(context.mPackageId, -1);
assertEquals(context.mBooleanStartIndex, -1);
- } catch(IOException ex){
+ } catch (IOException ex) {
assertTrue(ex.toString(), false);
}
}
@@ -97,12 +101,11 @@
public void testFlagContextQuery() {
MappedByteBuffer flagMap = null;
try {
- flagMap = AconfigStorageReadAPI.mapStorageFile(
- mStorageDir + "/maps/mockup.flag.map");
- } catch(IOException ex){
+ flagMap = AconfigStorageReadAPI.mapStorageFile(mStorageDir + "/maps/mockup.flag.map");
+ } catch (IOException ex) {
assertTrue(ex.toString(), false);
}
- assertTrue(flagMap!= null);
+ assertTrue(flagMap != null);
class Baseline {
public int mPackageId;
@@ -110,10 +113,8 @@
public StoredFlagType mFlagType;
public int mFlagIndex;
- public Baseline(int packageId,
- String flagName,
- StoredFlagType flagType,
- int flagIndex) {
+ public Baseline(
+ int packageId, String flagName, StoredFlagType flagType, int flagIndex) {
mPackageId = packageId;
mFlagName = flagName;
mFlagType = flagType;
@@ -133,8 +134,9 @@
try {
for (Baseline baseline : baselines) {
- FlagReadContext context = AconfigStorageReadAPI.getFlagReadContext(
- flagMap, baseline.mPackageId, baseline.mFlagName);
+ FlagReadContext context =
+ AconfigStorageReadAPI.getFlagReadContext(
+ flagMap, baseline.mPackageId, baseline.mFlagName);
assertEquals(context.mFlagType, baseline.mFlagType);
assertEquals(context.mFlagIndex, baseline.mFlagIndex);
}
@@ -147,21 +149,19 @@
public void testNonExistFlagContextQuery() {
MappedByteBuffer flagMap = null;
try {
- flagMap = AconfigStorageReadAPI.mapStorageFile(
- mStorageDir + "/maps/mockup.flag.map");
- } catch(IOException ex){
+ flagMap = AconfigStorageReadAPI.mapStorageFile(mStorageDir + "/maps/mockup.flag.map");
+ } catch (IOException ex) {
assertTrue(ex.toString(), false);
}
- assertTrue(flagMap!= null);
+ assertTrue(flagMap != null);
try {
- FlagReadContext context = AconfigStorageReadAPI.getFlagReadContext(
- flagMap, 0, "unknown");
+ FlagReadContext context =
+ AconfigStorageReadAPI.getFlagReadContext(flagMap, 0, "unknown");
assertEquals(context.mFlagType, null);
assertEquals(context.mFlagIndex, -1);
- context = AconfigStorageReadAPI.getFlagReadContext(
- flagMap, 3, "enabled_ro");
+ context = AconfigStorageReadAPI.getFlagReadContext(flagMap, 3, "enabled_ro");
assertEquals(context.mFlagType, null);
assertEquals(context.mFlagIndex, -1);
} catch (IOException ex) {
@@ -173,12 +173,11 @@
public void testBooleanFlagValueQuery() {
MappedByteBuffer flagVal = null;
try {
- flagVal = AconfigStorageReadAPI.mapStorageFile(
- mStorageDir + "/boot/mockup.val");
+ flagVal = AconfigStorageReadAPI.mapStorageFile(mStorageDir + "/boot/mockup.val");
} catch (IOException ex) {
assertTrue(ex.toString(), false);
}
- assertTrue(flagVal!= null);
+ assertTrue(flagVal != null);
boolean[] baselines = {false, true, true, false, true, true, true, true};
for (int i = 0; i < 8; ++i) {
@@ -195,12 +194,11 @@
public void testInvalidBooleanFlagValueQuery() {
MappedByteBuffer flagVal = null;
try {
- flagVal = AconfigStorageReadAPI.mapStorageFile(
- mStorageDir + "/boot/mockup.val");
+ flagVal = AconfigStorageReadAPI.mapStorageFile(mStorageDir + "/boot/mockup.val");
} catch (IOException ex) {
assertTrue(ex.toString(), false);
}
- assertTrue(flagVal!= null);
+ assertTrue(flagVal != null);
try {
Boolean value = AconfigStorageReadAPI.getBooleanFlagValue(flagVal, 9);
@@ -210,4 +208,63 @@
assertTrue(ex.toString(), ex.toString().contains(expectedErrmsg));
}
}
- }
+
+ @Test
+ public void testRustJavaEqualHash() throws IOException {
+ List<parsed_flag> flags = DeviceProtos.loadAndParseFlagProtos();
+ for (parsed_flag flag : flags) {
+ String packageName = flag.package_;
+ String flagName = flag.name;
+ long rHash = AconfigStorageReadAPI.hash(packageName);
+ long jHash = SipHasher13.hash(packageName.getBytes());
+ assertEquals(rHash, jHash);
+
+ String fullFlagName = packageName + "/" + flagName;
+ rHash = AconfigStorageReadAPI.hash(fullFlagName);
+ jHash = SipHasher13.hash(fullFlagName.getBytes());
+ assertEquals(rHash, jHash);
+ }
+ }
+
+ @Test
+ public void testRustJavaEqualFlag() throws IOException {
+ List<parsed_flag> flags = DeviceProtos.loadAndParseFlagProtos();
+
+ String mapPath = "/metadata/aconfig/maps/";
+ String flagsPath = "/metadata/aconfig/boot/";
+
+ for (parsed_flag flag : flags) {
+
+ String container = flag.container;
+ String packageName = flag.package_;
+ String flagName = flag.name;
+ String fullFlagName = packageName + "/" + flagName;
+
+ MappedByteBuffer packageMap =
+ AconfigStorageReadAPI.mapStorageFile(mapPath + container + ".package.map");
+ MappedByteBuffer flagMap =
+ AconfigStorageReadAPI.mapStorageFile(mapPath + container + ".flag.map");
+ MappedByteBuffer flagValList =
+ AconfigStorageReadAPI.mapStorageFile(flagsPath + container + ".val");
+
+ PackageReadContext packageContext =
+ AconfigStorageReadAPI.getPackageReadContext(packageMap, packageName);
+
+ FlagReadContext flagContext =
+ AconfigStorageReadAPI.getFlagReadContext(
+ flagMap, packageContext.mPackageId, flagName);
+
+ boolean rVal =
+ AconfigStorageReadAPI.getBooleanFlagValue(
+ flagValList,
+ packageContext.mBooleanStartIndex + flagContext.mFlagIndex);
+
+ StorageInternalReader reader = new StorageInternalReader(container, packageName);
+ boolean jVal = reader.getBooleanFlagValue(flagContext.mFlagIndex);
+
+ long rHash = AconfigStorageReadAPI.hash(packageName);
+ long jHash = SipHasher13.hash(packageName.getBytes());
+ assertEquals(rVal, jVal);
+ }
+ }
+}
diff --git a/tools/aconfig/aconfig_storage_read_api/tests/java/Android.bp b/tools/aconfig/aconfig_storage_read_api/tests/java/Android.bp
index 11b3824..3d4e9ad 100644
--- a/tools/aconfig/aconfig_storage_read_api/tests/java/Android.bp
+++ b/tools/aconfig/aconfig_storage_read_api/tests/java/Android.bp
@@ -2,6 +2,8 @@
name: "aconfig_storage_read_api.test.java",
srcs: ["./**/*.java"],
static_libs: [
+ "aconfig_device_paths_java",
+ "aconfig_storage_file_java",
"aconfig_storage_reader_java",
"androidx.test.rules",
"libaconfig_storage_read_api_java",
diff --git a/tools/aconfig/aconfig_storage_write_api/aconfig_storage_write_api.cpp b/tools/aconfig/aconfig_storage_write_api/aconfig_storage_write_api.cpp
index 7b43574..03a8fa2 100644
--- a/tools/aconfig/aconfig_storage_write_api/aconfig_storage_write_api.cpp
+++ b/tools/aconfig/aconfig_storage_write_api/aconfig_storage_write_api.cpp
@@ -100,18 +100,4 @@
return {};
}
-android::base::Result<void> create_flag_info(
- std::string const& package_map,
- std::string const& flag_map,
- std::string const& flag_info_out) {
- auto creation_cxx = create_flag_info_cxx(
- rust::Str(package_map.c_str()),
- rust::Str(flag_map.c_str()),
- rust::Str(flag_info_out.c_str()));
- if (creation_cxx.success) {
- return {};
- } else {
- return android::base::Error() << creation_cxx.error_message;
- }
-}
} // namespace aconfig_storage
diff --git a/tools/aconfig/aconfig_storage_write_api/include/aconfig_storage/aconfig_storage_write_api.hpp b/tools/aconfig/aconfig_storage_write_api/include/aconfig_storage/aconfig_storage_write_api.hpp
index 0bba7ff..50a5188 100644
--- a/tools/aconfig/aconfig_storage_write_api/include/aconfig_storage/aconfig_storage_write_api.hpp
+++ b/tools/aconfig/aconfig_storage_write_api/include/aconfig_storage/aconfig_storage_write_api.hpp
@@ -36,13 +36,4 @@
uint32_t offset,
bool value);
-/// Create flag info file based on package and flag map
-/// \input package_map: package map file
-/// \input flag_map: flag map file
-/// \input flag_info_out: flag info file to be created
-android::base::Result<void> create_flag_info(
- std::string const& package_map,
- std::string const& flag_map,
- std::string const& flag_info_out);
-
} // namespace aconfig_storage
diff --git a/tools/aconfig/aconfig_storage_write_api/src/lib.rs b/tools/aconfig/aconfig_storage_write_api/src/lib.rs
index 0396a63..09bb41f 100644
--- a/tools/aconfig/aconfig_storage_write_api/src/lib.rs
+++ b/tools/aconfig/aconfig_storage_write_api/src/lib.rs
@@ -24,15 +24,10 @@
#[cfg(test)]
mod test_utils;
-use aconfig_storage_file::{
- AconfigStorageError, FlagInfoHeader, FlagInfoList, FlagInfoNode, FlagTable, FlagValueType,
- PackageTable, StorageFileType, StoredFlagType, FILE_VERSION,
-};
+use aconfig_storage_file::{AconfigStorageError, FlagValueType};
use anyhow::anyhow;
use memmap2::MmapMut;
-use std::fs::File;
-use std::io::{Read, Write};
/// Get read write mapped storage files.
///
@@ -104,86 +99,6 @@
})
}
-/// Read in storage file as bytes
-fn read_file_to_bytes(file_path: &str) -> Result<Vec<u8>, AconfigStorageError> {
- let mut file = File::open(file_path).map_err(|errmsg| {
- AconfigStorageError::FileReadFail(anyhow!("Failed to open file {}: {}", file_path, errmsg))
- })?;
- let mut buffer = Vec::new();
- file.read_to_end(&mut buffer).map_err(|errmsg| {
- AconfigStorageError::FileReadFail(anyhow!(
- "Failed to read bytes from file {}: {}",
- file_path,
- errmsg
- ))
- })?;
- Ok(buffer)
-}
-
-/// Create flag info file given package map file and flag map file
-/// \input package_map: package map file
-/// \input flag_map: flag map file
-/// \output flag_info_out: created flag info file
-pub fn create_flag_info(
- package_map: &str,
- flag_map: &str,
- flag_info_out: &str,
-) -> Result<(), AconfigStorageError> {
- let package_table = PackageTable::from_bytes(&read_file_to_bytes(package_map)?)?;
- let flag_table = FlagTable::from_bytes(&read_file_to_bytes(flag_map)?)?;
-
- if package_table.header.container != flag_table.header.container {
- return Err(AconfigStorageError::FileCreationFail(anyhow!(
- "container for package map {} and flag map {} does not match",
- package_table.header.container,
- flag_table.header.container,
- )));
- }
-
- let mut package_start_index = vec![0; package_table.header.num_packages as usize];
- for node in package_table.nodes.iter() {
- package_start_index[node.package_id as usize] = node.boolean_start_index;
- }
-
- let mut is_flag_rw = vec![false; flag_table.header.num_flags as usize];
- for node in flag_table.nodes.iter() {
- let flag_index = package_start_index[node.package_id as usize] + node.flag_index as u32;
- is_flag_rw[flag_index as usize] = node.flag_type == StoredFlagType::ReadWriteBoolean;
- }
-
- let mut list = FlagInfoList {
- header: FlagInfoHeader {
- version: FILE_VERSION,
- container: flag_table.header.container,
- file_type: StorageFileType::FlagInfo as u8,
- file_size: 0,
- num_flags: flag_table.header.num_flags,
- boolean_flag_offset: 0,
- },
- nodes: is_flag_rw.iter().map(|&rw| FlagInfoNode::create(rw)).collect(),
- };
-
- list.header.boolean_flag_offset = list.header.into_bytes().len() as u32;
- list.header.file_size = list.into_bytes().len() as u32;
-
- let mut file = File::create(flag_info_out).map_err(|errmsg| {
- AconfigStorageError::FileCreationFail(anyhow!(
- "fail to create file {}: {}",
- flag_info_out,
- errmsg
- ))
- })?;
- file.write_all(&list.into_bytes()).map_err(|errmsg| {
- AconfigStorageError::FileCreationFail(anyhow!(
- "fail to write to file {}: {}",
- flag_info_out,
- errmsg
- ))
- })?;
-
- Ok(())
-}
-
// *************************************** //
// CC INTERLOP
// *************************************** //
@@ -212,12 +127,6 @@
pub error_message: String,
}
- // Flag info file creation return for cc interlop
- pub struct FlagInfoCreationCXX {
- pub success: bool,
- pub error_message: String,
- }
-
// Rust export to c++
extern "Rust" {
pub fn update_boolean_flag_value_cxx(
@@ -239,12 +148,6 @@
offset: u32,
value: bool,
) -> FlagHasLocalOverrideUpdateCXX;
-
- pub fn create_flag_info_cxx(
- package_map: &str,
- flag_map: &str,
- flag_info_out: &str,
- ) -> FlagInfoCreationCXX;
}
}
@@ -329,34 +232,15 @@
}
}
-/// Create flag info file cc interlop
-pub(crate) fn create_flag_info_cxx(
- package_map: &str,
- flag_map: &str,
- flag_info_out: &str,
-) -> ffi::FlagInfoCreationCXX {
- match create_flag_info(package_map, flag_map, flag_info_out) {
- Ok(()) => ffi::FlagInfoCreationCXX { success: true, error_message: String::from("") },
- Err(errmsg) => {
- ffi::FlagInfoCreationCXX { success: false, error_message: format!("{:?}", errmsg) }
- }
- }
-}
-
#[cfg(test)]
mod tests {
use super::*;
use crate::test_utils::copy_to_temp_file;
- use aconfig_storage_file::test_utils::{
- create_test_flag_info_list, create_test_flag_table, create_test_package_table,
- write_bytes_to_temp_file,
- };
use aconfig_storage_file::FlagInfoBit;
use aconfig_storage_read_api::flag_info_query::find_flag_attribute;
use aconfig_storage_read_api::flag_value_query::find_boolean_flag_value;
use std::fs::File;
use std::io::Read;
- use tempfile::NamedTempFile;
fn get_boolean_flag_value_at_offset(file: &str, offset: u32) -> bool {
let mut f = File::open(&file).unwrap();
@@ -439,31 +323,4 @@
}
}
}
-
- fn create_empty_temp_file() -> Result<NamedTempFile, AconfigStorageError> {
- let file = NamedTempFile::new().map_err(|_| {
- AconfigStorageError::FileCreationFail(anyhow!("Failed to create temp file"))
- })?;
- Ok(file)
- }
-
- #[test]
- // this test point locks down the flag info creation
- fn test_create_flag_info() {
- let package_table =
- write_bytes_to_temp_file(&create_test_package_table().into_bytes()).unwrap();
- let flag_table = write_bytes_to_temp_file(&create_test_flag_table().into_bytes()).unwrap();
- let flag_info = create_empty_temp_file().unwrap();
-
- let package_table_path = package_table.path().display().to_string();
- let flag_table_path = flag_table.path().display().to_string();
- let flag_info_path = flag_info.path().display().to_string();
-
- assert!(create_flag_info(&package_table_path, &flag_table_path, &flag_info_path).is_ok());
-
- let flag_info =
- FlagInfoList::from_bytes(&read_file_to_bytes(&flag_info_path).unwrap()).unwrap();
- let expected_flag_info = create_test_flag_info_list();
- assert_eq!(flag_info, expected_flag_info);
- }
}
diff --git a/tools/aconfig/aflags/Android.bp b/tools/aconfig/aflags/Android.bp
index c48585a..2040cc6 100644
--- a/tools/aconfig/aflags/Android.bp
+++ b/tools/aconfig/aflags/Android.bp
@@ -10,6 +10,7 @@
srcs: ["src/main.rs"],
rustlibs: [
"libaconfig_device_paths",
+ "libaconfig_flags",
"libaconfig_protos",
"libaconfigd_protos",
"libaconfig_storage_read_api",
@@ -24,6 +25,7 @@
rust_binary {
name: "aflags",
+ host_supported: true,
defaults: ["aflags.defaults"],
}
diff --git a/tools/aconfig/aflags/Cargo.toml b/tools/aconfig/aflags/Cargo.toml
index 7dc3436..7efce6d 100644
--- a/tools/aconfig/aflags/Cargo.toml
+++ b/tools/aconfig/aflags/Cargo.toml
@@ -15,3 +15,4 @@
aconfig_storage_read_api = { version = "0.1.0", path = "../aconfig_storage_read_api" }
clap = {version = "4.5.2" }
aconfig_device_paths = { version = "0.1.0", path = "../aconfig_device_paths" }
+aconfig_flags = { version = "0.1.0", path = "../aconfig_flags" }
\ No newline at end of file
diff --git a/tools/aconfig/aflags/src/aconfig_storage_source.rs b/tools/aconfig/aflags/src/aconfig_storage_source.rs
index a2c6012..68edf7d 100644
--- a/tools/aconfig/aflags/src/aconfig_storage_source.rs
+++ b/tools/aconfig/aflags/src/aconfig_storage_source.rs
@@ -1,3 +1,4 @@
+use crate::load_protos;
use crate::{Flag, FlagSource};
use crate::{FlagPermission, FlagValue, ValuePickedFrom};
use aconfigd_protos::{
@@ -9,13 +10,18 @@
use anyhow::Result;
use protobuf::Message;
use protobuf::SpecialFields;
+use std::collections::HashMap;
use std::io::{Read, Write};
use std::net::Shutdown;
use std::os::unix::net::UnixStream;
pub struct AconfigStorageSource {}
-fn convert(msg: ProtoFlagQueryReturnMessage) -> Result<Flag> {
+fn load_flag_to_container() -> Result<HashMap<String, String>> {
+ Ok(load_protos::load()?.into_iter().map(|p| (p.qualified_name(), p.container)).collect())
+}
+
+fn convert(msg: ProtoFlagQueryReturnMessage, containers: &HashMap<String, String>) -> Result<Flag> {
let (value, value_picked_from) = match (
&msg.boot_flag_value,
msg.default_flag_value,
@@ -55,15 +61,21 @@
None => return Err(anyhow!("missing permission")),
};
+ let name = msg.flag_name.ok_or(anyhow!("missing flag name"))?;
+ let package = msg.package_name.ok_or(anyhow!("missing package name"))?;
+ let qualified_name = format!("{package}.{name}");
Ok(Flag {
- name: msg.flag_name.ok_or(anyhow!("missing flag name"))?,
- package: msg.package_name.ok_or(anyhow!("missing package name"))?,
+ name,
+ package,
value,
permission,
value_picked_from,
staged_value,
- container: msg.container.ok_or(anyhow!("missing container"))?,
-
+ container: containers
+ .get(&qualified_name)
+ .cloned()
+ .unwrap_or_else(|| "<no container>".to_string())
+ .to_string(),
// TODO: remove once DeviceConfig is not in the CLI.
namespace: "-".to_string(),
})
@@ -114,9 +126,13 @@
impl FlagSource for AconfigStorageSource {
fn list_flags() -> Result<Vec<Flag>> {
+ let containers = load_flag_to_container()?;
read_from_socket()
.map(|query_messages| {
- query_messages.iter().map(|message| convert(message.clone())).collect::<Vec<_>>()
+ query_messages
+ .iter()
+ .map(|message| convert(message.clone(), &containers))
+ .collect::<Vec<_>>()
})?
.into_iter()
.collect()
diff --git a/tools/aconfig/aflags/src/main.rs b/tools/aconfig/aflags/src/main.rs
index d8912a9..07b7243 100644
--- a/tools/aconfig/aflags/src/main.rs
+++ b/tools/aconfig/aflags/src/main.rs
@@ -116,9 +116,10 @@
}
fn display_staged_value(&self) -> String {
- match self.staged_value {
- Some(v) => format!("(->{})", v),
- None => "-".to_string(),
+ match (&self.permission, self.staged_value) {
+ (FlagPermission::ReadOnly, _) => "-".to_string(),
+ (FlagPermission::ReadWrite, None) => "-".to_string(),
+ (FlagPermission::ReadWrite, Some(v)) => format!("(->{})", v),
}
}
}
@@ -164,10 +165,6 @@
enum Command {
/// List all aconfig flags on this device.
List {
- /// Read from the new flag storage.
- #[clap(long)]
- use_new_storage: bool,
-
/// Optionally filter by container name.
#[clap(short = 'c', long = "container")]
container: Option<String>,
@@ -184,6 +181,9 @@
/// <package>.<flag_name>
qualified_name: String,
},
+
+ /// Display which flag storage backs aconfig flags.
+ WhichBacking,
}
struct PaddingInfo {
@@ -282,21 +282,31 @@
Ok(result)
}
+fn display_which_backing() -> String {
+ if aconfig_flags::auto_generated::enable_only_new_storage() {
+ "aconfig_storage".to_string()
+ } else {
+ "device_config".to_string()
+ }
+}
+
fn main() -> Result<()> {
ensure!(nix::unistd::Uid::current().is_root(), "must be root");
let cli = Cli::parse();
let output = match cli.command {
- Command::List { use_new_storage: true, container } => {
- list(FlagSourceType::AconfigStorage, container)
- .map_err(|err| anyhow!("storage may not be enabled: {err}"))
- .map(Some)
- }
- Command::List { use_new_storage: false, container } => {
- list(FlagSourceType::DeviceConfig, container).map(Some)
+ Command::List { container } => {
+ if aconfig_flags::auto_generated::enable_only_new_storage() {
+ list(FlagSourceType::AconfigStorage, container)
+ .map_err(|err| anyhow!("storage may not be enabled: {err}"))
+ .map(Some)
+ } else {
+ list(FlagSourceType::DeviceConfig, container).map(Some)
+ }
}
Command::Enable { qualified_name } => set_flag(&qualified_name, "true").map(|_| None),
Command::Disable { qualified_name } => set_flag(&qualified_name, "false").map(|_| None),
+ Command::WhichBacking => Ok(Some(display_which_backing())),
};
match output {
Ok(Some(text)) => println!("{text}"),
diff --git a/tools/aconfig/fake_device_config/Android.bp b/tools/aconfig/fake_device_config/Android.bp
index 7704742..1f17e6b 100644
--- a/tools/aconfig/fake_device_config/Android.bp
+++ b/tools/aconfig/fake_device_config/Android.bp
@@ -15,9 +15,7 @@
java_library {
name: "fake_device_config",
srcs: [
- "src/android/util/Log.java",
- "src/android/provider/DeviceConfig.java",
- "src/android/os/StrictMode.java",
+ "src/**/*.java",
],
sdk_version: "none",
system_modules: "core-all-system-modules",
diff --git a/tools/aconfig/fake_device_config/src/android/provider/AconfigPackage.java b/tools/aconfig/fake_device_config/src/android/provider/AconfigPackage.java
new file mode 100644
index 0000000..2f01b8c
--- /dev/null
+++ b/tools/aconfig/fake_device_config/src/android/provider/AconfigPackage.java
@@ -0,0 +1,42 @@
+/*
+ * Copyright (C) 2024 The Android Open Source Project
+ *
+ * Licensed under the Apache License, Version 2.0 (the "License");
+ * you may not use this file except in compliance with the License.
+ * You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package android.provider;
+
+/*
+ * This class allows generated aconfig code to compile independently of the framework.
+ */
+public class AconfigPackage {
+
+ /** Flag value is true */
+ public static final int FLAG_BOOLEAN_VALUE_TRUE = 1;
+
+ /** Flag value is false */
+ public static final int FLAG_BOOLEAN_VALUE_FALSE = 0;
+
+ /** Flag value doesn't exist */
+ public static final int FLAG_BOOLEAN_VALUE_NOT_EXIST = 2;
+
+ public static int getBooleanFlagValue(String packageName, String flagName) {
+ return 0;
+ }
+
+ public AconfigPackage(String packageName) {}
+
+ public int getBooleanFlagValue(String flagName) {
+ return 0;
+ }
+}
\ No newline at end of file
diff --git a/tools/aconfig/fake_device_config/src/android/util/Log.java b/tools/aconfig/fake_device_config/src/android/util/Log.java
index 3e7fd0f..79de680 100644
--- a/tools/aconfig/fake_device_config/src/android/util/Log.java
+++ b/tools/aconfig/fake_device_config/src/android/util/Log.java
@@ -5,6 +5,10 @@
return 0;
}
+ public static int w(String tag, String msg) {
+ return 0;
+ }
+
public static int e(String tag, String msg) {
return 0;
}
diff --git a/tools/auto_gen_test_config.py b/tools/auto_gen_test_config.py
index 8ee599a..d54c412 100755
--- a/tools/auto_gen_test_config.py
+++ b/tools/auto_gen_test_config.py
@@ -34,6 +34,7 @@
PLACEHOLDER_PACKAGE = '{PACKAGE}'
PLACEHOLDER_RUNNER = '{RUNNER}'
PLACEHOLDER_TEST_TYPE = '{TEST_TYPE}'
+PLACEHOLDER_EXTRA_TEST_RUNNER_CONFIGS = '{EXTRA_TEST_RUNNER_CONFIGS}'
def main(argv):
@@ -59,6 +60,7 @@
"instrumentation_test_config_template",
help="Path to the instrumentation test config template.")
parser.add_argument("--extra-configs", default="")
+ parser.add_argument("--extra-test-runner-configs", default="")
args = parser.parse_args(argv)
target_config = args.target_config
@@ -66,6 +68,7 @@
empty_config = args.empty_config
instrumentation_test_config_template = args.instrumentation_test_config_template
extra_configs = '\n'.join(args.extra_configs.split('\\n'))
+ extra_test_runner_configs = '\n'.join(args.extra_test_runner_configs.split('\\n'))
module = os.path.splitext(os.path.basename(target_config))[0]
@@ -131,6 +134,7 @@
config = config.replace(PLACEHOLDER_PACKAGE, package)
config = config.replace(PLACEHOLDER_TEST_TYPE, test_type)
config = config.replace(PLACEHOLDER_EXTRA_CONFIGS, extra_configs)
+ config = config.replace(PLACEHOLDER_EXTRA_TEST_RUNNER_CONFIGS, extra_test_runner_configs)
config = config.replace(PLACEHOLDER_RUNNER, runner)
with open(target_config, 'w') as config_file:
config_file.write(config)
diff --git a/tools/edit_monitor/Android.bp b/tools/edit_monitor/Android.bp
new file mode 100644
index 0000000..b939633
--- /dev/null
+++ b/tools/edit_monitor/Android.bp
@@ -0,0 +1,44 @@
+// Copyright 2024 The Android Open Source Project
+//
+// Licensed under the Apache License, Version 2.0 (the "License");
+// you may not use this file except in compliance with the License.
+// You may obtain a copy of the License at
+//
+// http://www.apache.org/licenses/LICENSE-2.0
+//
+// Unless required by applicable law or agreed to in writing, software
+// distributed under the License is distributed on an "AS IS" BASIS,
+// WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+// See the License for the specific language governing permissions and
+// limitations under the License.
+
+// Set of error prone rules to ensure code quality
+// PackageLocation check requires the androidCompatible=false otherwise it does not do anything.
+
+package {
+ default_applicable_licenses: ["Android-Apache-2.0"],
+ default_team: "trendy_team_adte",
+}
+
+python_library_host {
+ name: "edit_monitor_lib",
+ pkg_path: "edit_monitor",
+ srcs: [
+ "daemon_manager.py",
+ ],
+}
+
+python_test_host {
+ name: "daemon_manager_test",
+ main: "daemon_manager_test.py",
+ pkg_path: "edit_monitor",
+ srcs: [
+ "daemon_manager_test.py",
+ ],
+ libs: [
+ "edit_monitor_lib",
+ ],
+ test_options: {
+ unit_test: true,
+ },
+}
diff --git a/tools/edit_monitor/OWNERS b/tools/edit_monitor/OWNERS
new file mode 100644
index 0000000..8f0f364
--- /dev/null
+++ b/tools/edit_monitor/OWNERS
@@ -0,0 +1 @@
+include platform/tools/asuite:/OWNERS_ADTE_TEAM
\ No newline at end of file
diff --git a/tools/edit_monitor/daemon_manager.py b/tools/edit_monitor/daemon_manager.py
new file mode 100644
index 0000000..1876451
--- /dev/null
+++ b/tools/edit_monitor/daemon_manager.py
@@ -0,0 +1,302 @@
+# Copyright 2024, The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+
+import hashlib
+import logging
+import multiprocessing
+import os
+import pathlib
+import signal
+import subprocess
+import sys
+import tempfile
+import time
+
+
+DEFAULT_PROCESS_TERMINATION_TIMEOUT_SECONDS = 1
+DEFAULT_MONITOR_INTERVAL_SECONDS = 5
+DEFAULT_MEMORY_USAGE_THRESHOLD = 2000
+DEFAULT_CPU_USAGE_THRESHOLD = 200
+DEFAULT_REBOOT_TIMEOUT_SECONDS = 60 * 60 * 24
+
+
+def default_daemon_target():
+ """Place holder for the default daemon target."""
+ print("default daemon target")
+
+
+class DaemonManager:
+ """Class to manage and monitor the daemon run as a subprocess."""
+
+ def __init__(
+ self,
+ binary_path: str,
+ daemon_target: callable = default_daemon_target,
+ daemon_args: tuple = (),
+ ):
+ self.binary_path = binary_path
+ self.daemon_target = daemon_target
+ self.daemon_args = daemon_args
+
+ self.pid = os.getpid()
+ self.daemon_process = None
+
+ self.max_memory_usage = 0
+ self.max_cpu_usage = 0
+
+ pid_file_dir = pathlib.Path(tempfile.gettempdir()).joinpath("edit_monitor")
+ pid_file_dir.mkdir(parents=True, exist_ok=True)
+ self.pid_file_path = self._get_pid_file_path(pid_file_dir)
+
+ def start(self):
+ """Writes the pidfile and starts the daemon proces."""
+ try:
+ self._stop_any_existing_instance()
+ self._write_pid_to_pidfile()
+ self._start_daemon_process()
+ except Exception as e:
+ logging.exception("Failed to start daemon manager with error %s", e)
+
+ def monitor_daemon(
+ self,
+ interval: int = DEFAULT_MONITOR_INTERVAL_SECONDS,
+ memory_threshold: float = DEFAULT_MEMORY_USAGE_THRESHOLD,
+ cpu_threshold: float = DEFAULT_CPU_USAGE_THRESHOLD,
+ reboot_timeout: int = DEFAULT_REBOOT_TIMEOUT_SECONDS,
+ ):
+ """Monits the daemon process status.
+
+ Periodically check the CPU/Memory usage of the daemon process as long as the
+ process is still running and kill the process if the resource usage is above
+ given thresholds.
+ """
+ logging.info("start monitoring daemon process %d.", self.daemon_process.pid)
+ reboot_time = time.time() + reboot_timeout
+ while self.daemon_process.is_alive():
+ if time.time() > reboot_time:
+ self.reboot()
+ try:
+ memory_usage = self._get_process_memory_percent(self.daemon_process.pid)
+ self.max_memory_usage = max(self.max_memory_usage, memory_usage)
+
+ cpu_usage = self._get_process_cpu_percent(self.daemon_process.pid)
+ self.max_cpu_usage = max(self.max_cpu_usage, cpu_usage)
+
+ time.sleep(interval)
+ except Exception as e:
+ # Logging the error and continue.
+ logging.warning("Failed to monitor daemon process with error: %s", e)
+
+ if (
+ self.max_memory_usage >= memory_threshold
+ or self.max_cpu_usage >= cpu_threshold
+ ):
+ logging.error(
+ "Daemon process is consuming too much resource, killing..."
+ ),
+ self._terminate_process(self.daemon_process.pid)
+
+ logging.info(
+ "Daemon process %d terminated. Max memory usage: %f, Max cpu"
+ " usage: %f.",
+ self.daemon_process.pid,
+ self.max_memory_usage,
+ self.max_cpu_usage,
+ )
+
+ def stop(self):
+ """Stops the daemon process and removes the pidfile."""
+
+ logging.debug("in daemon manager cleanup.")
+ try:
+ if self.daemon_process and self.daemon_process.is_alive():
+ self._terminate_process(self.daemon_process.pid)
+ self._remove_pidfile()
+ logging.debug("Successfully stopped daemon manager.")
+ except Exception as e:
+ logging.exception("Failed to stop daemon manager with error %s", e)
+
+ def reboot(self):
+ """Reboots the current process.
+
+ Stops the current daemon manager and reboots the entire process based on
+ the binary file. Exits directly If the binary file no longer exists.
+ """
+ logging.debug("Rebooting process based on binary %s.", self.binary_path)
+
+ # Stop the current daemon manager first.
+ self.stop()
+
+ # If the binary no longer exists, exit directly.
+ if not os.path.exists(self.binary_path):
+ logging.info("binary %s no longer exists, exiting.", self.binary_path)
+ sys.exit(0)
+
+ try:
+ os.execv(self.binary_path, sys.argv)
+ except OSError as e:
+ logging.exception("Failed to reboot process with error: %s.", e)
+ sys.exit(1) # Indicate an error occurred
+
+ def _stop_any_existing_instance(self):
+ if not self.pid_file_path.exists():
+ logging.debug("No existing instances.")
+ return
+
+ ex_pid = self._read_pid_from_pidfile()
+
+ if ex_pid:
+ logging.info("Found another instance with pid %d.", ex_pid)
+ self._terminate_process(ex_pid)
+ self._remove_pidfile()
+
+ def _read_pid_from_pidfile(self):
+ with open(self.pid_file_path, "r") as f:
+ return int(f.read().strip())
+
+ def _write_pid_to_pidfile(self):
+ """Creates a pidfile and writes the current pid to the file.
+
+ Raise FileExistsError if the pidfile already exists.
+ """
+ try:
+ # Use the 'x' mode to open the file for exclusive creation
+ with open(self.pid_file_path, "x") as f:
+ f.write(f"{self.pid}")
+ except FileExistsError as e:
+ # This could be caused due to race condition that a user is trying
+ # to start two edit monitors at the same time. Or because there is
+ # already an existing edit monitor running and we can not kill it
+ # for some reason.
+ logging.exception("pidfile %s already exists.", self.pid_file_path)
+ raise e
+
+ def _start_daemon_process(self):
+ """Starts a subprocess to run the daemon."""
+ p = multiprocessing.Process(
+ target=self.daemon_target, args=self.daemon_args
+ )
+ p.start()
+
+ logging.info("Start subprocess with PID %d", p.pid)
+ self.daemon_process = p
+
+ def _terminate_process(
+ self, pid: int, timeout: int = DEFAULT_PROCESS_TERMINATION_TIMEOUT_SECONDS
+ ):
+ """Terminates a process with given pid.
+
+ It first sends a SIGTERM to the process to allow it for proper
+ termination with a timeout. If the process is not terminated within
+ the timeout, kills it forcefully.
+ """
+ try:
+ os.kill(pid, signal.SIGTERM)
+ if not self._wait_for_process_terminate(pid, timeout):
+ logging.warning(
+ "Process %d not terminated within timeout, try force kill", pid
+ )
+ os.kill(pid, signal.SIGKILL)
+ except ProcessLookupError:
+ logging.info("Process with PID %d not found (already terminated)", pid)
+
+ def _wait_for_process_terminate(self, pid: int, timeout: int) -> bool:
+ start_time = time.time()
+
+ while time.time() < start_time + timeout:
+ if not self._is_process_alive(pid):
+ return True
+ time.sleep(1)
+
+ logging.error("Process %d not terminated within %d seconds.", pid, timeout)
+ return False
+
+ def _is_process_alive(self, pid: int) -> bool:
+ try:
+ output = subprocess.check_output(
+ ["ps", "-p", str(pid), "-o", "state="], text=True
+ ).strip()
+ state = output.split()[0]
+ return state != "Z" # Check if the state is not 'Z' (zombie)
+ except subprocess.CalledProcessError:
+ # Process not found (already dead).
+ return False
+ except (FileNotFoundError, OSError, ValueError) as e:
+ logging.warning(
+ "Unable to check the status for process %d with error: %s.", pid, e
+ )
+ return True
+
+ def _remove_pidfile(self):
+ try:
+ os.remove(self.pid_file_path)
+ except FileNotFoundError:
+ logging.info("pid file %s already removed.", self.pid_file_path)
+
+ def _get_pid_file_path(self, pid_file_dir: pathlib.Path) -> pathlib.Path:
+ """Generates the path to store the pidfile.
+
+ The file path should have the format of "/tmp/edit_monitor/xxxx.lock"
+ where xxxx is a hashed value based on the binary path that starts the
+ process.
+ """
+ hash_object = hashlib.sha256()
+ hash_object.update(self.binary_path.encode("utf-8"))
+ pid_file_path = pid_file_dir.joinpath(hash_object.hexdigest() + ".lock")
+ logging.info("pid_file_path: %s", pid_file_path)
+
+ return pid_file_path
+
+ def _get_process_memory_percent(self, pid: int) -> float:
+ try:
+ with open(f"/proc/{pid}/stat", "r") as f:
+ stat_data = f.readline().split()
+ # RSS is the 24th field in /proc/[pid]/stat
+ rss_pages = int(stat_data[23])
+ return rss_pages * 4 / 1024 # Covert to MB
+ except (FileNotFoundError, IndexError, ValueError, IOError) as e:
+ logging.exception("Failed to get memory usage.")
+ raise e
+
+ def _get_process_cpu_percent(self, pid: int, interval: int = 1) -> float:
+ try:
+ total_start_time = self._get_total_cpu_time(pid)
+ with open("/proc/uptime", "r") as f:
+ uptime_start = float(f.readline().split()[0])
+
+ time.sleep(interval)
+
+ total_end_time = self._get_total_cpu_time(pid)
+ with open("/proc/uptime", "r") as f:
+ uptime_end = float(f.readline().split()[0])
+
+ return (
+ (total_end_time - total_start_time)
+ / (uptime_end - uptime_start)
+ * 100
+ )
+ except (FileNotFoundError, IndexError, ValueError, IOError) as e:
+ logging.exception("Failed to get CPU usage.")
+ raise e
+
+ def _get_total_cpu_time(self, pid: int) -> float:
+ with open(f"/proc/{str(pid)}/stat", "r") as f:
+ stats = f.readline().split()
+ # utime is the 14th field in /proc/[pid]/stat measured in clock ticks.
+ utime = int(stats[13])
+ # stime is the 15th field in /proc/[pid]/stat measured in clock ticks.
+ stime = int(stats[14])
+ return (utime + stime) / os.sysconf(os.sysconf_names["SC_CLK_TCK"])
+
diff --git a/tools/edit_monitor/daemon_manager_test.py b/tools/edit_monitor/daemon_manager_test.py
new file mode 100644
index 0000000..bcfa850
--- /dev/null
+++ b/tools/edit_monitor/daemon_manager_test.py
@@ -0,0 +1,360 @@
+# Copyright 2024, The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+"""Unittests for DaemonManager."""
+
+import logging
+import multiprocessing
+import os
+import pathlib
+import signal
+import subprocess
+import sys
+import tempfile
+import time
+import unittest
+from unittest import mock
+from edit_monitor import daemon_manager
+
+TEST_BINARY_FILE = '/path/to/test_binary'
+TEST_PID_FILE_PATH = (
+ '587239c2d1050afdf54512e2d799f3b929f86b43575eb3c7b4bab105dd9bd25e.lock'
+)
+
+
+def simple_daemon(output_file):
+ with open(output_file, 'w') as f:
+ f.write('running daemon target')
+
+
+def long_running_daemon():
+ while True:
+ time.sleep(1)
+
+
+def memory_consume_daemon_target(size_mb):
+ try:
+ size_bytes = size_mb * 1024 * 1024
+ dummy_data = bytearray(size_bytes)
+ time.sleep(10)
+ except MemoryError:
+ print(f'Process failed to allocate {size_mb} MB of memory.')
+
+
+def cpu_consume_daemon_target(target_usage_percent):
+ while True:
+ start_time = time.time()
+ while time.time() - start_time < target_usage_percent / 100:
+ pass # Busy loop to consume CPU
+
+ # Sleep to reduce CPU usage
+ time.sleep(1 - target_usage_percent / 100)
+
+
+class DaemonManagerTest(unittest.TestCase):
+
+ @classmethod
+ def setUpClass(cls):
+ super().setUpClass()
+ # Configure to print logging to stdout.
+ logging.basicConfig(filename=None, level=logging.DEBUG)
+ console = logging.StreamHandler(sys.stdout)
+ logging.getLogger('').addHandler(console)
+
+ def setUp(self):
+ super().setUp()
+ self.original_tempdir = tempfile.tempdir
+ self.working_dir = tempfile.TemporaryDirectory()
+ # Sets the tempdir under the working dir so any temp files created during
+ # tests will be cleaned.
+ tempfile.tempdir = self.working_dir.name
+
+ def tearDown(self):
+ # Cleans up any child processes left by the tests.
+ self._cleanup_child_processes()
+ self.working_dir.cleanup()
+ # Restores tempdir.
+ tempfile.tempdir = self.original_tempdir
+ super().tearDown()
+
+ def test_start_success_with_no_existing_instance(self):
+ self.assert_run_simple_daemon_success()
+
+ def test_start_success_with_existing_instance_running(self):
+ # Create a long running subprocess
+ p = multiprocessing.Process(target=long_running_daemon)
+ p.start()
+
+ # Create a pidfile with the subprocess pid
+ pid_file_path_dir = pathlib.Path(self.working_dir.name).joinpath(
+ 'edit_monitor'
+ )
+ pid_file_path_dir.mkdir(parents=True, exist_ok=True)
+ with open(pid_file_path_dir.joinpath(TEST_PID_FILE_PATH), 'w') as f:
+ f.write(str(p.pid))
+
+ self.assert_run_simple_daemon_success()
+ p.terminate()
+
+ def test_start_success_with_existing_instance_already_dead(self):
+ # Create a pidfile with pid that does not exist.
+ pid_file_path_dir = pathlib.Path(self.working_dir.name).joinpath(
+ 'edit_monitor'
+ )
+ pid_file_path_dir.mkdir(parents=True, exist_ok=True)
+ with open(pid_file_path_dir.joinpath(TEST_PID_FILE_PATH), 'w') as f:
+ f.write('123456')
+
+ self.assert_run_simple_daemon_success()
+
+ def test_start_success_with_existing_instance_from_different_binary(self):
+ # First start an instance based on "some_binary_path"
+ existing_dm = daemon_manager.DaemonManager(
+ 'some_binary_path',
+ daemon_target=long_running_daemon,
+ )
+ existing_dm.start()
+
+ self.assert_run_simple_daemon_success()
+ existing_dm.stop()
+
+ @mock.patch('os.kill')
+ def test_start_failed_to_kill_existing_instance(self, mock_kill):
+ mock_kill.side_effect = OSError('Unknown OSError')
+ pid_file_path_dir = pathlib.Path(self.working_dir.name).joinpath(
+ 'edit_monitor'
+ )
+ pid_file_path_dir.mkdir(parents=True, exist_ok=True)
+ with open(pid_file_path_dir.joinpath(TEST_PID_FILE_PATH), 'w') as f:
+ f.write('123456')
+
+ dm = daemon_manager.DaemonManager(TEST_BINARY_FILE)
+ dm.start()
+
+ # Verify no daemon process is started.
+ self.assertIsNone(dm.daemon_process)
+
+ def test_start_failed_to_write_pidfile(self):
+ pid_file_path_dir = pathlib.Path(self.working_dir.name).joinpath(
+ 'edit_monitor'
+ )
+ pid_file_path_dir.mkdir(parents=True, exist_ok=True)
+ # Makes the directory read-only so write pidfile will fail.
+ os.chmod(pid_file_path_dir, 0o555)
+
+ dm = daemon_manager.DaemonManager(TEST_BINARY_FILE)
+ dm.start()
+
+ # Verifies no daemon process is started.
+ self.assertIsNone(dm.daemon_process)
+
+ def test_start_failed_to_start_daemon_process(self):
+ dm = daemon_manager.DaemonManager(
+ TEST_BINARY_FILE, daemon_target='wrong_target', daemon_args=(1)
+ )
+ dm.start()
+
+ # Verifies no daemon process is started.
+ self.assertIsNone(dm.daemon_process)
+
+ def test_monitor_daemon_subprocess_killed_high_memory_usage(self):
+ dm = daemon_manager.DaemonManager(
+ TEST_BINARY_FILE,
+ daemon_target=memory_consume_daemon_target,
+ daemon_args=(2,),
+ )
+ dm.start()
+ dm.monitor_daemon(interval=1, memory_threshold=2)
+
+ self.assertTrue(dm.max_memory_usage >= 2)
+ self.assert_no_subprocess_running()
+
+ def test_monitor_daemon_subprocess_killed_high_cpu_usage(self):
+ dm = daemon_manager.DaemonManager(
+ TEST_BINARY_FILE,
+ daemon_target=cpu_consume_daemon_target,
+ daemon_args=(20,),
+ )
+ dm.start()
+ dm.monitor_daemon(interval=1, cpu_threshold=20)
+
+ self.assertTrue(dm.max_cpu_usage >= 20)
+ self.assert_no_subprocess_running()
+
+ @mock.patch('subprocess.check_output')
+ def test_monitor_daemon_failed_does_not_matter(self, mock_output):
+ mock_output.side_effect = OSError('Unknown OSError')
+ self.assert_run_simple_daemon_success()
+
+ @mock.patch('os.execv')
+ def test_monitor_daemon_reboot_triggered(self, mock_execv):
+ binary_file = tempfile.NamedTemporaryFile(
+ dir=self.working_dir.name, delete=False
+ )
+
+ dm = daemon_manager.DaemonManager(
+ binary_file.name, daemon_target=long_running_daemon
+ )
+ dm.start()
+ dm.monitor_daemon(reboot_timeout=0.5)
+ mock_execv.assert_called_once()
+
+ def test_stop_success(self):
+ dm = daemon_manager.DaemonManager(
+ TEST_BINARY_FILE, daemon_target=long_running_daemon
+ )
+ dm.start()
+ dm.stop()
+
+ self.assert_no_subprocess_running()
+ self.assertFalse(dm.pid_file_path.exists())
+
+ @mock.patch('os.kill')
+ def test_stop_failed_to_kill_daemon_process(self, mock_kill):
+ mock_kill.side_effect = OSError('Unknown OSError')
+ dm = daemon_manager.DaemonManager(
+ TEST_BINARY_FILE, daemon_target=long_running_daemon
+ )
+ dm.start()
+ dm.stop()
+
+ self.assertTrue(dm.daemon_process.is_alive())
+ self.assertTrue(dm.pid_file_path.exists())
+
+ @mock.patch('os.remove')
+ def test_stop_failed_to_remove_pidfile(self, mock_remove):
+ mock_remove.side_effect = OSError('Unknown OSError')
+
+ dm = daemon_manager.DaemonManager(
+ TEST_BINARY_FILE, daemon_target=long_running_daemon
+ )
+ dm.start()
+ dm.stop()
+
+ self.assert_no_subprocess_running()
+ self.assertTrue(dm.pid_file_path.exists())
+
+ @mock.patch('os.execv')
+ def test_reboot_success(self, mock_execv):
+ binary_file = tempfile.NamedTemporaryFile(
+ dir=self.working_dir.name, delete=False
+ )
+
+ dm = daemon_manager.DaemonManager(
+ binary_file.name, daemon_target=long_running_daemon
+ )
+ dm.start()
+ dm.reboot()
+
+ # Verifies the old process is stopped
+ self.assert_no_subprocess_running()
+ self.assertFalse(dm.pid_file_path.exists())
+
+ mock_execv.assert_called_once()
+
+ @mock.patch('os.execv')
+ def test_reboot_binary_no_longer_exists(self, mock_execv):
+ dm = daemon_manager.DaemonManager(
+ TEST_BINARY_FILE, daemon_target=long_running_daemon
+ )
+ dm.start()
+
+ with self.assertRaises(SystemExit) as cm:
+ dm.reboot()
+ mock_execv.assert_not_called()
+ self.assertEqual(cm.exception.code, 0)
+
+ @mock.patch('os.execv')
+ def test_reboot_failed(self, mock_execv):
+ mock_execv.side_effect = OSError('Unknown OSError')
+ binary_file = tempfile.NamedTemporaryFile(
+ dir=self.working_dir.name, delete=False
+ )
+
+ dm = daemon_manager.DaemonManager(
+ binary_file.name, daemon_target=long_running_daemon
+ )
+ dm.start()
+
+ with self.assertRaises(SystemExit) as cm:
+ dm.reboot()
+ self.assertEqual(cm.exception.code, 1)
+
+ def assert_run_simple_daemon_success(self):
+ damone_output_file = tempfile.NamedTemporaryFile(
+ dir=self.working_dir.name, delete=False
+ )
+ dm = daemon_manager.DaemonManager(
+ TEST_BINARY_FILE,
+ daemon_target=simple_daemon,
+ daemon_args=(damone_output_file.name,),
+ )
+ dm.start()
+ dm.monitor_daemon(interval=1)
+
+ # Verifies the expected pid file is created.
+ expected_pid_file_path = pathlib.Path(self.working_dir.name).joinpath(
+ 'edit_monitor', TEST_PID_FILE_PATH
+ )
+ self.assertTrue(expected_pid_file_path.exists())
+
+ # Verify the daemon process is executed successfully.
+ with open(damone_output_file.name, 'r') as f:
+ contents = f.read()
+ self.assertEqual(contents, 'running daemon target')
+
+ def assert_no_subprocess_running(self):
+ child_pids = self._get_child_processes(os.getpid())
+ for child_pid in child_pids:
+ self.assertFalse(
+ self._is_process_alive(child_pid), f'process {child_pid} still alive'
+ )
+
+ def _get_child_processes(self, parent_pid):
+ try:
+ output = subprocess.check_output(
+ ['ps', '-o', 'pid,ppid', '--no-headers'], text=True
+ )
+
+ child_processes = []
+ for line in output.splitlines():
+ pid, ppid = line.split()
+ if int(ppid) == parent_pid:
+ child_processes.append(int(pid))
+ return child_processes
+ except subprocess.CalledProcessError as e:
+ self.fail(f'failed to get child process, error: {e}')
+
+ def _is_process_alive(self, pid):
+ try:
+ output = subprocess.check_output(
+ ['ps', '-p', str(pid), '-o', 'state='], text=True
+ ).strip()
+ state = output.split()[0]
+ return state != 'Z' # Check if the state is not 'Z' (zombie)
+ except subprocess.CalledProcessError:
+ return False
+
+ def _cleanup_child_processes(self):
+ child_pids = self._get_child_processes(os.getpid())
+ for child_pid in child_pids:
+ try:
+ os.kill(child_pid, signal.SIGKILL)
+ except ProcessLookupError:
+ # process already terminated
+ pass
+
+
+if __name__ == '__main__':
+ unittest.main()
diff --git a/tools/filelistdiff/Android.bp b/tools/filelistdiff/Android.bp
index 632ada3..ab766d6 100644
--- a/tools/filelistdiff/Android.bp
+++ b/tools/filelistdiff/Android.bp
@@ -24,4 +24,4 @@
prebuilt_etc_host {
name: "system_image_diff_allowlist",
src: "allowlist",
-}
+}
\ No newline at end of file
diff --git a/tools/filelistdiff/OWNERS b/tools/filelistdiff/OWNERS
new file mode 100644
index 0000000..690fb17
--- /dev/null
+++ b/tools/filelistdiff/OWNERS
@@ -0,0 +1 @@
+per-file allowlist = justinyun@google.com, jeongik@google.com, kiyoungkim@google.com, inseob@google.com
diff --git a/tools/filelistdiff/allowlist b/tools/filelistdiff/allowlist
index 943f955..ae8a662 100644
--- a/tools/filelistdiff/allowlist
+++ b/tools/filelistdiff/allowlist
@@ -1,87 +1,11 @@
# Known diffs only in the KATI system image
-etc/NOTICE.xml.gz
-etc/compatconfig/TeleService-platform-compat-config.xml
-etc/compatconfig/calendar-provider-compat-config.xml
-etc/compatconfig/contacts-provider-platform-compat-config.xml
-etc/compatconfig/documents-ui-compat-config.xml
-etc/compatconfig/framework-location-compat-config.xml
-etc/compatconfig/framework-platform-compat-config.xml
-etc/compatconfig/icu4j-platform-compat-config.xml
-etc/compatconfig/services-platform-compat-config.xml
-etc/permissions/android.software.credentials.xml
-etc/permissions/android.software.preview_sdk.xml
-etc/permissions/android.software.webview.xml
-etc/permissions/android.software.window_magnification.xml
-etc/permissions/com.android.adservices.sdksandbox.xml
-etc/security/otacerts.zip
-etc/vintf/compatibility_matrix.202404.xml
-etc/vintf/compatibility_matrix.202504.xml
-etc/vintf/compatibility_matrix.5.xml
-etc/vintf/compatibility_matrix.6.xml
-etc/vintf/compatibility_matrix.7.xml
-etc/vintf/compatibility_matrix.8.xml
-etc/vintf/compatibility_matrix.device.xml
-etc/vintf/manifest.xml
-framework/boot-apache-xml.vdex
-framework/boot-apache-xml.vdex.fsv_meta
-framework/boot-bouncycastle.vdex
-framework/boot-bouncycastle.vdex.fsv_meta
-framework/boot-core-icu4j.vdex
-framework/boot-core-icu4j.vdex.fsv_meta
-framework/boot-core-libart.vdex
-framework/boot-core-libart.vdex.fsv_meta
-framework/boot-ext.vdex
-framework/boot-ext.vdex.fsv_meta
-framework/boot-framework-adservices.vdex
-framework/boot-framework-adservices.vdex.fsv_meta
-framework/boot-framework-graphics.vdex
-framework/boot-framework-graphics.vdex.fsv_meta
-framework/boot-framework-location.vdex
-framework/boot-framework-location.vdex.fsv_meta
-framework/boot-framework.vdex
-framework/boot-framework.vdex.fsv_meta
-framework/boot-ims-common.vdex
-framework/boot-ims-common.vdex.fsv_meta
-framework/boot-okhttp.vdex
-framework/boot-okhttp.vdex.fsv_meta
-framework/boot-telephony-common.vdex
-framework/boot-telephony-common.vdex.fsv_meta
-framework/boot-voip-common.vdex
-framework/boot-voip-common.vdex.fsv_meta
-framework/boot.vdex
-framework/boot.vdex.fsv_meta
framework/oat/x86_64/apex@com.android.compos@javalib@service-compos.jar@classes.odex
framework/oat/x86_64/apex@com.android.compos@javalib@service-compos.jar@classes.odex.fsv_meta
framework/oat/x86_64/apex@com.android.compos@javalib@service-compos.jar@classes.vdex
framework/oat/x86_64/apex@com.android.compos@javalib@service-compos.jar@classes.vdex.fsv_meta
-lib/aaudio-aidl-cpp.so
-lib/android.hardware.biometrics.fingerprint@2.1.so
-lib/android.hardware.radio.config@1.0.so
-lib/android.hardware.radio.deprecated@1.0.so
-lib/android.hardware.radio@1.0.so
-lib/android.hardware.radio@1.1.so
-lib/android.hardware.radio@1.2.so
-lib/android.hardware.radio@1.3.so
-lib/android.hardware.radio@1.4.so
-lib/android.hardware.secure_element@1.0.so
-lib/com.android.media.aaudio-aconfig-cc.so
-lib/heapprofd_client.so
-lib/heapprofd_client_api.so
-lib/libaaudio.so
-lib/libaaudio_internal.so
-lib/libalarm_jni.so
-lib/libamidi.so
-lib/libcups.so
-lib/libjni_deviceAsWebcam.so
-lib/libprintspooler_jni.so
-lib/libvendorsupport.so
-lib/libwfds.so
-lib/libyuv.so
-# b/351258461
-adb_keys
+# Known diffs that are installed in either system image with the configuration
+# b/353429422
init.environ.rc
-
-# Known diffs only in the Soong system image
-lib/libhidcommand_jni.so
-lib/libuinputcommand_jni.so
\ No newline at end of file
+# b/338342381
+etc/NOTICE.xml.gz
diff --git a/tools/filelistdiff/file_list_diff.py b/tools/filelistdiff/file_list_diff.py
index cdc5b2e..30ed107 100644
--- a/tools/filelistdiff/file_list_diff.py
+++ b/tools/filelistdiff/file_list_diff.py
@@ -34,23 +34,24 @@
unique_in_soong = set(filter(is_unknown_diff, soong_files - kati_files))
if unique_in_kati:
- print(f'{COLOR_ERROR}Please add following modules into system image module {system_module_name}.{COLOR_NORMAL}')
- print(f'{COLOR_WARNING}KATI only module(s):{COLOR_NORMAL}')
+ print('')
+ print(f'{COLOR_ERROR}Missing required modules in {system_module_name} module.{COLOR_NORMAL}')
+ print(f'To resolve this issue, please add the modules to the Android.bp file for the {system_module_name} to install the following KATI only installed files.')
+ print(f'You can find the correct Android.bp file using the command "gomod {system_module_name}".')
+ print(f'{COLOR_WARNING}KATI only installed file(s):{COLOR_NORMAL}')
for item in sorted(unique_in_kati):
- print(item)
+ print(' '+item)
if unique_in_soong:
- if unique_in_kati:
- print('')
-
- print(f'{COLOR_ERROR}Please add following modules into build/make/target/product/base_system.mk.{COLOR_NORMAL}')
- print(f'{COLOR_WARNING}Soong only module(s):{COLOR_NORMAL}')
+ print('')
+ print(f'{COLOR_ERROR}Missing packages in base_system.mk.{COLOR_NORMAL}')
+ print('Please add packages into build/make/target/product/base_system.mk or build/make/tools/filelistdiff/allowlist to install or skip the following Soong only installed files.')
+ print(f'{COLOR_WARNING}Soong only installed file(s):{COLOR_NORMAL}')
for item in sorted(unique_in_soong):
- print(item)
+ print(' '+item)
if unique_in_kati or unique_in_soong:
print('')
- print(f'{COLOR_ERROR}FAILED: System image from KATI and SOONG differs from installed file list.{COLOR_NORMAL}')
sys.exit(1)
diff --git a/tools/finalization/README.md b/tools/finalization/README.md
index d0aed69..5e2aecd 100644
--- a/tools/finalization/README.md
+++ b/tools/finalization/README.md
@@ -19,3 +19,8 @@
## Utility:
[Full cleanup](./cleanup.sh). Remove all local changes and switch each project into head-less state. This is the best state to sync/rebase/finalize the branch.
+
+## Dry run:
+[Full cleanup](./dryrun-cleanup.sh). Remove all local changes and switch each project into head-less state. Also removes "DryRun" branches.
+[SDK](./dryrun-step-1.sh). Perform SDK finalization and upload the CLs to Gerrit.
+[SDK and REL](./dryrun-step-1-and-2.sh). Perform SDK finalization, plus all necessary changes to switch configuration to REL, and upload the CLs to Gerrit.
\ No newline at end of file
diff --git a/tools/finalization/build-step-1-and-m.sh b/tools/finalization/build-step-1-and-m.sh
index 0e7129f..88bb347 100755
--- a/tools/finalization/build-step-1-and-m.sh
+++ b/tools/finalization/build-step-1-and-m.sh
@@ -9,10 +9,9 @@
local m="$top/build/soong/soong_ui.bash --make-mode TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
# This command tests:
- # The release state for AIDL.
# ABI difference between user and userdebug builds.
# Resource/SDK finalization.
- AIDL_FROZEN_REL=true $m
+ $m
}
finalize_main_step1_and_m
diff --git a/tools/finalization/cleanup.sh b/tools/finalization/cleanup.sh
index cd87b1d..e2a0592 100755
--- a/tools/finalization/cleanup.sh
+++ b/tools/finalization/cleanup.sh
@@ -14,8 +14,8 @@
repo forall -c '\
git checkout . ; git revert --abort ; git clean -fdx ;\
- git checkout @ ; git branch fina-step1 -D ; git reset --hard; \
- repo start fina-step1 ; git checkout @ ; git b fina-step1 -D ;'
+ git checkout @ --detach ; git branch fina-step1 -D ; git reset --hard; \
+ repo start fina-step1 ; git checkout @ --detach ; git b fina-step1 -D ;'
}
finalize_revert_local_changes_main
diff --git a/tools/finalization/command-line-options.sh b/tools/finalization/command-line-options.sh
new file mode 100644
index 0000000..d9397c2
--- /dev/null
+++ b/tools/finalization/command-line-options.sh
@@ -0,0 +1,8 @@
+ARGV=$(getopt --options '' --long dry-run -- "$@")
+eval set -- "$ARGV"
+while true; do
+ case "$1" in
+ --dry-run) repo_upload_dry_run_arg="--dry-run"; repo_branch="finalization-dry-run"; shift ;;
+ *) break
+ esac
+done
diff --git a/tools/finalization/dryrun-cleanup.sh b/tools/finalization/dryrun-cleanup.sh
new file mode 100755
index 0000000..ddaffae
--- /dev/null
+++ b/tools/finalization/dryrun-cleanup.sh
@@ -0,0 +1,25 @@
+#!/bin/bash
+# Brings local repository to a remote head state. Also removes all dryrun branches.
+
+# set -ex
+
+function finalize_revert_local_changes_main() {
+ local top="$(dirname "$0")"/../../../..
+ source $top/build/make/tools/finalization/environment.sh
+
+ local m="$top/build/soong/soong_ui.bash --make-mode TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+
+ # remove the out folder
+ $m clobber
+
+ repo selfupdate
+
+ repo forall -c '\
+ git checkout . ; git revert --abort ; git clean -fdx ;\
+ git checkout @ --detach ; git branch fina-step1 -D ; git reset --hard; \
+ repo start fina-step1 ; git checkout @ --detach ; git b fina-step1 -D ; \
+ git b $FINAL_PLATFORM_CODENAME-SDK-Finalization-DryRun -D; \
+ git b $FINAL_PLATFORM_CODENAME-SDK-Finalization-DryRun-Rel -D; '
+}
+
+finalize_revert_local_changes_main
diff --git a/tools/finalization/dryrun-step-1-and-2.sh b/tools/finalization/dryrun-step-1-and-2.sh
new file mode 100755
index 0000000..f883bca
--- /dev/null
+++ b/tools/finalization/dryrun-step-1-and-2.sh
@@ -0,0 +1,39 @@
+#!/bin/bash
+# Script to perform 1st and 2nd step of Android Finalization, create CLs and upload to Gerrit.
+
+function commit_step_2_changes() {
+ repo forall -c '\
+ if [[ $(git status --short) ]]; then
+ repo start "$FINAL_PLATFORM_CODENAME-SDK-Finalization-DryRun-Rel" ;
+ git add -A . ;
+ git commit -m "$FINAL_PLATFORM_CODENAME/$FINAL_PLATFORM_SDK_VERSION is now REL" \
+ -m "Ignore-AOSP-First: $FINAL_PLATFORM_CODENAME Finalization
+Bug: $FINAL_BUG_ID
+Test: build";
+
+ repo upload --cbr --no-verify -o nokeycheck -t -y . ;
+ fi'
+}
+
+function finalize_step_2_main() {
+ local top="$(dirname "$0")"/../../../..
+ source $top/build/make/tools/finalization/environment.sh
+
+ source $top/build/make/tools/finalization/finalize-sdk-resources.sh
+
+ source $top/build/make/tools/finalization/localonly-steps.sh
+
+ source $top/build/make/tools/finalization/finalize-sdk-rel.sh
+
+ # move all changes to finalization branch/topic and upload to gerrit
+ commit_step_2_changes
+
+ # build to confirm everything is OK
+ local m_next="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=next TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ $m_next
+
+ local m_fina="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=fina_2 TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ $m_fina
+}
+
+finalize_step_2_main
diff --git a/tools/finalization/dryrun-step-1.sh b/tools/finalization/dryrun-step-1.sh
new file mode 100755
index 0000000..0f2bc63
--- /dev/null
+++ b/tools/finalization/dryrun-step-1.sh
@@ -0,0 +1,35 @@
+#!/bin/bash
+# Script to perform a dry run of step 1 of Android Finalization, create CLs and upload to Gerrit.
+
+function commit_step_1_changes() {
+ repo forall -c '\
+ if [[ $(git status --short) ]]; then
+ repo start "$FINAL_PLATFORM_CODENAME-SDK-Finalization-DryRun" ;
+ git add -A . ;
+ git commit -m "$FINAL_PLATFORM_CODENAME is now $FINAL_PLATFORM_SDK_VERSION" \
+ -m "Ignore-AOSP-First: $FINAL_PLATFORM_CODENAME Finalization
+Bug: $FINAL_BUG_ID
+Test: build";
+
+ repo upload --cbr --no-verify -o nokeycheck -t -y . ;
+ fi'
+}
+
+function finalize_step_1_main() {
+ local top="$(dirname "$0")"/../../../..
+ source $top/build/make/tools/finalization/environment.sh
+
+ source $top/build/make/tools/finalization/finalize-sdk-resources.sh
+
+ # move all changes to finalization branch/topic and upload to gerrit
+ commit_step_1_changes
+
+ # build to confirm everything is OK
+ local m_next="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=next TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ $m_next
+
+ local m_fina="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=fina_1 TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ $m_fina
+}
+
+finalize_step_1_main
diff --git a/tools/finalization/environment.sh b/tools/finalization/environment.sh
index 7961e8b..f68a68b 100755
--- a/tools/finalization/environment.sh
+++ b/tools/finalization/environment.sh
@@ -15,14 +15,14 @@
# We might or might not fix this in future, but for now let's keep it +1.
export FINAL_PLATFORM_SDK_VERSION='35'
# Feel free to randomize once in a while to detect buggy version detection code.
-export FINAL_MAINLINE_EXTENSION='58'
+export FINAL_MAINLINE_EXTENSION='13'
# Options:
# 'unfinalized' - branch is in development state,
# 'vintf' - VINTF is finalized
# 'sdk' - VINTF and SDK/API are finalized
# 'rel' - branch is finalized, switched to REL
-export FINAL_STATE='vintf'
+export FINAL_STATE='rel'
export BUILD_FROM_SOURCE_STUB=true
diff --git a/tools/finalization/finalize-sdk-rel.sh b/tools/finalization/finalize-sdk-rel.sh
index 59fe28c..c49f974 100755
--- a/tools/finalization/finalize-sdk-rel.sh
+++ b/tools/finalization/finalize-sdk-rel.sh
@@ -8,12 +8,6 @@
fi
}
-function revert_resources_sdk_int_fix() {
- if grep -q 'public static final int RESOURCES_SDK_INT = SDK_INT;' "$top/frameworks/base/core/java/android/os/Build.java" ; then
- patch --strip=1 --no-backup-if-mismatch --directory="$top/frameworks/base" --input=../../build/make/tools/finalization/frameworks_base.revert_resource_sdk_int.diff
- fi
-}
-
function apply_prerelease_sdk_hack() {
if ! grep -q 'STOPSHIP: hack for the pre-release SDK' "$top/frameworks/base/core/java/android/content/pm/parsing/FrameworkParsingPackageUtils.java" ; then
patch --strip=1 --no-backup-if-mismatch --directory="$top/frameworks/base" --input=../../build/make/tools/finalization/frameworks_base.apply_hack.diff
@@ -30,25 +24,18 @@
# let the apps built with pre-release SDK parse
apply_prerelease_sdk_hack
- # in REL mode, resources would correctly set the resources_sdk_int, no fix required
- revert_resources_sdk_int_fix
-
# cts
- echo "$FINAL_PLATFORM_VERSION" > "$top/cts/tests/tests/os/assets/platform_versions.txt"
+ if ! grep -q "${FINAL_PLATFORM_VERSION}" "$top/cts/tests/tests/os/assets/platform_versions.txt" ; then
+ echo ${FINAL_PLATFORM_VERSION} >> "$top/cts/tests/tests/os/assets/platform_versions.txt"
+ fi
if [ "$FINAL_PLATFORM_CODENAME" != "$CURRENT_PLATFORM_CODENAME" ]; then
echo "$CURRENT_PLATFORM_CODENAME" >> "./cts/tests/tests/os/assets/platform_versions.txt"
fi
git -C "$top/cts" mv hostsidetests/theme/assets/${FINAL_PLATFORM_CODENAME} hostsidetests/theme/assets/${FINAL_PLATFORM_SDK_VERSION}
# prebuilts/abi-dumps/platform
- mkdir -p "$top/prebuilts/abi-dumps/platform/$FINAL_PLATFORM_SDK_VERSION"
- cp -r "$top/prebuilts/abi-dumps/platform/current/64/" "$top/prebuilts/abi-dumps/platform/$FINAL_PLATFORM_SDK_VERSION/"
-
- # TODO(b/309880485)
- # uncomment and update
- # prebuilts/abi-dumps/ndk
- #mkdir -p "$top/prebuilts/abi-dumps/ndk/$FINAL_PLATFORM_SDK_VERSION"
- #cp -r "$top/prebuilts/abi-dumps/ndk/current/64/" "$top/prebuilts/abi-dumps/ndk/$FINAL_PLATFORM_SDK_VERSION/"
+ "$top/build/soong/soong_ui.bash" --make-mode TARGET_RELEASE=next TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug create_reference_dumps
+ ANDROID_BUILD_TOP="$top" "$top/out/host/linux-x86/bin/create_reference_dumps" -release next --build-variant userdebug --lib-variant APEX
}
finalize_sdk_rel
diff --git a/tools/finalization/finalize-sdk-resources.sh b/tools/finalization/finalize-sdk-resources.sh
index 596f803..10266ed 100755
--- a/tools/finalization/finalize-sdk-resources.sh
+++ b/tools/finalization/finalize-sdk-resources.sh
@@ -9,13 +9,6 @@
fi
}
-function apply_resources_sdk_int_fix() {
- if ! grep -q 'public static final int RESOURCES_SDK_INT = SDK_INT;' "$top/frameworks/base/core/java/android/os/Build.java" ; then
- local base_git_root="$(readlink -f $top/frameworks/base)"
- patch --strip=1 --no-backup-if-mismatch --directory="$base_git_root" --input=../../build/make/tools/finalization/frameworks_base.apply_resource_sdk_int.diff
- fi
-}
-
function finalize_bionic_ndk() {
# Adding __ANDROID_API_<>__.
# If this hasn't done then it's not used and not really needed. Still, let's check and add this.
@@ -41,7 +34,8 @@
echo " /** Checks if the device is running on a release version of Android $FINAL_PLATFORM_CODENAME or newer */
@ChecksSdkIntAtLeast(api = $FINAL_PLATFORM_SDK_VERSION /* BUILD_VERSION_CODES.$FINAL_PLATFORM_CODENAME */)
public static boolean isAtLeast${FINAL_PLATFORM_CODENAME:0:1}() {
- return SDK_INT >= $FINAL_PLATFORM_SDK_VERSION;
+ return SDK_INT >= $FINAL_PLATFORM_SDK_VERSION ||
+ (SDK_INT == $(($FINAL_PLATFORM_SDK_VERSION - 1)) && isAtLeastPreReleaseCodename(\"$FINAL_PLATFORM_CODENAME\"));
}" > "$tmpfile"
local javaFuncRegex='\/\*\*[^{]*isAtLeast'"${shortCodename}"'() {[^{}]*}'
@@ -55,7 +49,11 @@
d}' $javaSdkLevel
echo "// Checks if the device is running on release version of Android ${FINAL_PLATFORM_CODENAME:0:1} or newer.
-inline bool IsAtLeast${FINAL_PLATFORM_CODENAME:0:1}() { return android_get_device_api_level() >= $FINAL_PLATFORM_SDK_VERSION; }" > "$tmpfile"
+inline bool IsAtLeast${FINAL_PLATFORM_CODENAME:0:1}() {
+ return android_get_device_api_level() >= $FINAL_PLATFORM_SDK_VERSION ||
+ (android_get_device_api_level() == $(($FINAL_PLATFORM_SDK_VERSION - 1)) &&
+ detail::IsAtLeastPreReleaseCodename(\"$FINAL_PLATFORM_CODENAME\"));
+}" > "$tmpfile"
local cppFuncRegex='\/\/[^{]*IsAtLeast'"${shortCodename}"'() {[^{}]*}'
local cppFuncReplace="N;N;N;N;N;N; s/$cppFuncRegex/$methodPlaceholder/; /$cppFuncRegex/!{P;D};"
@@ -123,14 +121,18 @@
sed -i -e 's/Pkg\.Revision.*/Pkg\.Revision=${PLATFORM_SDK_VERSION}.0.0/g' $build_tools_source
# build/soong
- local codename_version="\"${FINAL_PLATFORM_CODENAME}\": ${FINAL_PLATFORM_SDK_VERSION}"
+ local codename_version="\"${FINAL_PLATFORM_CODENAME}\": ${FINAL_PLATFORM_SDK_VERSION}"
if ! grep -q "$codename_version" "$top/build/soong/android/api_levels.go" ; then
sed -i -e "/:.*$((${FINAL_PLATFORM_SDK_VERSION}-1)),/a \\\t\t$codename_version," "$top/build/soong/android/api_levels.go"
fi
# cts
- echo ${FINAL_PLATFORM_VERSION} > "$top/cts/tests/tests/os/assets/platform_releases.txt"
- sed -i -e "s/EXPECTED_SDK = $((${FINAL_PLATFORM_SDK_VERSION}-1))/EXPECTED_SDK = ${FINAL_PLATFORM_SDK_VERSION}/g" "$top/cts/tests/tests/os/src/android/os/cts/BuildVersionTest.java"
+ if ! grep -q "${FINAL_PLATFORM_VERSION}" "$top/cts/tests/tests/os/assets/platform_releases.txt" ; then
+ echo ${FINAL_PLATFORM_VERSION} >> "$top/cts/tests/tests/os/assets/platform_releases.txt"
+ fi
+ if ! grep -q "$((${FINAL_PLATFORM_SDK_VERSION}-1)), ${FINAL_PLATFORM_VERSION}" "$top/cts/tests/tests/os/src/android/os/cts/BuildVersionTest.java" ; then
+ sed -i -e "s/.*EXPECTED_SDKS = List.of(.*$((${FINAL_PLATFORM_SDK_VERSION}-1))/&, $FINAL_PLATFORM_SDK_VERSION/" "$top/cts/tests/tests/os/src/android/os/cts/BuildVersionTest.java"
+ fi
# libcore
sed -i "s%$SDK_CODENAME%$SDK_VERSION%g" "$top/libcore/dalvik/src/main/java/dalvik/annotation/compat/VersionCodes.java"
@@ -153,7 +155,6 @@
# frameworks/base
sed -i "s%$SDK_CODENAME%$SDK_VERSION%g" "$top/frameworks/base/core/java/android/os/Build.java"
- apply_resources_sdk_int_fix
sed -i -e "/=.*$((${FINAL_PLATFORM_SDK_VERSION}-1)),/a \\ SDK_${FINAL_PLATFORM_CODENAME_JAVA} = ${FINAL_PLATFORM_SDK_VERSION}," "$top/frameworks/base/tools/aapt/SdkConstants.h"
sed -i -e "/=.*$((${FINAL_PLATFORM_SDK_VERSION}-1)),/a \\ SDK_${FINAL_PLATFORM_CODENAME_JAVA} = ${FINAL_PLATFORM_SDK_VERSION}," "$top/frameworks/base/tools/aapt2/SdkConstants.h"
diff --git a/tools/finalization/finalize-vintf-resources.sh b/tools/finalization/finalize-vintf-resources.sh
index a55d8e1..d532b25 100755
--- a/tools/finalization/finalize-vintf-resources.sh
+++ b/tools/finalization/finalize-vintf-resources.sh
@@ -16,8 +16,6 @@
export TARGET_RELEASE=fina_0
export TARGET_PRODUCT=aosp_arm64
- # TODO(b/314010764): finalize LL_NDK
-
# system/sepolicy
"$top/system/sepolicy/tools/finalize-vintf-resources.sh" "$top" "$FINAL_BOARD_API_LEVEL"
@@ -25,7 +23,11 @@
# pre-finalization build target (trunk)
local aidl_m="$top/build/soong/soong_ui.bash --make-mode"
- AIDL_TRANSITIVE_FREEZE=true $aidl_m aidl-freeze-api
+ AIDL_TRANSITIVE_FREEZE=true $aidl_m aidl-freeze-api create_reference_dumps
+
+ # Generate LLNDK ABI dumps
+ # This command depends on ANDROID_BUILD_TOP
+ "$ANDROID_HOST_OUT/bin/create_reference_dumps" -release "$TARGET_RELEASE" --build-variant "$TARGET_BUILD_VARIANT" --lib-variant LLNDK
}
function create_new_compat_matrix_and_kernel_configs() {
diff --git a/tools/finalization/frameworks_base.apply_resource_sdk_int.diff b/tools/finalization/frameworks_base.apply_resource_sdk_int.diff
deleted file mode 100644
index f0576d0..0000000
--- a/tools/finalization/frameworks_base.apply_resource_sdk_int.diff
+++ /dev/null
@@ -1,24 +0,0 @@
-From cdb47fc90b8d6860ec1dc5efada1f9ccd471618b Mon Sep 17 00:00:00 2001
-From: Alex Buynytskyy <alexbuy@google.com>
-Date: Tue, 11 Apr 2023 22:12:44 +0000
-Subject: [PATCH] Don't force +1 for resource resolution.
-
-Bug: 277674088
-Fixes: 277674088
-Test: boots, no crashes
-Change-Id: I17e743a0f1cf6f98fddd40c358dea5a8b9cc7723
----
-
-diff --git a/core/java/android/os/Build.java b/core/java/android/os/Build.java
-index eb47170..4d3e92b 100755
---- a/core/java/android/os/Build.java
-+++ b/core/java/android/os/Build.java
-@@ -493,7 +493,7 @@
- * @hide
- */
- @TestApi
-- public static final int RESOURCES_SDK_INT = SDK_INT + ACTIVE_CODENAMES.length;
-+ public static final int RESOURCES_SDK_INT = SDK_INT;
-
- /**
- * The current lowest supported value of app target SDK. Applications targeting
diff --git a/tools/finalization/frameworks_base.revert_resource_sdk_int.diff b/tools/finalization/frameworks_base.revert_resource_sdk_int.diff
deleted file mode 100644
index 2ade499..0000000
--- a/tools/finalization/frameworks_base.revert_resource_sdk_int.diff
+++ /dev/null
@@ -1,27 +0,0 @@
-From c7e460bb19071d867cd7ca04282ce42694f4f358 Mon Sep 17 00:00:00 2001
-From: Alex Buynytskyy <alexbuy@google.com>
-Date: Wed, 12 Apr 2023 01:06:26 +0000
-Subject: [PATCH] Revert "Don't force +1 for resource resolution."
-
-It's not required for master.
-
-This reverts commit f1cb683988f81579a76ddbf9993848a4a06dd28c.
-
-Bug: 277674088
-Test: boots, no crashes
-Change-Id: Ia1692548f26496fdc6f1e4f0557213c7996d6823
----
-
-diff --git a/core/java/android/os/Build.java b/core/java/android/os/Build.java
-index 4d3e92b..eb47170 100755
---- a/core/java/android/os/Build.java
-+++ b/core/java/android/os/Build.java
-@@ -493,7 +493,7 @@
- * @hide
- */
- @TestApi
-- public static final int RESOURCES_SDK_INT = SDK_INT;
-+ public static final int RESOURCES_SDK_INT = SDK_INT + ACTIVE_CODENAMES.length;
-
- /**
- * The current lowest supported value of app target SDK. Applications targeting
diff --git a/tools/finalization/localonly-steps.sh b/tools/finalization/localonly-steps.sh
index bebd563..94ee368 100755
--- a/tools/finalization/localonly-steps.sh
+++ b/tools/finalization/localonly-steps.sh
@@ -10,14 +10,15 @@
local m="$top/build/soong/soong_ui.bash --make-mode TARGET_PRODUCT=aosp_arm64 TARGET_RELEASE=fina_1 TARGET_BUILD_VARIANT=userdebug DIST_DIR=out/dist"
# adb keys
- $m adb
- LOGNAME=android-eng HOSTNAME=google.com "$top/out/host/linux-x86/bin/adb" keygen "$top/vendor/google/security/adb/${FINAL_PLATFORM_VERSION}.adb_key"
+ # The keys are already generated for Android 15. Keeping the command (commented out) for future reference.
+ # $m adb
+ # LOGNAME=android-eng HOSTNAME=google.com "$top/out/host/linux-x86/bin/adb" keygen "$top/vendor/google/security/adb/${FINAL_PLATFORM_VERSION}.adb_key"
# Build Platform SDKs.
$top/build/soong/soong_ui.bash --make-mode TARGET_PRODUCT=sdk TARGET_RELEASE=fina_1 TARGET_BUILD_VARIANT=userdebug sdk dist sdk_repo DIST_DIR=out/dist
# Build Modules SDKs.
- TARGET_RELEASE=fina_1 TARGET_BUILD_VARIANT=userdebug UNBUNDLED_BUILD_SDKS_FROM_SOURCE=true DIST_DIR=out/dist "$top/vendor/google/build/mainline_modules_sdks.sh" --build-release=latest
+ TARGET_RELEASE=fina_1 TARGET_BUILD_VARIANT=userdebug UNBUNDLED_BUILD_SDKS_FROM_SOURCE=true DIST_DIR=out/dist "$top/vendor/google/build/mainline_modules_sdks.sh" --build-release=next
# Update prebuilts.
"$top/prebuilts/build-tools/path/linux-x86/python3" -W ignore::DeprecationWarning "$top/prebuilts/sdk/update_prebuilts.py" --local_mode -f ${FINAL_PLATFORM_SDK_VERSION} -e ${FINAL_MAINLINE_EXTENSION} --bug 1 1
diff --git a/tools/finalization/step-0.sh b/tools/finalization/step-0.sh
index e61c644..2087f6e 100755
--- a/tools/finalization/step-0.sh
+++ b/tools/finalization/step-0.sh
@@ -9,19 +9,21 @@
set +e
repo forall -c '\
if [[ $(git status --short) ]]; then
- repo start "VINTF-$FINAL_BOARD_API_LEVEL-Finalization" ;
+ repo start "'$repo_branch'" ;
git add -A . ;
git commit -m "Vendor API level $FINAL_BOARD_API_LEVEL is now frozen" \
-m "Ignore-AOSP-First: VINTF $FINAL_BOARD_API_LEVEL Finalization
Bug: $FINAL_BUG_ID
Test: build";
- repo upload --cbr --no-verify -o nokeycheck -t -y . ;
+ repo upload '"$repo_upload_dry_run_arg"' --cbr --no-verify -o nokeycheck -t -y . ;
fi'
}
function finalize_step_0_main() {
local top="$(dirname "$0")"/../../../..
source $top/build/make/tools/finalization/environment.sh
+ local repo_branch="VINTF-$FINAL_BOARD_API_LEVEL-Finalization"
+ source $top/build/make/tools/finalization/command-line-options.sh
local m="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=next TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
@@ -34,4 +36,4 @@
AIDL_FROZEN_REL=true $m
}
-finalize_step_0_main
+finalize_step_0_main $@
diff --git a/tools/finalization/step-1.sh b/tools/finalization/step-1.sh
index 0e483d5..736d641 100755
--- a/tools/finalization/step-1.sh
+++ b/tools/finalization/step-1.sh
@@ -7,21 +7,21 @@
set +e
repo forall -c '\
if [[ $(git status --short) ]]; then
- repo start "$FINAL_PLATFORM_CODENAME-SDK-Finalization" ;
+ repo start "'$repo_branch'" ;
git add -A . ;
git commit -m "$FINAL_PLATFORM_CODENAME is now $FINAL_PLATFORM_SDK_VERSION and extension version $FINAL_MAINLINE_EXTENSION" \
-m "Ignore-AOSP-First: $FINAL_PLATFORM_CODENAME Finalization
Bug: $FINAL_BUG_ID
Test: build";
- repo upload --cbr --no-verify -o nokeycheck -t -y . ;
+ repo upload '"$repo_upload_dry_run_arg"' --cbr --no-verify -o nokeycheck -t -y . ;
fi'
}
function finalize_step_1_main() {
local top="$(dirname "$0")"/../../../..
source $top/build/make/tools/finalization/environment.sh
-
- local m="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=next TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ local repo_branch="$FINAL_PLATFORM_CODENAME-SDK-Finalization"
+ source $top/build/make/tools/finalization/command-line-options.sh
source $top/build/make/tools/finalization/finalize-sdk-resources.sh
@@ -29,7 +29,11 @@
commit_step_1_changes
# build to confirm everything is OK
- AIDL_FROZEN_REL=true $m
+ local m_next="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=next TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ $m_next
+
+ local m_fina="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=fina_1 TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ $m_fina
}
-finalize_step_1_main
+finalize_step_1_main $@
diff --git a/tools/finalization/step-2.sh b/tools/finalization/step-2.sh
index 356cad0..52e3887 100755
--- a/tools/finalization/step-2.sh
+++ b/tools/finalization/step-2.sh
@@ -4,22 +4,22 @@
function commit_step_2_changes() {
repo forall -c '\
if [[ $(git status --short) ]]; then
- repo start "$FINAL_PLATFORM_CODENAME-SDK-Finalization-Rel" ;
+ repo start "'$repo_branch'" ;
git add -A . ;
git commit -m "$FINAL_PLATFORM_CODENAME/$FINAL_PLATFORM_SDK_VERSION is now REL" \
-m "Ignore-AOSP-First: $FINAL_PLATFORM_CODENAME Finalization
Bug: $FINAL_BUG_ID
Test: build";
- repo upload --cbr --no-verify -o nokeycheck -t -y . ;
+ repo upload '"$repo_upload_dry_run_arg"' --cbr --no-verify -o nokeycheck -t -y . ;
fi'
}
function finalize_step_2_main() {
local top="$(dirname "$0")"/../../../..
source $top/build/make/tools/finalization/environment.sh
-
- local m="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=next TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ local repo_branch="$FINAL_PLATFORM_CODENAME-SDK-Finalization-Rel"
+ source $top/build/make/tools/finalization/command-line-options.sh
# prebuilts etc
source $top/build/make/tools/finalization/finalize-sdk-rel.sh
@@ -28,7 +28,11 @@
commit_step_2_changes
# build to confirm everything is OK
- AIDL_FROZEN_REL=true $m
+ local m_next="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=next TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ $m_next
+
+ local m_fina="$top/build/soong/soong_ui.bash --make-mode TARGET_RELEASE=fina_2 TARGET_PRODUCT=aosp_arm64 TARGET_BUILD_VARIANT=userdebug"
+ $m_fina
}
-finalize_step_2_main
+finalize_step_2_main $@
diff --git a/tools/ide_query/ide_query.sh b/tools/ide_query/ide_query.sh
index 6f9b0c4..8dfffc1 100755
--- a/tools/ide_query/ide_query.sh
+++ b/tools/ide_query/ide_query.sh
@@ -19,7 +19,7 @@
require_top
# Ensure cogsetup (out/ will be symlink outside the repo)
-. ${TOP}/build/make/cogsetup.sh
+setup_cog_env_if_needed
case $(uname -s) in
Linux)
diff --git a/tools/releasetools/Android.bp b/tools/releasetools/Android.bp
index cf7e2ae..8c71044 100644
--- a/tools/releasetools/Android.bp
+++ b/tools/releasetools/Android.bp
@@ -368,6 +368,9 @@
libs: [
"ota_utils_lib",
],
+ required: [
+ "signapk",
+ ],
}
python_binary_host {
@@ -555,6 +558,7 @@
"releasetools_common",
"ota_metadata_proto",
"ota_utils_lib",
+ "update_payload",
],
}
diff --git a/tools/releasetools/add_img_to_target_files.py b/tools/releasetools/add_img_to_target_files.py
index b39a82c..c25ff27 100644
--- a/tools/releasetools/add_img_to_target_files.py
+++ b/tools/releasetools/add_img_to_target_files.py
@@ -464,6 +464,7 @@
dtbo_prebuilt_path = os.path.join(
OPTIONS.input_tmp, "PREBUILT_IMAGES", "dtbo.img")
assert os.path.exists(dtbo_prebuilt_path)
+ os.makedirs(os.path.dirname(img.name), exist_ok=True)
shutil.copy(dtbo_prebuilt_path, img.name)
# AVB-sign the image as needed.
diff --git a/tools/releasetools/apex_utils.py b/tools/releasetools/apex_utils.py
index 73f36c7..54df955 100644
--- a/tools/releasetools/apex_utils.py
+++ b/tools/releasetools/apex_utils.py
@@ -200,7 +200,7 @@
# --signing_args "--signing_helper_with_files=%path" to apexer
if signing_args:
generate_image_cmd.extend(
- ['--signing_args', signing_args])
+ ['--signing_args', '"{}"'.format(signing_args)])
# optional arguments for apex repacking
manifest_json = os.path.join(apex_dir, 'apex_manifest.json')
diff --git a/tools/releasetools/common.py b/tools/releasetools/common.py
index edd4366..f04dfb7 100644
--- a/tools/releasetools/common.py
+++ b/tools/releasetools/common.py
@@ -3238,7 +3238,9 @@
return t
def WriteToDir(self, d):
- with open(os.path.join(d, self.name), "wb") as fp:
+ output_path = os.path.join(d, self.name)
+ os.makedirs(os.path.dirname(output_path), exist_ok=True)
+ with open(output_path, "wb") as fp:
fp.write(self.data)
def AddToZip(self, z, compression=None):
diff --git a/tools/releasetools/ota_from_raw_img.py b/tools/releasetools/ota_from_raw_img.py
index 03b44f1..3b9374a 100644
--- a/tools/releasetools/ota_from_raw_img.py
+++ b/tools/releasetools/ota_from_raw_img.py
@@ -105,9 +105,6 @@
if args.package_key:
logger.info("Signing payload...")
- # TODO: remove OPTIONS when no longer used as fallback in payload_signer
- common.OPTIONS.payload_signer_args = None
- common.OPTIONS.payload_signer_maximum_signature_size = None
signer = PayloadSigner(args.package_key, args.private_key_suffix,
key_passwords[args.package_key],
payload_signer=args.payload_signer,
diff --git a/tools/releasetools/ota_from_target_files.py b/tools/releasetools/ota_from_target_files.py
index 985cd56..6446e1f 100755
--- a/tools/releasetools/ota_from_target_files.py
+++ b/tools/releasetools/ota_from_target_files.py
@@ -264,6 +264,10 @@
--compression_factor
Specify the maximum block size to be compressed at once during OTA. supported options: 4k, 8k, 16k, 32k, 64k, 128k, 256k
+
+ --full_ota_partitions
+ Specify list of partitions should be updated in full OTA fashion, even if
+ an incremental OTA is about to be generated
"""
from __future__ import print_function
@@ -283,7 +287,7 @@
import ota_utils
import payload_signer
from ota_utils import (VABC_COMPRESSION_PARAM_SUPPORT, FinalizeMetadata, GetPackageMetadata,
- PayloadGenerator, SECURITY_PATCH_LEVEL_PROP_NAME, ExtractTargetFiles, CopyTargetFilesDir)
+ PayloadGenerator, SECURITY_PATCH_LEVEL_PROP_NAME, ExtractTargetFiles, CopyTargetFilesDir, TARGET_FILES_IMAGES_SUBDIR)
from common import DoesInputFileContain, IsSparseImage
import target_files_diff
from non_ab_ota import GenerateNonAbOtaPackage
@@ -337,6 +341,7 @@
OPTIONS.max_threads = None
OPTIONS.vabc_cow_version = None
OPTIONS.compression_factor = None
+OPTIONS.full_ota_partitions = None
POSTINSTALL_CONFIG = 'META/postinstall_config.txt'
@@ -892,6 +897,14 @@
if source_file is not None:
source_file = ExtractTargetFiles(source_file)
+ if OPTIONS.full_ota_partitions:
+ for partition in OPTIONS.full_ota_partitions:
+ for subdir in TARGET_FILES_IMAGES_SUBDIR:
+ image_path = os.path.join(source_file, subdir, partition + ".img")
+ if os.path.exists(image_path):
+ logger.info(
+ "Ignoring source image %s for partition %s because it is configured to use full OTA", image_path, partition)
+ os.remove(image_path)
assert "ab_partitions" in OPTIONS.source_info_dict, \
"META/ab_partitions.txt is required for ab_update."
assert "ab_partitions" in OPTIONS.target_info_dict, \
@@ -1193,7 +1206,7 @@
def main(argv):
- def option_handler(o, a):
+ def option_handler(o, a: str):
if o in ("-i", "--incremental_from"):
OPTIONS.incremental_source = a
elif o == "--full_radio":
@@ -1320,6 +1333,9 @@
else:
raise ValueError("Cannot parse value %r for option %r - only "
"integers are allowed." % (a, o))
+ elif o == "--full_ota_partitions":
+ OPTIONS.full_ota_partitions = set(
+ a.strip().strip("\"").strip("'").split(","))
else:
return False
return True
@@ -1370,6 +1386,7 @@
"max_threads=",
"vabc_cow_version=",
"compression_factor=",
+ "full_ota_partitions=",
], extra_option_handler=[option_handler, payload_signer.signer_options])
common.InitLogging()
diff --git a/tools/releasetools/sign_target_files_apks.py b/tools/releasetools/sign_target_files_apks.py
index b485440..4ad97e0 100755
--- a/tools/releasetools/sign_target_files_apks.py
+++ b/tools/releasetools/sign_target_files_apks.py
@@ -184,16 +184,17 @@
import shutil
import stat
import sys
+import shlex
import tempfile
import zipfile
from xml.etree import ElementTree
import add_img_to_target_files
import ota_from_raw_img
-import ota_utils
import apex_utils
import common
import payload_signer
+import update_payload
from payload_signer import SignOtaPackage, PAYLOAD_BIN
@@ -223,6 +224,7 @@
OPTIONS.allow_gsi_debug_sepolicy = False
OPTIONS.override_apk_keys = None
OPTIONS.override_apex_keys = None
+OPTIONS.input_tmp = None
AVB_FOOTER_ARGS_BY_PARTITION = {
@@ -581,18 +583,34 @@
filename.endswith("/prop.default")
+def GetOtaSigningArgs():
+ args = []
+ if OPTIONS.package_key:
+ args.extend(["--package_key", OPTIONS.package_key])
+ if OPTIONS.payload_signer:
+ args.extend(["--payload_signer=" + OPTIONS.payload_signer])
+ if OPTIONS.payload_signer_args:
+ args.extend(["--payload_signer_args=" + shlex.join(OPTIONS.payload_signer_args)])
+ if OPTIONS.search_path:
+ args.extend(["--search_path", OPTIONS.search_path])
+ if OPTIONS.payload_signer_maximum_signature_size:
+ args.extend(["--payload_signer_maximum_signature_size",
+ OPTIONS.payload_signer_maximum_signature_size])
+ if OPTIONS.private_key_suffix:
+ args.extend(["--private_key_suffix", OPTIONS.private_key_suffix])
+ return args
+
+
def RegenerateKernelPartitions(input_tf_zip: zipfile.ZipFile, output_tf_zip: zipfile.ZipFile, misc_info):
"""Re-generate boot and dtbo partitions using new signing configuration"""
+ files_to_unzip = [
+ "PREBUILT_IMAGES/*", "BOOTABLE_IMAGES/*.img", "*/boot_16k.img", "*/dtbo_16k.img"]
if OPTIONS.input_tmp is None:
- OPTIONS.input_tmp = common.UnzipTemp(input_tf_zip.filename, [
- "*/boot.img", "*/dtbo.img"])
+ OPTIONS.input_tmp = common.UnzipTemp(input_tf_zip.filename, files_to_unzip)
else:
- common.UnzipToDir(input_tf_zip, OPTIONS.input_tmp, [
- "*/boot.img", "*/dtbo.img"])
+ common.UnzipToDir(input_tf_zip.filename, OPTIONS.input_tmp, files_to_unzip)
unzip_dir = OPTIONS.input_tmp
- image_dir = os.path.join(unzip_dir, "IMAGES")
- shutil.rmtree(image_dir)
- os.makedirs(image_dir, exist_ok=True)
+ os.makedirs(os.path.join(unzip_dir, "IMAGES"), exist_ok=True)
boot_image = common.GetBootableImage(
"IMAGES/boot.img", "boot.img", unzip_dir, "BOOT", misc_info)
@@ -601,37 +619,64 @@
boot_image = os.path.join(unzip_dir, boot_image.name)
common.ZipWrite(output_tf_zip, boot_image, "IMAGES/boot.img",
compress_type=zipfile.ZIP_STORED)
- add_img_to_target_files.AddDtbo(output_tf_zip)
+ if misc_info.get("has_dtbo") == "true":
+ add_img_to_target_files.AddDtbo(output_tf_zip)
return unzip_dir
-def RegenerateBootOTA(input_tf_zip: zipfile.ZipFile, output_tf_zip: zipfile.ZipFile, misc_info, filename, input_ota):
- if filename not in ["VENDOR/boot_otas/boot_ota_4k.zip", "SYSTEM/boot_otas/boot_ota_4k.zip"]:
- # We only need to re-generate 4K boot OTA, for other OTA packages
- # simply copy as is
- with input_tf_zip.open(filename, "r") as in_fp:
- shutil.copyfileobj(in_fp, input_ota)
- input_ota.flush()
+def RegenerateBootOTA(input_tf_zip: zipfile.ZipFile, filename, input_ota):
+ with input_tf_zip.open(filename, "r") as in_fp:
+ payload = update_payload.Payload(in_fp)
+ is_incremental = any([part.HasField('old_partition_info')
+ for part in payload.manifest.partitions])
+ is_boot_ota = filename.startswith(
+ "VENDOR/boot_otas/") or filename.startswith("SYSTEM/boot_otas/")
+ if not is_boot_ota:
return
- timestamp = misc_info["build.prop"].GetProp(
- "ro.system.build.date.utc")
- unzip_dir = RegenerateKernelPartitions(
- input_tf_zip, output_tf_zip, misc_info)
- signed_boot_image = os.path.join(unzip_dir, "IMAGES/boot.img")
- signed_dtbo_image = os.path.join(unzip_dir, "IMAGES/dtbo.img")
+ is_4k_boot_ota = filename in [
+ "VENDOR/boot_otas/boot_ota_4k.zip", "SYSTEM/boot_otas/boot_ota_4k.zip"]
+ # Only 4K boot image is re-generated, so if 16K boot ota isn't incremental,
+ # we do not need to re-generate
+ if not is_4k_boot_ota and not is_incremental:
+ return
+ timestamp = str(payload.manifest.max_timestamp)
+ partitions = [part.partition_name for part in payload.manifest.partitions]
+ unzip_dir = OPTIONS.input_tmp
+ signed_boot_image = os.path.join(unzip_dir, "IMAGES", "boot.img")
if not os.path.exists(signed_boot_image):
logger.warn("Need to re-generate boot OTA {} but failed to get signed boot image. 16K dev option will be impacted, after rolling back to 4K user would need to sideload/flash their device to continue receiving OTAs.")
return
- logger.info(
- "Re-generating boot OTA {} with timestamp {}".format(filename, timestamp))
- args = ["ota_from_raw_img", "--package_key", OPTIONS.package_key,
+ signed_dtbo_image = os.path.join(unzip_dir, "IMAGES", "dtbo.img")
+ if "dtbo" in partitions and not os.path.exists(signed_dtbo_image):
+ raise ValueError(
+ "Boot OTA {} has dtbo partition, but no dtbo image found in target files.".format(filename))
+ if is_incremental:
+ signed_16k_boot_image = os.path.join(
+ unzip_dir, "IMAGES", "boot_16k.img")
+ signed_16k_dtbo_image = os.path.join(
+ unzip_dir, "IMAGES", "dtbo_16k.img")
+ if is_4k_boot_ota:
+ if os.path.exists(signed_16k_boot_image):
+ signed_boot_image = signed_16k_boot_image + ":" + signed_boot_image
+ if os.path.exists(signed_16k_dtbo_image):
+ signed_dtbo_image = signed_16k_dtbo_image + ":" + signed_dtbo_image
+ else:
+ if os.path.exists(signed_16k_boot_image):
+ signed_boot_image += ":" + signed_16k_boot_image
+ if os.path.exists(signed_16k_dtbo_image):
+ signed_dtbo_image += ":" + signed_16k_dtbo_image
+
+ args = ["ota_from_raw_img",
"--max_timestamp", timestamp, "--output", input_ota.name]
- if os.path.exists(signed_dtbo_image):
+ args.extend(GetOtaSigningArgs())
+ if "dtbo" in partitions:
args.extend(["--partition_name", "boot,dtbo",
signed_boot_image, signed_dtbo_image])
else:
args.extend(["--partition_name", "boot", signed_boot_image])
+ logger.info(
+ "Re-generating boot OTA {} using cmd {}".format(filename, args))
ota_from_raw_img.main(args)
@@ -657,6 +702,8 @@
if misc_info.get('avb_enable') == 'true':
RewriteAvbProps(misc_info)
+ RegenerateKernelPartitions(input_tf_zip, output_tf_zip, misc_info)
+
for info in input_tf_zip.infolist():
filename = info.filename
if filename.startswith("IMAGES/"):
@@ -667,10 +714,10 @@
if filename.startswith("OTA/") and filename.endswith(".img"):
continue
- data = input_tf_zip.read(filename)
- out_info = copy.copy(info)
(is_apk, is_compressed, should_be_skipped) = GetApkFileInfo(
filename, compressed_extension, OPTIONS.skip_apks_with_path_prefix)
+ data = input_tf_zip.read(filename)
+ out_info = copy.copy(info)
if is_apk and should_be_skipped:
# Copy skipped APKs verbatim.
@@ -734,8 +781,7 @@
elif filename.endswith(".zip") and IsEntryOtaPackage(input_tf_zip, filename):
logger.info("Re-signing OTA package {}".format(filename))
with tempfile.NamedTemporaryFile() as input_ota, tempfile.NamedTemporaryFile() as output_ota:
- RegenerateBootOTA(input_tf_zip, output_tf_zip,
- misc_info, filename, input_ota)
+ RegenerateBootOTA(input_tf_zip, filename, input_ota)
SignOtaPackage(input_ota.name, output_ota.name)
common.ZipWrite(output_tf_zip, output_ota.name, filename,
@@ -1131,9 +1177,9 @@
common.ZipWriteStr(output_zip, filename, temp_file.getvalue())
-def ReplaceOtaKeys(input_tf_zip, output_tf_zip, misc_info):
+def ReplaceOtaKeys(input_tf_zip: zipfile.ZipFile, output_tf_zip, misc_info):
try:
- keylist = input_tf_zip.read("META/otakeys.txt").split()
+ keylist = input_tf_zip.read("META/otakeys.txt").decode().split()
except KeyError:
raise common.ExternalError("can't read META/otakeys.txt from input")
diff --git a/tools/sbom/Android.bp b/tools/sbom/Android.bp
index 6901b06..4f6d3b7 100644
--- a/tools/sbom/Android.bp
+++ b/tools/sbom/Android.bp
@@ -33,6 +33,13 @@
],
}
+python_library_host {
+ name: "compliance_metadata",
+ srcs: [
+ "compliance_metadata.py",
+ ],
+}
+
python_binary_host {
name: "gen_sbom",
srcs: [
@@ -44,6 +51,7 @@
},
},
libs: [
+ "compliance_metadata",
"metadata_file_proto_py",
"libprotobuf-python",
"sbom_lib",
@@ -109,3 +117,17 @@
"sbom_lib",
],
}
+
+python_binary_host {
+ name: "gen_notice_xml",
+ srcs: [
+ "gen_notice_xml.py",
+ ],
+ version: {
+ py3: {
+ embedded_launcher: true,
+ },
+ },
+ libs: [
+ ],
+}
diff --git a/tools/sbom/compliance_metadata.py b/tools/sbom/compliance_metadata.py
new file mode 100644
index 0000000..9910217
--- /dev/null
+++ b/tools/sbom/compliance_metadata.py
@@ -0,0 +1,204 @@
+#!/usr/bin/env python3
+#
+# Copyright (C) 2024 The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+import sqlite3
+
+class MetadataDb:
+ def __init__(self, db):
+ self.conn = sqlite3.connect(':memory')
+ self.conn.row_factory = sqlite3.Row
+ with sqlite3.connect(db) as c:
+ c.backup(self.conn)
+ self.reorg()
+
+ def reorg(self):
+ # package_license table
+ self.conn.execute("create table package_license as "
+ "select name as package, pkg_default_applicable_licenses as license "
+ "from modules "
+ "where module_type = 'package' ")
+ cursor = self.conn.execute("select package,license from package_license where license like '% %'")
+ multi_licenses_packages = cursor.fetchall()
+ cursor.close()
+ rows = []
+ for p in multi_licenses_packages:
+ licenses = p['license'].strip().split(' ')
+ for lic in licenses:
+ rows.append((p['package'], lic))
+ self.conn.executemany('insert into package_license values (?, ?)', rows)
+ self.conn.commit()
+
+ self.conn.execute("delete from package_license where license like '% %'")
+ self.conn.commit()
+
+ # module_license table
+ self.conn.execute("create table module_license as "
+ "select distinct name as module, package, licenses as license "
+ "from modules "
+ "where licenses != '' ")
+ cursor = self.conn.execute("select module,package,license from module_license where license like '% %'")
+ multi_licenses_modules = cursor.fetchall()
+ cursor.close()
+ rows = []
+ for m in multi_licenses_modules:
+ licenses = m['license'].strip().split(' ')
+ for lic in licenses:
+ rows.append((m['module'], m['package'],lic))
+ self.conn.executemany('insert into module_license values (?, ?, ?)', rows)
+ self.conn.commit()
+
+ self.conn.execute("delete from module_license where license like '% %'")
+ self.conn.commit()
+
+ # module_installed_file table
+ self.conn.execute("create table module_installed_file as "
+ "select id as module_id, name as module_name, package, installed_files as installed_file "
+ "from modules "
+ "where installed_files != '' ")
+ cursor = self.conn.execute("select module_id, module_name, package, installed_file "
+ "from module_installed_file where installed_file like '% %'")
+ multi_installed_file_modules = cursor.fetchall()
+ cursor.close()
+ rows = []
+ for m in multi_installed_file_modules:
+ installed_files = m['installed_file'].strip().split(' ')
+ for f in installed_files:
+ rows.append((m['module_id'], m['module_name'], m['package'], f))
+ self.conn.executemany('insert into module_installed_file values (?, ?, ?, ?)', rows)
+ self.conn.commit()
+
+ self.conn.execute("delete from module_installed_file where installed_file like '% %'")
+ self.conn.commit()
+
+ # module_built_file table
+ self.conn.execute("create table module_built_file as "
+ "select id as module_id, name as module_name, package, built_files as built_file "
+ "from modules "
+ "where built_files != '' ")
+ cursor = self.conn.execute("select module_id, module_name, package, built_file "
+ "from module_built_file where built_file like '% %'")
+ multi_built_file_modules = cursor.fetchall()
+ cursor.close()
+ rows = []
+ for m in multi_built_file_modules:
+ built_files = m['installed_file'].strip().split(' ')
+ for f in built_files:
+ rows.append((m['module_id'], m['module_name'], m['package'], f))
+ self.conn.executemany('insert into module_built_file values (?, ?, ?, ?)', rows)
+ self.conn.commit()
+
+ self.conn.execute("delete from module_built_file where built_file like '% %'")
+ self.conn.commit()
+
+
+ # Indexes
+ self.conn.execute('create index idx_modules_id on modules (id)')
+ self.conn.execute('create index idx_modules_name on modules (name)')
+ self.conn.execute('create index idx_package_licnese_package on package_license (package)')
+ self.conn.execute('create index idx_package_licnese_license on package_license (license)')
+ self.conn.execute('create index idx_module_licnese_module on module_license (module)')
+ self.conn.execute('create index idx_module_licnese_license on module_license (license)')
+ self.conn.execute('create index idx_module_installed_file_module_id on module_installed_file (module_id)')
+ self.conn.execute('create index idx_module_installed_file_installed_file on module_installed_file (installed_file)')
+ self.conn.execute('create index idx_module_built_file_module_id on module_built_file (module_id)')
+ self.conn.execute('create index idx_module_built_file_built_file on module_built_file (built_file)')
+ self.conn.commit()
+
+ def dump_debug_db(self, debug_db):
+ with sqlite3.connect(debug_db) as c:
+ self.conn.backup(c)
+
+ def get_installed_files(self):
+ # Get all records from table make_metadata, which contains all installed files and corresponding make modules' metadata
+ cursor = self.conn.execute('select installed_file, module_path, is_prebuilt_make_module, product_copy_files, kernel_module_copy_files, is_platform_generated, license_text from make_metadata')
+ rows = cursor.fetchall()
+ cursor.close()
+ installed_files_metadata = []
+ for row in rows:
+ metadata = dict(zip(row.keys(), row))
+ installed_files_metadata.append(metadata)
+ return installed_files_metadata
+
+ def get_soong_modules(self):
+ # Get all records from table modules, which contains metadata of all soong modules
+ cursor = self.conn.execute('select name, package, package as module_path, module_type as soong_module_type, built_files, installed_files, static_dep_files, whole_static_dep_files from modules')
+ rows = cursor.fetchall()
+ cursor.close()
+ soong_modules = []
+ for row in rows:
+ soong_module = dict(zip(row.keys(), row))
+ soong_modules.append(soong_module)
+ return soong_modules
+
+ def get_package_licenses(self, package):
+ cursor = self.conn.execute('select m.name, m.package, m.lic_license_text as license_text '
+ 'from package_license pl join modules m on pl.license = m.name '
+ 'where pl.package = ?',
+ ('//' + package,))
+ rows = cursor.fetchall()
+ licenses = {}
+ for r in rows:
+ licenses[r['name']] = r['license_text']
+ return licenses
+
+ def get_module_licenses(self, module_name, package):
+ licenses = {}
+ # If property "licenses" is defined on module
+ cursor = self.conn.execute('select m.name, m.package, m.lic_license_text as license_text '
+ 'from module_license ml join modules m on ml.license = m.name '
+ 'where ml.module = ? and ml.package = ?',
+ (module_name, package))
+ rows = cursor.fetchall()
+ for r in rows:
+ licenses[r['name']] = r['license_text']
+ if len(licenses) > 0:
+ return licenses
+
+ # Use default package license
+ cursor = self.conn.execute('select m.name, m.package, m.lic_license_text as license_text '
+ 'from package_license pl join modules m on pl.license = m.name '
+ 'where pl.package = ?',
+ ('//' + package,))
+ rows = cursor.fetchall()
+ for r in rows:
+ licenses[r['name']] = r['license_text']
+ return licenses
+
+ def get_soong_module_of_installed_file(self, installed_file):
+ cursor = self.conn.execute('select name, m.package, m.package as module_path, module_type as soong_module_type, built_files, installed_files, static_dep_files, whole_static_dep_files '
+ 'from modules m join module_installed_file mif on m.id = mif.module_id '
+ 'where mif.installed_file = ?',
+ (installed_file,))
+ rows = cursor.fetchall()
+ cursor.close()
+ if rows:
+ soong_module = dict(zip(rows[0].keys(), rows[0]))
+ return soong_module
+
+ return None
+
+ def get_soong_module_of_built_file(self, built_file):
+ cursor = self.conn.execute('select name, m.package, m.package as module_path, module_type as soong_module_type, built_files, installed_files, static_dep_files, whole_static_dep_files '
+ 'from modules m join module_built_file mbf on m.id = mbf.module_id '
+ 'where mbf.built_file = ?',
+ (built_file,))
+ rows = cursor.fetchall()
+ cursor.close()
+ if rows:
+ soong_module = dict(zip(rows[0].keys(), rows[0]))
+ return soong_module
+
+ return None
\ No newline at end of file
diff --git a/tools/sbom/gen_notice_xml.py b/tools/sbom/gen_notice_xml.py
new file mode 100644
index 0000000..eaa6e5a
--- /dev/null
+++ b/tools/sbom/gen_notice_xml.py
@@ -0,0 +1,81 @@
+# !/usr/bin/env python3
+#
+# Copyright (C) 2024 The Android Open Source Project
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
+
+"""
+Generate NOTICE.xml.gz of a partition.
+Usage example:
+ gen_notice_xml.py --output_file out/soong/.intermediate/.../NOTICE.xml.gz \
+ --metadata out/soong/compliance-metadata/aosp_cf_x86_64_phone/compliance-metadata.db \
+ --partition system \
+ --product_out out/target/vsoc_x86_64 \
+ --soong_out out/soong
+"""
+
+import argparse
+
+
+FILE_HEADER = '''\
+<?xml version="1.0" encoding="utf-8"?>
+<licenses>
+'''
+FILE_FOOTER = '''\
+</licenses>
+'''
+
+
+def get_args():
+ parser = argparse.ArgumentParser()
+ parser.add_argument('-v', '--verbose', action='store_true', default=False, help='Print more information.')
+ parser.add_argument('-d', '--debug', action='store_true', default=True, help='Debug mode')
+ parser.add_argument('--output_file', required=True, help='The path of the generated NOTICE.xml.gz file.')
+ parser.add_argument('--partition', required=True, help='The name of partition for which the NOTICE.xml.gz is generated.')
+ parser.add_argument('--metadata', required=True, help='The path of compliance metadata DB file.')
+ parser.add_argument('--product_out', required=True, help='The path of PRODUCT_OUT, e.g. out/target/product/vsoc_x86_64.')
+ parser.add_argument('--soong_out', required=True, help='The path of Soong output directory, e.g. out/soong')
+
+ return parser.parse_args()
+
+
+def log(*info):
+ if args.verbose:
+ for i in info:
+ print(i)
+
+
+def new_file_name_tag(file_metadata, package_name):
+ file_path = file_metadata['installed_file'].removeprefix(args.product_out)
+ lib = 'Android'
+ if package_name:
+ lib = package_name
+ return f'<file-name contentId="" lib="{lib}">{file_path}</file-name>\n'
+
+
+def new_file_content_tag():
+ pass
+
+
+def main():
+ global args
+ args = get_args()
+ log('Args:', vars(args))
+
+ with open(args.output_file, 'w', encoding="utf-8") as notice_xml_file:
+ notice_xml_file.write(FILE_HEADER)
+ notice_xml_file.write(FILE_FOOTER)
+
+
+if __name__ == '__main__':
+ main()
diff --git a/tools/sbom/gen_sbom.py b/tools/sbom/gen_sbom.py
index a203258..9c3a8be 100644
--- a/tools/sbom/gen_sbom.py
+++ b/tools/sbom/gen_sbom.py
@@ -26,6 +26,7 @@
"""
import argparse
+import compliance_metadata
import datetime
import google.protobuf.text_format as text_format
import hashlib
@@ -35,7 +36,6 @@
import metadata_file_pb2
import sbom_data
import sbom_writers
-import sqlite3
# Package type
PKG_SOURCE = 'SOURCE'
@@ -568,202 +568,16 @@
return sorted(all_static_dep_files.keys())
-class MetadataDb:
- def __init__(self, db):
- self.conn = sqlite3.connect(':memory')
- self.conn.row_factory = sqlite3.Row
- with sqlite3.connect(db) as c:
- c.backup(self.conn)
- self.reorg()
-
- def reorg(self):
- # package_license table
- self.conn.execute("create table package_license as "
- "select name as package, pkg_default_applicable_licenses as license "
- "from modules "
- "where module_type = 'package' ")
- cursor = self.conn.execute("select package,license from package_license where license like '% %'")
- multi_licenses_packages = cursor.fetchall()
- cursor.close()
- rows = []
- for p in multi_licenses_packages:
- licenses = p['license'].strip().split(' ')
- for lic in licenses:
- rows.append((p['package'], lic))
- self.conn.executemany('insert into package_license values (?, ?)', rows)
- self.conn.commit()
-
- self.conn.execute("delete from package_license where license like '% %'")
- self.conn.commit()
-
- # module_license table
- self.conn.execute("create table module_license as "
- "select distinct name as module, package, licenses as license "
- "from modules "
- "where licenses != '' ")
- cursor = self.conn.execute("select module,package,license from module_license where license like '% %'")
- multi_licenses_modules = cursor.fetchall()
- cursor.close()
- rows = []
- for m in multi_licenses_modules:
- licenses = m['license'].strip().split(' ')
- for lic in licenses:
- rows.append((m['module'], m['package'],lic))
- self.conn.executemany('insert into module_license values (?, ?, ?)', rows)
- self.conn.commit()
-
- self.conn.execute("delete from module_license where license like '% %'")
- self.conn.commit()
-
- # module_installed_file table
- self.conn.execute("create table module_installed_file as "
- "select id as module_id, name as module_name, package, installed_files as installed_file "
- "from modules "
- "where installed_files != '' ")
- cursor = self.conn.execute("select module_id, module_name, package, installed_file "
- "from module_installed_file where installed_file like '% %'")
- multi_installed_file_modules = cursor.fetchall()
- cursor.close()
- rows = []
- for m in multi_installed_file_modules:
- installed_files = m['installed_file'].strip().split(' ')
- for f in installed_files:
- rows.append((m['module_id'], m['module_name'], m['package'], f))
- self.conn.executemany('insert into module_installed_file values (?, ?, ?, ?)', rows)
- self.conn.commit()
-
- self.conn.execute("delete from module_installed_file where installed_file like '% %'")
- self.conn.commit()
-
- # module_built_file table
- self.conn.execute("create table module_built_file as "
- "select id as module_id, name as module_name, package, built_files as built_file "
- "from modules "
- "where built_files != '' ")
- cursor = self.conn.execute("select module_id, module_name, package, built_file "
- "from module_built_file where built_file like '% %'")
- multi_built_file_modules = cursor.fetchall()
- cursor.close()
- rows = []
- for m in multi_built_file_modules:
- built_files = m['installed_file'].strip().split(' ')
- for f in built_files:
- rows.append((m['module_id'], m['module_name'], m['package'], f))
- self.conn.executemany('insert into module_built_file values (?, ?, ?, ?)', rows)
- self.conn.commit()
-
- self.conn.execute("delete from module_built_file where built_file like '% %'")
- self.conn.commit()
-
-
- # Indexes
- self.conn.execute('create index idx_modules_id on modules (id)')
- self.conn.execute('create index idx_modules_name on modules (name)')
- self.conn.execute('create index idx_package_licnese_package on package_license (package)')
- self.conn.execute('create index idx_package_licnese_license on package_license (license)')
- self.conn.execute('create index idx_module_licnese_module on module_license (module)')
- self.conn.execute('create index idx_module_licnese_license on module_license (license)')
- self.conn.execute('create index idx_module_installed_file_module_id on module_installed_file (module_id)')
- self.conn.execute('create index idx_module_installed_file_installed_file on module_installed_file (installed_file)')
- self.conn.execute('create index idx_module_built_file_module_id on module_built_file (module_id)')
- self.conn.execute('create index idx_module_built_file_built_file on module_built_file (built_file)')
- self.conn.commit()
-
- if args.debug:
- with sqlite3.connect(os.path.dirname(args.metadata) + '/compliance-metadata-debug.db') as c:
- self.conn.backup(c)
-
-
- def get_installed_files(self):
- # Get all records from table make_metadata, which contains all installed files and corresponding make modules' metadata
- cursor = self.conn.execute('select installed_file, module_path, is_prebuilt_make_module, product_copy_files, kernel_module_copy_files, is_platform_generated, license_text from make_metadata')
- rows = cursor.fetchall()
- cursor.close()
- installed_files_metadata = []
- for row in rows:
- metadata = dict(zip(row.keys(), row))
- installed_files_metadata.append(metadata)
- return installed_files_metadata
-
- def get_soong_modules(self):
- # Get all records from table modules, which contains metadata of all soong modules
- cursor = self.conn.execute('select name, package, package as module_path, module_type as soong_module_type, built_files, installed_files, static_dep_files, whole_static_dep_files from modules')
- rows = cursor.fetchall()
- cursor.close()
- soong_modules = []
- for row in rows:
- soong_module = dict(zip(row.keys(), row))
- soong_modules.append(soong_module)
- return soong_modules
-
- def get_package_licenses(self, package):
- cursor = self.conn.execute('select m.name, m.package, m.lic_license_text as license_text '
- 'from package_license pl join modules m on pl.license = m.name '
- 'where pl.package = ?',
- ('//' + package,))
- rows = cursor.fetchall()
- licenses = {}
- for r in rows:
- licenses[r['name']] = r['license_text']
- return licenses
-
- def get_module_licenses(self, module_name, package):
- licenses = {}
- # If property "licenses" is defined on module
- cursor = self.conn.execute('select m.name, m.package, m.lic_license_text as license_text '
- 'from module_license ml join modules m on ml.license = m.name '
- 'where ml.module = ? and ml.package = ?',
- (module_name, package))
- rows = cursor.fetchall()
- for r in rows:
- licenses[r['name']] = r['license_text']
- if len(licenses) > 0:
- return licenses
-
- # Use default package license
- cursor = self.conn.execute('select m.name, m.package, m.lic_license_text as license_text '
- 'from package_license pl join modules m on pl.license = m.name '
- 'where pl.package = ?',
- ('//' + package,))
- rows = cursor.fetchall()
- for r in rows:
- licenses[r['name']] = r['license_text']
- return licenses
-
- def get_soong_module_of_installed_file(self, installed_file):
- cursor = self.conn.execute('select name, m.package, m.package as module_path, module_type as soong_module_type, built_files, installed_files, static_dep_files, whole_static_dep_files '
- 'from modules m join module_installed_file mif on m.id = mif.module_id '
- 'where mif.installed_file = ?',
- (installed_file,))
- rows = cursor.fetchall()
- cursor.close()
- if rows:
- soong_module = dict(zip(rows[0].keys(), rows[0]))
- return soong_module
-
- return None
-
- def get_soong_module_of_built_file(self, built_file):
- cursor = self.conn.execute('select name, m.package, m.package as module_path, module_type as soong_module_type, built_files, installed_files, static_dep_files, whole_static_dep_files '
- 'from modules m join module_built_file mbf on m.id = mbf.module_id '
- 'where mbf.built_file = ?',
- (built_file,))
- rows = cursor.fetchall()
- cursor.close()
- if rows:
- soong_module = dict(zip(rows[0].keys(), rows[0]))
- return soong_module
-
- return None
-
-
def main():
global args
args = get_args()
log('Args:', vars(args))
global db
- db = MetadataDb(args.metadata)
+ db = compliance_metadata.MetadataDb(args.metadata)
+ if args.debug:
+ db.dump_debug_db(os.path.dirname(args.output_file) + '/compliance-metadata-debug.db')
+
global metadata_file_protos
metadata_file_protos = {}
global licenses_text
diff --git a/tools/signapk/src/com/android/signapk/SignApk.java b/tools/signapk/src/com/android/signapk/SignApk.java
index 6b2341b..654e196 100644
--- a/tools/signapk/src/com/android/signapk/SignApk.java
+++ b/tools/signapk/src/com/android/signapk/SignApk.java
@@ -302,7 +302,6 @@
final KeyStore keyStore, final String keyName)
throws CertificateException, KeyStoreException, NoSuchAlgorithmException,
UnrecoverableKeyException, UnrecoverableEntryException {
- final Key key = keyStore.getKey(keyName, readPassword(keyName));
final PrivateKeyEntry privateKeyEntry = (PrivateKeyEntry) keyStore.getEntry(keyName, null);
if (privateKeyEntry == null) {
throw new Error(