|  | // | 
|  | // Copyright (C) 2017 The Android Open Source Project | 
|  | // | 
|  | // Licensed under the Apache License, Version 2.0 (the "License"); | 
|  | // you may not use this file except in compliance with the License. | 
|  | // You may obtain a copy of the License at | 
|  | // | 
|  | //      http://www.apache.org/licenses/LICENSE-2.0 | 
|  | // | 
|  | // Unless required by applicable law or agreed to in writing, software | 
|  | // distributed under the License is distributed on an "AS IS" BASIS, | 
|  | // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. | 
|  | // See the License for the specific language governing permissions and | 
|  | // limitations under the License. | 
|  | // | 
|  |  | 
|  | #include "update_engine/payload_generator/squashfs_filesystem.h" | 
|  |  | 
|  | #include <fcntl.h> | 
|  |  | 
|  | #include <algorithm> | 
|  | #include <string> | 
|  | #include <utility> | 
|  |  | 
|  | #include <base/files/file_util.h> | 
|  | #include <base/logging.h> | 
|  | #include <base/strings/string_number_conversions.h> | 
|  | #include <base/strings/string_split.h> | 
|  | #include <brillo/streams/file_stream.h> | 
|  |  | 
|  | #include "update_engine/common/subprocess.h" | 
|  | #include "update_engine/common/utils.h" | 
|  | #include "update_engine/payload_generator/deflate_utils.h" | 
|  | #include "update_engine/payload_generator/delta_diff_generator.h" | 
|  | #include "update_engine/payload_generator/extent_ranges.h" | 
|  | #include "update_engine/payload_generator/extent_utils.h" | 
|  | #include "update_engine/update_metadata.pb.h" | 
|  |  | 
|  | using std::string; | 
|  | using std::unique_ptr; | 
|  | using std::vector; | 
|  |  | 
|  | namespace chromeos_update_engine { | 
|  |  | 
|  | namespace { | 
|  |  | 
|  | Extent ExtentForBytes(uint64_t block_size, | 
|  | uint64_t start_bytes, | 
|  | uint64_t size_bytes) { | 
|  | uint64_t start_block = start_bytes / block_size; | 
|  | uint64_t end_block = (start_bytes + size_bytes + block_size - 1) / block_size; | 
|  | return ExtentForRange(start_block, end_block - start_block); | 
|  | } | 
|  |  | 
|  | // The size of the squashfs super block. | 
|  | constexpr size_t kSquashfsSuperBlockSize = 96; | 
|  | constexpr uint64_t kSquashfsCompressedBit = 1 << 24; | 
|  | constexpr uint32_t kSquashfsZlibCompression = 1; | 
|  |  | 
|  | bool ReadSquashfsHeader(const brillo::Blob blob, | 
|  | SquashfsFilesystem::SquashfsHeader* header) { | 
|  | if (blob.size() < kSquashfsSuperBlockSize) { | 
|  | return false; | 
|  | } | 
|  |  | 
|  | memcpy(&header->magic, blob.data(), 4); | 
|  | memcpy(&header->block_size, blob.data() + 12, 4); | 
|  | memcpy(&header->compression_type, blob.data() + 20, 2); | 
|  | memcpy(&header->major_version, blob.data() + 28, 2); | 
|  | return true; | 
|  | } | 
|  |  | 
|  | bool CheckHeader(const SquashfsFilesystem::SquashfsHeader& header) { | 
|  | return header.magic == 0x73717368 && header.major_version == 4; | 
|  | } | 
|  |  | 
|  | bool GetFileMapContent(const string& sqfs_path, string* map) { | 
|  | // Create a tmp file | 
|  | string map_file; | 
|  | TEST_AND_RETURN_FALSE( | 
|  | utils::MakeTempFile("squashfs_file_map.XXXXXX", &map_file, nullptr)); | 
|  | ScopedPathUnlinker map_unlinker(map_file); | 
|  |  | 
|  | // Run unsquashfs to get the system file map. | 
|  | // unsquashfs -m <map-file> <squashfs-file> | 
|  | vector<string> cmd = {"unsquashfs", "-m", map_file, sqfs_path}; | 
|  | string stdout; | 
|  | int exit_code; | 
|  | if (!Subprocess::SynchronousExec(cmd, &exit_code, &stdout) || | 
|  | exit_code != 0) { | 
|  | LOG(ERROR) << "Failed to run unsquashfs -m. The stdout content was: " | 
|  | << stdout; | 
|  | return false; | 
|  | } | 
|  | TEST_AND_RETURN_FALSE(utils::ReadFile(map_file, map)); | 
|  | return true; | 
|  | } | 
|  |  | 
|  | }  // namespace | 
|  |  | 
|  | bool SquashfsFilesystem::Init(const string& map, | 
|  | const string& sqfs_path, | 
|  | size_t size, | 
|  | const SquashfsHeader& header, | 
|  | bool extract_deflates) { | 
|  | size_ = size; | 
|  |  | 
|  | bool is_zlib = header.compression_type == kSquashfsZlibCompression; | 
|  | if (!is_zlib) { | 
|  | LOG(WARNING) << "Filesystem is not Gzipped. Not filling deflates!"; | 
|  | } | 
|  | vector<puffin::ByteExtent> zlib_blks; | 
|  |  | 
|  | // Reading files map. For the format of the file map look at the comments for | 
|  | // |CreateFromFileMap()|. | 
|  | auto lines = base::SplitStringPiece(map, | 
|  | "\n", | 
|  | base::WhitespaceHandling::KEEP_WHITESPACE, | 
|  | base::SplitResult::SPLIT_WANT_NONEMPTY); | 
|  | for (const auto& line : lines) { | 
|  | auto splits = | 
|  | base::SplitStringPiece(line, | 
|  | " \t", | 
|  | base::WhitespaceHandling::TRIM_WHITESPACE, | 
|  | base::SplitResult::SPLIT_WANT_NONEMPTY); | 
|  | // Only filename is invalid. | 
|  | TEST_AND_RETURN_FALSE(splits.size() > 1); | 
|  | uint64_t start; | 
|  | TEST_AND_RETURN_FALSE(base::StringToUint64(splits[1], &start)); | 
|  | uint64_t cur_offset = start; | 
|  | for (size_t i = 2; i < splits.size(); ++i) { | 
|  | uint64_t blk_size; | 
|  | TEST_AND_RETURN_FALSE(base::StringToUint64(splits[i], &blk_size)); | 
|  | // TODO(ahassani): For puffin push it into a proper list if uncompressed. | 
|  | auto new_blk_size = blk_size & ~kSquashfsCompressedBit; | 
|  | TEST_AND_RETURN_FALSE(new_blk_size <= header.block_size); | 
|  | if (new_blk_size > 0 && !(blk_size & kSquashfsCompressedBit)) { | 
|  | // Compressed block | 
|  | if (is_zlib && extract_deflates) { | 
|  | zlib_blks.emplace_back(cur_offset, new_blk_size); | 
|  | } | 
|  | } | 
|  | cur_offset += new_blk_size; | 
|  | } | 
|  |  | 
|  | // If size is zero do not add the file. | 
|  | if (cur_offset - start > 0) { | 
|  | File file; | 
|  | file.name = splits[0].as_string(); | 
|  | file.extents = {ExtentForBytes(kBlockSize, start, cur_offset - start)}; | 
|  | files_.emplace_back(file); | 
|  | } | 
|  | } | 
|  |  | 
|  | // Sort all files by their offset in the squashfs. | 
|  | std::sort(files_.begin(), files_.end(), [](const File& a, const File& b) { | 
|  | return a.extents[0].start_block() < b.extents[0].start_block(); | 
|  | }); | 
|  | // If there is any overlap between two consecutive extents, remove them. Here | 
|  | // we are assuming all files have exactly one extent. If this assumption | 
|  | // changes then this implementation needs to change too. | 
|  | for (auto first = files_.begin(), second = first + 1; | 
|  | first != files_.end() && second != files_.end(); | 
|  | second = first + 1) { | 
|  | auto first_begin = first->extents[0].start_block(); | 
|  | auto first_end = first_begin + first->extents[0].num_blocks(); | 
|  | auto second_begin = second->extents[0].start_block(); | 
|  | auto second_end = second_begin + second->extents[0].num_blocks(); | 
|  | // Remove the first file if the size is zero. | 
|  | if (first_end == first_begin) { | 
|  | first = files_.erase(first); | 
|  | } else if (first_end > second_begin) {  // We found a collision. | 
|  | if (second_end <= first_end) { | 
|  | // Second file is inside the first file, remove the second file. | 
|  | second = files_.erase(second); | 
|  | } else if (first_begin == second_begin) { | 
|  | // First file is inside the second file, remove the first file. | 
|  | first = files_.erase(first); | 
|  | } else { | 
|  | // Remove overlapping extents from the first file. | 
|  | first->extents[0].set_num_blocks(second_begin - first_begin); | 
|  | ++first; | 
|  | } | 
|  | } else { | 
|  | ++first; | 
|  | } | 
|  | } | 
|  |  | 
|  | // Find all the metadata including superblock and add them to the list of | 
|  | // files. | 
|  | ExtentRanges file_extents; | 
|  | for (const auto& file : files_) { | 
|  | file_extents.AddExtents(file.extents); | 
|  | } | 
|  | vector<Extent> full = { | 
|  | ExtentForRange(0, (size_ + kBlockSize - 1) / kBlockSize)}; | 
|  | auto metadata_extents = FilterExtentRanges(full, file_extents); | 
|  | // For now there should be at most two extents. One for superblock and one for | 
|  | // metadata at the end. Just create appropriate files with <metadata-i> name. | 
|  | // We can add all these extents as one metadata too, but that violates the | 
|  | // contiguous write optimization. | 
|  | for (size_t i = 0; i < metadata_extents.size(); i++) { | 
|  | File file; | 
|  | file.name = "<metadata-" + std::to_string(i) + ">"; | 
|  | file.extents = {metadata_extents[i]}; | 
|  | files_.emplace_back(file); | 
|  | } | 
|  |  | 
|  | // Do one last sort before returning. | 
|  | std::sort(files_.begin(), files_.end(), [](const File& a, const File& b) { | 
|  | return a.extents[0].start_block() < b.extents[0].start_block(); | 
|  | }); | 
|  |  | 
|  | if (is_zlib && extract_deflates) { | 
|  | // If it is infact gzipped, then the sqfs_path should be valid to read its | 
|  | // content. | 
|  | TEST_AND_RETURN_FALSE(!sqfs_path.empty()); | 
|  | if (zlib_blks.empty()) { | 
|  | return true; | 
|  | } | 
|  |  | 
|  | // Sort zlib blocks. | 
|  | std::sort(zlib_blks.begin(), | 
|  | zlib_blks.end(), | 
|  | [](const puffin::ByteExtent& a, const puffin::ByteExtent& b) { | 
|  | return a.offset < b.offset; | 
|  | }); | 
|  |  | 
|  | // Sanity check. Make sure zlib blocks are not overlapping. | 
|  | auto result = std::adjacent_find( | 
|  | zlib_blks.begin(), | 
|  | zlib_blks.end(), | 
|  | [](const puffin::ByteExtent& a, const puffin::ByteExtent& b) { | 
|  | return (a.offset + a.length) > b.offset; | 
|  | }); | 
|  | TEST_AND_RETURN_FALSE(result == zlib_blks.end()); | 
|  |  | 
|  | vector<puffin::BitExtent> deflates; | 
|  | TEST_AND_RETURN_FALSE( | 
|  | puffin::LocateDeflatesInZlibBlocks(sqfs_path, zlib_blks, &deflates)); | 
|  |  | 
|  | // Add deflates for each file. | 
|  | for (auto& file : files_) { | 
|  | file.deflates = deflate_utils::FindDeflates(file.extents, deflates); | 
|  | } | 
|  | } | 
|  | return true; | 
|  | } | 
|  |  | 
|  | unique_ptr<SquashfsFilesystem> SquashfsFilesystem::CreateFromFile( | 
|  | const string& sqfs_path, bool extract_deflates) { | 
|  | if (sqfs_path.empty()) | 
|  | return nullptr; | 
|  |  | 
|  | brillo::StreamPtr sqfs_file = | 
|  | brillo::FileStream::Open(base::FilePath(sqfs_path), | 
|  | brillo::Stream::AccessMode::READ, | 
|  | brillo::FileStream::Disposition::OPEN_EXISTING, | 
|  | nullptr); | 
|  | if (!sqfs_file) { | 
|  | LOG(ERROR) << "Unable to open " << sqfs_path << " for reading."; | 
|  | return nullptr; | 
|  | } | 
|  |  | 
|  | SquashfsHeader header; | 
|  | brillo::Blob blob(kSquashfsSuperBlockSize); | 
|  | if (!sqfs_file->ReadAllBlocking(blob.data(), blob.size(), nullptr)) { | 
|  | LOG(ERROR) << "Unable to read from file: " << sqfs_path; | 
|  | return nullptr; | 
|  | } | 
|  | if (!ReadSquashfsHeader(blob, &header) || !CheckHeader(header)) { | 
|  | // This is not necessary an error. | 
|  | return nullptr; | 
|  | } | 
|  |  | 
|  | // Read the map file. | 
|  | string filemap; | 
|  | if (!GetFileMapContent(sqfs_path, &filemap)) { | 
|  | LOG(ERROR) << "Failed to produce squashfs map file: " << sqfs_path; | 
|  | return nullptr; | 
|  | } | 
|  |  | 
|  | unique_ptr<SquashfsFilesystem> sqfs(new SquashfsFilesystem()); | 
|  | if (!sqfs->Init( | 
|  | filemap, sqfs_path, sqfs_file->GetSize(), header, extract_deflates)) { | 
|  | LOG(ERROR) << "Failed to initialized the Squashfs file system"; | 
|  | return nullptr; | 
|  | } | 
|  |  | 
|  | return sqfs; | 
|  | } | 
|  |  | 
|  | unique_ptr<SquashfsFilesystem> SquashfsFilesystem::CreateFromFileMap( | 
|  | const string& filemap, size_t size, const SquashfsHeader& header) { | 
|  | if (!CheckHeader(header)) { | 
|  | LOG(ERROR) << "Invalid Squashfs super block!"; | 
|  | return nullptr; | 
|  | } | 
|  |  | 
|  | unique_ptr<SquashfsFilesystem> sqfs(new SquashfsFilesystem()); | 
|  | if (!sqfs->Init(filemap, "", size, header, false)) { | 
|  | LOG(ERROR) << "Failed to initialize the Squashfs file system using filemap"; | 
|  | return nullptr; | 
|  | } | 
|  | // TODO(ahassani): Add a function that initializes the puffin related extents. | 
|  | return sqfs; | 
|  | } | 
|  |  | 
|  | size_t SquashfsFilesystem::GetBlockSize() const { | 
|  | return kBlockSize; | 
|  | } | 
|  |  | 
|  | size_t SquashfsFilesystem::GetBlockCount() const { | 
|  | return size_ / kBlockSize; | 
|  | } | 
|  |  | 
|  | bool SquashfsFilesystem::GetFiles(vector<File>* files) const { | 
|  | files->insert(files->end(), files_.begin(), files_.end()); | 
|  | return true; | 
|  | } | 
|  |  | 
|  | bool SquashfsFilesystem::LoadSettings(brillo::KeyValueStore* store) const { | 
|  | // Settings not supported in squashfs. | 
|  | LOG(ERROR) << "squashfs doesn't support LoadSettings()."; | 
|  | return false; | 
|  | } | 
|  |  | 
|  | bool SquashfsFilesystem::IsSquashfsImage(const brillo::Blob& blob) { | 
|  | SquashfsHeader header; | 
|  | return ReadSquashfsHeader(blob, &header) && CheckHeader(header); | 
|  | } | 
|  | }  // namespace chromeos_update_engine |