# Framework for mounting and patching macOS root volume # Copyright (C) 2020-2022, Dhinak G, Mykola Grymalyuk # System based off of Apple's Kernel Debug Kit (KDK) # - https://developer.apple.com/download/all/ # The system relies on mounting the APFS volume as a live read/write volume # We perform our required edits, then create a new snapshot for the system boot # The manual process is as follows: # 1. Find the Root Volume # 'diskutil info / | grep "Device Node:"' # 2. Convert Snapshot Device Node to Root Volume Device Node # /dev/disk3s1s1 -> /dev/disk3s1 (strip last 's1') # 3. Mount the APFS volume as a read/write volume # 'sudo mount -o nobrowse -t apfs /dev/disk5s5 /System/Volumes/Update/mnt1' # 4. Perform edits to the system (ie. create new KernelCollection) # 'sudo kmutil install --volume-root /System/Volumes/Update/mnt1/ --update-all' # 5. Create a new snapshot for the system boot # 'sudo bless --folder /System/Volumes/Update/mnt1/System/Library/CoreServices --bootefi --create-snapshot' # Additionally Apple's APFS snapshot system supports system rollbacks: # 'sudo bless --mount /System/Volumes/Update/mnt1 --bootefi --last-sealed-snapshot' # Note: root volume rollbacks are unstable in Big Sur due to quickly discarding the original snapshot # - Generally within 2~ boots, the original snapshot is discarded # - Monterey always preserves the original snapshot allowing for reliable rollbacks # Alternative to mounting via 'mount', Apple's update system uses 'mount_apfs' directly # '/sbin/mount_apfs -R /dev/disk5s5 /System/Volumes/Update/mnt1' # With macOS Ventura, you will also need to install the KDK onto root if you plan to use kmutil # This is because Apple removed on-disk binaries (ref: https://github.com/dortania/OpenCore-Legacy-Patcher/issues/998) # 'sudo ditto /Library/Developer/KDKs//System /System/Volumes/Update/mnt1/System' import logging import plistlib import subprocess import applescript from pathlib import Path from datetime import datetime from resources import constants, utilities, kdk_handler from resources.sys_patch import sys_patch_detect, sys_patch_auto, sys_patch_helpers, sys_patch_generate from data import os_data class PatchSysVolume: def __init__(self, model: str, global_constants: constants.Constants, hardware_details: list = None) -> None: self.model = model self.constants: constants.Constants = global_constants self.computer = self.constants.computer self.root_mount_path = None self.root_supports_snapshot = utilities.check_if_root_is_apfs_snapshot() self.constants.root_patcher_succeeded = False # Reset Variable each time we start self.constants.needs_to_open_preferences = False self.patch_set_dictionary = {} self.needs_kmutil_exemptions = False # For '/Library/Extensions' rebuilds self.kdk_path = None # GUI will detect hardware patches before starting PatchSysVolume() # However the TUI will not, so allow for data to be passed in manually avoiding multiple calls if hardware_details is None: hardware_details = sys_patch_detect.DetectRootPatch(self.computer.real_model, self.constants).detect_patch_set() self.hardware_details = hardware_details self._init_pathing(custom_root_mount_path=None, custom_data_mount_path=None) self.skip_root_kmutil_requirement = self.hardware_details["Settings: Supports Auxiliary Cache"] def _init_pathing(self, custom_root_mount_path: Path = None, custom_data_mount_path: Path = None) -> None: """ Initializes the pathing for root volume patching Parameters: custom_root_mount_path (Path): Custom path to mount the root volume custom_data_mount_path (Path): Custom path to mount the data volume """ if custom_root_mount_path and custom_data_mount_path: self.mount_location = custom_root_mount_path self.data_mount_location = custom_data_mount_path elif self.root_supports_snapshot is True: # Big Sur and newer use APFS snapshots self.mount_location = "/System/Volumes/Update/mnt1" self.mount_location_data = "" else: self.mount_location = "" self.mount_location_data = "" self.mount_extensions = f"{self.mount_location}/System/Library/Extensions" self.mount_application_support = f"{self.mount_location_data}/Library/Application Support" def _mount_root_vol(self) -> bool: """ Attempts to mount the booted APFS volume as a writable volume at /System/Volumes/Update/mnt1 Manual invocation: 'sudo mount -o nobrowse -t apfs /dev/diskXsY /System/Volumes/Update/mnt1' Returns: bool: True if successful, False if not """ # Returns boolean if Root Volume is available self.root_mount_path = utilities.get_disk_path() if self.root_mount_path.startswith("disk"): logging.info(f"- Found Root Volume at: {self.root_mount_path}") if Path(self.mount_extensions).exists(): logging.info("- Root Volume is already mounted") return True else: if self.root_supports_snapshot is True: logging.info("- Mounting APFS Snapshot as writable") result = utilities.elevated(["mount", "-o", "nobrowse", "-t", "apfs", f"/dev/{self.root_mount_path}", self.mount_location], stdout=subprocess.PIPE, stderr=subprocess.STDOUT) if result.returncode == 0: logging.info(f"- Mounted APFS Snapshot as writable at: {self.mount_location}") if Path(self.mount_extensions).exists(): logging.info("- Successfully mounted the Root Volume") return True else: logging.info("- Root Volume appears to have unmounted unexpectedly") else: logging.info("- Unable to mount APFS Snapshot as writable") logging.info("Reason for mount failure:") logging.info(result.stdout.decode().strip()) return False def _merge_kdk_with_root(self, save_hid_cs=False) -> None: """ Merge Kernel Debug Kit (KDK) with the root volume If no KDK is present, will call kdk_handler to download and install it Parameters: save_hid_cs (bool): If True, will save the HID CS file before merging KDK Required for USB 1.1 downgrades on Ventura and newer """ if self.skip_root_kmutil_requirement is True: return if self.constants.detected_os < os_data.os_data.ventura: return if self.constants.kdk_download_path.exists(): if kdk_handler.KernelDebugKitUtilities().install_kdk_dmg(self.constants.kdk_download_path) is False: logging.info("Failed to install KDK") raise Exception("Failed to install KDK") kdk_obj = kdk_handler.KernelDebugKitObject(self.constants, self.constants.detected_os_build, self.constants.detected_os_version) if kdk_obj.success is False: logging.info(f"Unable to get KDK info: {kdk_obj.error_msg}") raise Exception(f"Unable to get KDK info: {kdk_obj.error_msg}") if kdk_obj.kdk_already_installed is False: kdk_download_obj = kdk_obj.retrieve_download() if not kdk_download_obj: logging.info(f"Could not retrieve KDK: {kdk_obj.error_msg}") # Hold thread until download is complete kdk_download_obj.download(spawn_thread=False) if kdk_download_obj.download_complete is False: error_msg = kdk_download_obj.error_msg logging.info(f"Could not download KDK: {error_msg}") raise Exception(f"Could not download KDK: {error_msg}") if kdk_obj.validate_kdk_checksum() is False: logging.info(f"KDK checksum validation failed: {kdk_obj.error_msg}") raise Exception(f"KDK checksum validation failed: {kdk_obj.error_msg}") kdk_handler.KernelDebugKitUtilities().install_kdk_dmg(self.constants.kdk_download_path) # re-init kdk_obj to get the new kdk_installed_path kdk_obj = kdk_handler.KernelDebugKitObject(self.constants, self.constants.detected_os_build, self.constants.detected_os_version) if kdk_obj.success is False: logging.info(f"Unable to get KDK info: {kdk_obj.error_msg}") raise Exception(f"Unable to get KDK info: {kdk_obj.error_msg}") if kdk_obj.kdk_already_installed is False: # We shouldn't get here, but just in case logging.warning(f"KDK was not installed, but should have been: {kdk_obj.error_msg}") raise Exception("KDK was not installed, but should have been: {kdk_obj.error_msg}") kdk_path = Path(kdk_obj.kdk_installed_path) if kdk_obj.kdk_installed_path != "" else None oclp_plist = Path("/System/Library/CoreServices/OpenCore-Legacy-Patcher.plist") if (Path(self.mount_location) / Path("System/Library/Extensions/System.kext/PlugIns/Libkern.kext/Libkern")).exists() and oclp_plist.exists(): # KDK was already merged, check if the KDK used is the same as the one we're using # If not, we'll rsync over with the new KDK try: oclp_plist_data = plistlib.load(open(oclp_plist, "rb")) if "Kernel Debug Kit Used" in oclp_plist_data: if oclp_plist_data["Kernel Debug Kit Used"] == str(kdk_path): logging.info("- Matching KDK determined to already be merged, skipping") return except: pass if kdk_path is None: logging.info(f"- Unable to find Kernel Debug Kit") raise Exception("Unable to find Kernel Debug Kit") self.kdk_path = kdk_path logging.info(f"- Found KDK at: {kdk_path}") # Due to some IOHIDFamily oddities, we need to ensure their CodeSignature is retained cs_path = Path(self.mount_location) / Path("System/Library/Extensions/IOHIDFamily.kext/Contents/PlugIns/IOHIDEventDriver.kext/Contents/_CodeSignature") if save_hid_cs is True and cs_path.exists(): logging.info("- Backing up IOHIDEventDriver CodeSignature") # Note it's a folder, not a file utilities.elevated(["/bin/cp", "-r", cs_path, f"{self.constants.payload_path}/IOHIDEventDriver_CodeSignature.bak"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT) logging.info(f"- Merging KDK with Root Volume: {kdk_path.name}") utilities.elevated( # Only merge '/System/Library/Extensions' # 'Kernels' and 'KernelSupport' is wasted space for root patching (we don't care above dev kernels) ["rsync", "-r", "-i", "-a", f"{kdk_path}/System/Library/Extensions/", f"{self.mount_location}/System/Library/Extensions"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT ) # During reversing, we found that kmutil uses this path to determine whether the KDK was successfully merged # Best to verify now before we cause any damage if not (Path(self.mount_location) / Path("System/Library/Extensions/System.kext/PlugIns/Libkern.kext/Libkern")).exists(): logging.info("- Failed to merge KDK with Root Volume") raise Exception("Failed to merge KDK with Root Volume") logging.info("- Successfully merged KDK with Root Volume") # Restore IOHIDEventDriver CodeSignature if save_hid_cs is True and Path(f"{self.constants.payload_path}/IOHIDEventDriver_CodeSignature.bak").exists(): logging.info("- Restoring IOHIDEventDriver CodeSignature") if not cs_path.exists(): logging.info(" - CodeSignature folder missing, creating") utilities.elevated(["/bin/mkdir", "-p", cs_path], stdout=subprocess.PIPE, stderr=subprocess.STDOUT) utilities.elevated(["/bin/cp", "-r", f"{self.constants.payload_path}/IOHIDEventDriver_CodeSignature.bak", cs_path], stdout=subprocess.PIPE, stderr=subprocess.STDOUT) utilities.elevated(["/bin/rm", "-rf", f"{self.constants.payload_path}/IOHIDEventDriver_CodeSignature.bak"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT) def _unpatch_root_vol(self): """ Reverts APFS snapshot and cleans up any changes made to the root and data volume """ if self.constants.detected_os <= os_data.os_data.big_sur or self.root_supports_snapshot is False: logging.info("- OS version does not support snapshotting, skipping revert") logging.info("- Reverting to last signed APFS snapshot") result = utilities.elevated(["bless", "--mount", self.mount_location, "--bootefi", "--last-sealed-snapshot"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT) if result.returncode != 0: logging.info("- Unable to revert root volume patches") logging.info("Reason for unpatch Failure:") logging.info(result.stdout.decode()) logging.info("- Failed to revert snapshot via Apple's 'bless' command") else: self._clean_skylight_plugins() self._delete_nonmetal_enforcement() self._clean_auxiliary_kc() self.constants.root_patcher_succeeded = True logging.info("- Unpatching complete") logging.info("\nPlease reboot the machine for patches to take effect") def _rebuild_root_volume(self) -> bool: """ Rebuilds the Root Volume: - Rebuilds the Kernel Collection - Updates the Preboot Kernel Cache - Rebuilds the dyld Shared Cache - Creates a new APFS Snapshot Returns: bool: True if successful, False if not """ if self._rebuild_kernel_collection() is True: self._update_preboot_kernel_cache() self._rebuild_dyld_shared_cache() if self._create_new_apfs_snapshot() is True: logging.info("- Patching complete") logging.info("\nPlease reboot the machine for patches to take effect") if self.needs_kmutil_exemptions is True: logging.info("Note: Apple will require you to open System Preferences -> Security to allow the new kernel extensions to be loaded") self.constants.root_patcher_succeeded = True return True return False def _rebuild_kernel_collection(self) -> bool: """ Rebuilds the Kernel Collection Supports following KC generation: - Boot/SysKC (11.0+) - AuxKC (11.0+) - PrelinkedKernel (10.15-) Returns: bool: True if successful, False if not """ logging.info("- Rebuilding Kernel Cache (This may take some time)") if self.constants.detected_os > os_data.os_data.catalina: # Base Arguments args = ["kmutil", "install"] if self.skip_root_kmutil_requirement is True: # Only rebuild the Auxiliary Kernel Collection args.append("--new") args.append("aux") args.append("--boot-path") args.append(f"{self.mount_location}/System/Library/KernelCollections/BootKernelExtensions.kc") args.append("--system-path") args.append(f"{self.mount_location}/System/Library/KernelCollections/SystemKernelExtensions.kc") else: # Rebuild Boot, System and Auxiliary Kernel Collections args.append("--volume-root") args.append(self.mount_location) # Build Boot, Sys and Aux KC args.append("--update-all") # If multiple kernels found, only build release KCs args.append("--variant-suffix") args.append("release") if self.constants.detected_os >= os_data.os_data.ventura: # With Ventura, we're required to provide a KDK in some form # to rebuild the Kernel Cache # # However since we already merged the KDK onto root with 'ditto', # We can add '--allow-missing-kdk' to skip parsing the KDK # # This allows us to only delete/overwrite kexts inside of # /System/Library/Extensions and not the entire KDK args.append("--allow-missing-kdk") # 'install' and '--update-all' cannot be used together in Ventura. # kmutil will request the usage of 'create' instead: # Warning: kmutil install's usage of --update-all is deprecated. # Use kmutil create --update-install instead' args[1] = "create" if self.needs_kmutil_exemptions is True: # When installing to '/Library/Extensions', following args skip kext consent # prompt in System Preferences when SIP's disabled logging.info(" (You will get a prompt by System Preferences, ignore for now)") args.append("--no-authentication") args.append("--no-authorization") else: args = ["kextcache", "-i", f"{self.mount_location}/"] result = utilities.elevated(args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT) # kextcache notes: # - kextcache always returns 0, even if it fails # - Check the output for 'KernelCache ID' to see if the cache was successfully rebuilt # kmutil notes: # - will return 71 on failure to build KCs # - will return 31 on 'No binaries or codeless kexts were provided' # - will return -10 if the volume is missing (ie. unmounted by another process) if result.returncode != 0 or (self.constants.detected_os < os_data.os_data.catalina and "KernelCache ID" not in result.stdout.decode()): logging.info("- Unable to build new kernel cache") logging.info(f"\nReason for Patch Failure ({result.returncode}):") logging.info(result.stdout.decode()) logging.info("") logging.info("\nPlease reboot the machine to avoid potential issues rerunning the patcher") return False if self.skip_root_kmutil_requirement is True: # Force rebuild the Auxiliary KC result = utilities.elevated(["/usr/bin/killall", "syspolicyd", "kernelmanagerd"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT) if result.returncode != 0: logging.info("- Unable to remove kernel extension policy files") logging.info(f"\nReason for Patch Failure ({result.returncode}):") logging.info(result.stdout.decode()) logging.info("") logging.info("\nPlease reboot the machine to avoid potential issues rerunning the patcher") return False for file in ["KextPolicy", "KextPolicy-shm", "KextPolicy-wal"]: self._remove_file("/private/var/db/SystemPolicyConfiguration/", file) else: # Install RSRHelper utility to handle desynced KCs sys_patch_helpers.SysPatchHelpers(self.constants).install_rsr_repair_binary() logging.info("- Successfully built new kernel cache") return True def _create_new_apfs_snapshot(self) -> bool: """ Creates a new APFS snapshot of the root volume Returns: bool: True if snapshot was created, False if not """ if self.root_supports_snapshot is True: logging.info("- Creating new APFS snapshot") bless = utilities.elevated( [ "bless", "--folder", f"{self.mount_location}/System/Library/CoreServices", "--bootefi", "--create-snapshot" ], stdout=subprocess.PIPE, stderr=subprocess.STDOUT ) if bless.returncode != 0: logging.info("- Unable to create new snapshot") logging.info("Reason for snapshot failure:") logging.info(bless.stdout.decode()) if "Can't use last-sealed-snapshot or create-snapshot on non system volume" in bless.stdout.decode(): logging.info("- This is an APFS bug with Monterey and newer! Perform a clean installation to ensure your APFS volume is built correctly") return False self._unmount_drive() return True def _unmount_drive(self) -> None: """ Unmount root volume """ if self.root_mount_path: logging.info("- Unmounting Root Volume (Don't worry if this fails)") utilities.elevated(["/usr/sbin/diskutil", "unmount", self.root_mount_path], stdout=subprocess.PIPE).stdout.decode().strip().encode() else: logging.info("- Skipping Root Volume unmount") def _rebuild_dyld_shared_cache(self) -> None: """ Rebuild the dyld shared cache Only required on Mojave and older """ if self.constants.detected_os > os_data.os_data.catalina: return logging.info("- Rebuilding dyld shared cache") utilities.process_status(utilities.elevated(["update_dyld_shared_cache", "-root", f"{self.mount_location}/"])) def _update_preboot_kernel_cache(self) -> None: """ Update the preboot kernel cache Only required on Catalina """ if self.constants.detected_os == os_data.os_data.catalina: logging.info("- Rebuilding preboot kernel cache") utilities.process_status(utilities.elevated(["kcditto"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) def _clean_skylight_plugins(self) -> None: """ Clean non-Metal's SkylightPlugins folder """ if (Path(self.mount_application_support) / Path("SkyLightPlugins/")).exists(): logging.info("- Found SkylightPlugins folder, removing old plugins") utilities.process_status(utilities.elevated(["/bin/rm", "-Rf", f"{self.mount_application_support}/SkyLightPlugins"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) utilities.process_status(utilities.elevated(["/bin/mkdir", f"{self.mount_application_support}/SkyLightPlugins"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) else: logging.info("- Creating SkylightPlugins folder") utilities.process_status(utilities.elevated(["/bin/mkdir", "-p", f"{self.mount_application_support}/SkyLightPlugins/"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) def _delete_nonmetal_enforcement(self) -> None: """ Remove defaults related to forced OpenGL rendering Primarily for development purposes """ for arg in ["useMetal", "useIOP"]: result = subprocess.run(["/usr/bin/defaults", "read", "/Library/Preferences/com.apple.CoreDisplay", arg], stdout=subprocess.PIPE, stderr=subprocess.DEVNULL).stdout.decode("utf-8").strip() if result in ["0", "false", "1", "true"]: logging.info(f"- Removing non-Metal Enforcement Preference: {arg}") utilities.elevated(["/usr/bin/defaults", "delete", "/Library/Preferences/com.apple.CoreDisplay", arg]) def _clean_auxiliary_kc(self) -> None: """ Clean the Auxiliary Kernel Collection Logic: When reverting root volume patches, the AuxKC will still retain the UUID it was built against. Thus when Boot/SysKC are reverted, Aux will break To resolve this, delete all installed kexts in /L*/E* and rebuild the AuxKC We can verify our binaries based off the OpenCore-Legacy-Patcher.plist file """ if self.constants.detected_os < os_data.os_data.big_sur: return logging.info("- Cleaning Auxiliary Kernel Collection") oclp_path = "/System/Library/CoreServices/OpenCore-Legacy-Patcher.plist" if Path(oclp_path).exists(): oclp_plist_data = plistlib.load(Path(oclp_path).open("rb")) for key in oclp_plist_data: if isinstance(oclp_plist_data[key], (bool, int)): continue if "Install" not in oclp_plist_data[key]: continue for location in oclp_plist_data[key]["Install"]: if not location.endswith("Extensions"): continue for file in oclp_plist_data[key]["Install"][location]: if not file.endswith(".kext"): continue self._remove_file("/Library/Extensions", file) # Handle situations where users migrated from older OSes with a lot of garbage in /L*/E* # ex. Nvidia Web Drivers, NetUSB, dosdude1's patches, etc. # Move if file's age is older than October 2021 (year before Ventura) if self.constants.detected_os < os_data.os_data.ventura: return relocation_path = "/Library/Relocated Extensions" if not Path(relocation_path).exists(): utilities.elevated(["/bin/mkdir", relocation_path]) for file in Path("/Library/Extensions").glob("*.kext"): try: if datetime.fromtimestamp(file.stat().st_mtime) < datetime(2021, 10, 1): logging.info(f" - Relocating {file.name} kext to {relocation_path}") if Path(relocation_path) / Path(file.name).exists(): utilities.elevated(["/bin/rm", "-Rf", relocation_path / Path(file.name)]) utilities.elevated(["/bin/mv", file, relocation_path]) except: # Some users have the most cursed /L*/E* folders # ex. Symlinks pointing to symlinks pointing to dead files pass def _write_patchset(self, patchset: dict) -> None: """ Write patchset information to Root Volume Parameters: patchset (dict): Patchset information (generated by GenerateRootPatchSets) """ destination_path = f"{self.mount_location}/System/Library/CoreServices" file_name = "OpenCore-Legacy-Patcher.plist" destination_path_file = f"{destination_path}/{file_name}" if sys_patch_helpers.SysPatchHelpers(self.constants).generate_patchset_plist(patchset, file_name, self.kdk_path): logging.info("- Writing patchset information to Root Volume") if Path(destination_path_file).exists(): utilities.process_status(utilities.elevated(["/bin/rm", destination_path_file], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) utilities.process_status(utilities.elevated(["/bin/cp", f"{self.constants.payload_path}/{file_name}", destination_path], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) def _add_auxkc_support(self, install_file: str, source_folder_path: str, install_patch_directory: str, destination_folder_path: str) -> str: """ Patch provided Kext to support Auxiliary Kernel Collection Logic: In macOS Ventura, KDKs are required to build new Boot and System KCs However for some patch sets, we're able to use the Auxiliary KCs with '/Library/Extensions' kernelmanagerd determines which kext is installed by their 'OSBundleRequired' entry If a kext is labeled as 'OSBundleRequired: Root' or 'OSBundleRequired: Safe Boot', kernelmanagerd will require the kext to be installed in the Boot/SysKC Additionally, kexts starting with 'com.apple.' are not natively allowed to be installed in the AuxKC. So we need to explicitly set our 'OSBundleRequired' to 'Auxiliary' Parameters: install_file (str): Kext file name source_folder_path (str): Source folder path install_patch_directory (str): Patch directory destination_folder_path (str): Destination folder path Returns: str: Updated destination folder path """ if self.skip_root_kmutil_requirement is False: return destination_folder_path if not install_file.endswith(".kext"): return destination_folder_path if install_patch_directory != "/System/Library/Extensions": return destination_folder_path if self.constants.detected_os < os_data.os_data.ventura: return destination_folder_path updated_install_location = str(self.mount_location_data) + "/Library/Extensions" logging.info(f" - Adding AuxKC support to {install_file}") plist_path = Path(Path(source_folder_path) / Path(install_file) / Path("Contents/Info.plist")) plist_data = plistlib.load((plist_path).open("rb")) # Check if we need to update the 'OSBundleRequired' entry if not plist_data["CFBundleIdentifier"].startswith("com.apple."): return updated_install_location if "OSBundleRequired" in plist_data: if plist_data["OSBundleRequired"] == "Auxiliary": return updated_install_location plist_data["OSBundleRequired"] = "Auxiliary" plistlib.dump(plist_data, plist_path.open("wb")) self._check_kexts_needs_authentication(install_file) return updated_install_location def _check_kexts_needs_authentication(self, kext_name: str): """ Verify whether the user needs to authenticate in System Preferences Sets 'needs_to_open_preferences' to True if the kext is not in the AuxKC Logic: Under 'private/var/db/KernelManagement/AuxKC/CurrentAuxKC/com.apple.kcgen.instructions.plist' ["kextsToBuild"][i]: ["bundlePathMainOS"] = /Library/Extensions/Test.kext ["cdHash"] = Bundle's CDHash (random on ad-hoc signed, static on dev signed) ["teamID"] = Team ID (blank on ad-hoc signed) To grab the CDHash of a kext, run 'codesign -dvvv ' Parameters: kext_name (str): Name of the kext to check """ try: aux_cache_path = Path(self.mount_location_data) / Path("/private/var/db/KernelExtensionManagement/AuxKC/CurrentAuxKC/com.apple.kcgen.instructions.plist") if aux_cache_path.exists(): aux_cache_data = plistlib.load((aux_cache_path).open("rb")) for kext in aux_cache_data["kextsToBuild"]: if "bundlePathMainOS" in aux_cache_data["kextsToBuild"][kext]: if aux_cache_data["kextsToBuild"][kext]["bundlePathMainOS"] == f"/Library/Extensions/{kext_name}": return except PermissionError: pass logging.info(f" - {kext_name} requires authentication in System Preferences") self.constants.needs_to_open_preferences = True # Notify in GUI to open System Preferences def _patch_root_vol(self): """ Patch root volume """ logging.info(f"- Running patches for {self.model}") if self.patch_set_dictionary != {}: self._execute_patchset(self.patch_set_dictionary) else: self._execute_patchset(sys_patch_generate.GenerateRootPatchSets(self.computer.real_model, self.constants, self.hardware_details).patchset) if self.constants.wxpython_variant is True and self.constants.detected_os >= os_data.os_data.big_sur: needs_daemon = False if self.constants.detected_os >= os_data.os_data.ventura and self.skip_root_kmutil_requirement is False: needs_daemon = True sys_patch_auto.AutomaticSysPatch(self.constants).install_auto_patcher_launch_agent(kdk_caching_needed=needs_daemon) self._rebuild_root_volume() def _execute_patchset(self, required_patches: dict): """ Executes provided patchset Parameters: required_patches (dict): Patchset to execute (generated by sys_patch_generate.GenerateRootPatchSets) """ source_files_path = str(self.constants.payload_local_binaries_root_path) self._preflight_checks(required_patches, source_files_path) for patch in required_patches: logging.info("- Installing Patchset: " + patch) for method_remove in ["Remove", "Remove Non-Root"]: if method_remove in required_patches[patch]: for remove_patch_directory in required_patches[patch][method_remove]: logging.info("- Remove Files at: " + remove_patch_directory) for remove_patch_file in required_patches[patch][method_remove][remove_patch_directory]: if method_remove == "Remove": destination_folder_path = str(self.mount_location) + remove_patch_directory else: destination_folder_path = str(self.mount_location_data) + remove_patch_directory self._remove_file(destination_folder_path, remove_patch_file) for method_install in ["Install", "Install Non-Root"]: if method_install in required_patches[patch]: for install_patch_directory in list(required_patches[patch][method_install]): logging.info(f"- Handling Installs in: {install_patch_directory}") for install_file in list(required_patches[patch][method_install][install_patch_directory]): source_folder_path = source_files_path + "/" + required_patches[patch][method_install][install_patch_directory][install_file] + install_patch_directory if method_install == "Install": destination_folder_path = str(self.mount_location) + install_patch_directory else: if install_patch_directory == "/Library/Extensions": self.needs_kmutil_exemptions = True self._check_kexts_needs_authentication(install_file) destination_folder_path = str(self.mount_location_data) + install_patch_directory updated_destination_folder_path = self._add_auxkc_support(install_file, source_folder_path, install_patch_directory, destination_folder_path) if destination_folder_path != updated_destination_folder_path: # Update required_patches to reflect the new destination folder path if updated_destination_folder_path not in required_patches[patch][method_install]: required_patches[patch][method_install].update({updated_destination_folder_path: {}}) required_patches[patch][method_install][updated_destination_folder_path].update({install_file: required_patches[patch][method_install][install_patch_directory][install_file]}) required_patches[patch][method_install][install_patch_directory].pop(install_file) destination_folder_path = updated_destination_folder_path self._install_new_file(source_folder_path, destination_folder_path, install_file) if "Processes" in required_patches[patch]: for process in required_patches[patch]["Processes"]: # Some processes need sudo, however we cannot directly call sudo in some scenarios # Instead, call elevated funtion if string's boolean is True if required_patches[patch]["Processes"][process] is True: logging.info(f"- Running Process as Root:\n{process}") utilities.process_status(utilities.elevated(process.split(" "), stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) else: logging.info(f"- Running Process:\n{process}") utilities.process_status(subprocess.run(process, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, shell=True)) if any(x in required_patches for x in ["AMD Legacy GCN", "AMD Legacy Polaris", "AMD Legacy Vega"]): sys_patch_helpers.SysPatchHelpers(self.constants).disable_window_server_caching() if "Metal 3802 Common Extended" in required_patches: sys_patch_helpers.SysPatchHelpers(self.constants).patch_gpu_compiler_libraries(mount_point=self.mount_location) self._write_patchset(required_patches) def _preflight_checks(self, required_patches: dict, source_files_path: Path) -> None: """ Runs preflight checks before patching Parameters: required_patches (dict): Patchset dictionary (from sys_patch_generate.GenerateRootPatchSets) source_files_path (Path): Path to the source files (PatcherSupportPkg) """ logging.info("- Running Preflight Checks before patching") # Make sure old SkyLight plugins aren't being used self._clean_skylight_plugins() # Make sure non-Metal Enforcement preferences are not present self._delete_nonmetal_enforcement() # Make sure we clean old kexts in /L*/E* that are not in the patchset self._clean_auxiliary_kc() # Make sure SNB kexts are compatible with the host if "Intel Sandy Bridge" in required_patches: sys_patch_helpers.SysPatchHelpers(self.constants).snb_board_id_patch(source_files_path) for patch in required_patches: # Check if all files are present for method_type in ["Install", "Install Non-Root"]: if method_type in required_patches[patch]: for install_patch_directory in required_patches[patch][method_type]: for install_file in required_patches[patch][method_type][install_patch_directory]: source_file = source_files_path + "/" + required_patches[patch][method_type][install_patch_directory][install_file] + install_patch_directory + "/" + install_file if not Path(source_file).exists(): raise Exception(f"Failed to find {source_file}") # Ensure KDK is properly installed self._merge_kdk_with_root(save_hid_cs=True if "Legacy USB 1.1" in required_patches else False) logging.info("- Finished Preflight, starting patching") def _install_new_file(self, source_folder: Path, destination_folder: Path, file_name: str) -> None: """ Installs a new file to the destination folder File handling logic: - .frameworks are merged with the destination folder - Other files are deleted and replaced (ex. .kexts, .apps) Parameters: source_folder (Path): Path to the source folder destination_folder (Path): Path to the destination folder file_name (str): Name of the file to install """ file_name_str = str(file_name) if not Path(destination_folder).exists(): logging.info(f" - Skipping {file_name}, cannot locate {source_folder}") return if file_name_str.endswith(".framework"): # merge with rsync logging.info(f" - Installing: {file_name}") utilities.elevated(["rsync", "-r", "-i", "-a", f"{source_folder}/{file_name}", f"{destination_folder}/"], stdout=subprocess.PIPE) self._fix_permissions(destination_folder + "/" + file_name) elif Path(source_folder + "/" + file_name_str).is_dir(): # Applicable for .kext, .app, .plugin, .bundle, all of which are directories if Path(destination_folder + "/" + file_name).exists(): logging.info(f" - Found existing {file_name}, overwriting...") utilities.process_status(utilities.elevated(["/bin/rm", "-R", f"{destination_folder}/{file_name}"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) else: logging.info(f" - Installing: {file_name}") utilities.process_status(utilities.elevated(["/bin/cp", "-R", f"{source_folder}/{file_name}", destination_folder], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) self._fix_permissions(destination_folder + "/" + file_name) else: # Assume it's an individual file, replace as normal if Path(destination_folder + "/" + file_name).exists(): logging.info(f" - Found existing {file_name}, overwriting...") utilities.process_status(utilities.elevated(["/bin/rm", f"{destination_folder}/{file_name}"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) else: logging.info(f" - Installing: {file_name}") utilities.process_status(utilities.elevated(["/bin/cp", f"{source_folder}/{file_name}", destination_folder], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) self._fix_permissions(destination_folder + "/" + file_name) def _remove_file(self, destination_folder: Path, file_name: str) -> None: """ Removes a file from the destination folder Parameters: destination_folder (Path): Path to the destination folder file_name (str): Name of the file to remove """ if Path(destination_folder + "/" + file_name).exists(): logging.info(f" - Removing: {file_name}") if Path(destination_folder + "/" + file_name).is_dir(): utilities.process_status(utilities.elevated(["/bin/rm", "-R", f"{destination_folder}/{file_name}"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) else: utilities.process_status(utilities.elevated(["/bin/rm", f"{destination_folder}/{file_name}"], stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) def _fix_permissions(self, destination_file: Path) -> None: """ Fix file permissions for a given file or directory """ chmod_args = ["chmod", "-Rf", "755", destination_file] chown_args = ["chown", "-Rf", "root:wheel", destination_file] if not Path(destination_file).is_dir(): # Strip recursive arguments chmod_args.pop(1) chown_args.pop(1) utilities.process_status(utilities.elevated(chmod_args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) utilities.process_status(utilities.elevated(chown_args, stdout=subprocess.PIPE, stderr=subprocess.STDOUT)) def _check_files(self) -> bool: """ Check if all files are present (primarily PatcherSupportPkg resources) Returns: bool: True if all files are present, False otherwise """ if Path(self.constants.payload_local_binaries_root_path).exists(): logging.info("- Local PatcherSupportPkg resources available, continuing...") return True if Path(self.constants.payload_local_binaries_root_path_dmg).exists(): logging.info("- Local PatcherSupportPkg resources available, mounting...") output = subprocess.run( [ "/usr/bin/hdiutil", "attach", "-noverify", f"{self.constants.payload_local_binaries_root_path_dmg}", "-mountpoint", Path(self.constants.payload_path / Path("Universal-Binaries")), "-nobrowse", "-shadow", Path(self.constants.payload_path / Path("Universal-Binaries_overlay")), "-passphrase", "password" ], stdout=subprocess.PIPE, stderr=subprocess.STDOUT ) if output.returncode != 0: logging.info("- Failed to mount Universal-Binaries.dmg") logging.info(f"Output: {output.stdout.decode()}") logging.info(f"Return Code: {output.returncode}") return False logging.info("- Mounted Universal-Binaries.dmg") if self.constants.cli_mode is False and Path(self.constants.overlay_psp_path_dmg).exists() and Path("~/.dortania_developer").expanduser().exists(): icon_path = str(self.constants.app_icon_path).replace("/", ":")[1:] msg = "Welcome to the DortaniaInternal Program, please provided the decryption key to access internal resources. Press cancel to skip." password = Path("~/.dortania_developer_key").expanduser().read_text().strip() if Path("~/.dortania_developer_key").expanduser().exists() else "" for i in range(3): try: if password == "": password = applescript.AppleScript( f""" set theResult to display dialog "{msg}" default answer "" with hidden answer with title "OpenCore Legacy Patcher" with icon file "{icon_path}" return the text returned of theResult """ ).run() result = subprocess.run( [ "/usr/bin/hdiutil", "attach", "-noverify", f"{self.constants.overlay_psp_path_dmg}", "-mountpoint", Path(self.constants.payload_path / Path("DortaniaInternal")), "-nobrowse", "-passphrase", password ], stdout=subprocess.PIPE, stderr=subprocess.STDOUT ) if result.returncode == 0: logging.info("- Mounted DortaniaInternal resources") result = subprocess.run( [ "/usr/bin/ditto", f"{self.constants.payload_path / Path('DortaniaInternal')}", f"{self.constants.payload_path / Path('Universal-Binaries')}" ], stdout=subprocess.PIPE, stderr=subprocess.STDOUT ) if result.returncode == 0: return True logging.info("- Failed to merge DortaniaInternal resources") logging.info(f"Output: {result.stdout.decode()}") logging.info(f"Return Code: {result.returncode}") return False logging.info("- Failed to mount DortaniaInternal resources") logging.info(f"Output: {result.stdout.decode()}") logging.info(f"Return Code: {result.returncode}") if "Authentication error" not in result.stdout.decode(): try: # Display that the disk image might be corrupted applescript.AppleScript( f""" display dialog "Failed to mount DortaniaInternal resources, please file an internal radar:\n\n{result.stdout.decode()}" with title "OpenCore Legacy Patcher" with icon file "{icon_path}" """ ).run() return False except Exception as e: pass break msg = f"Decryption failed, please try again. {2 - i} attempts remaining. " password = "" if i == 2: applescript.AppleScript( f""" display dialog "Failed to mount DortaniaInternal resources, too many incorrect passwords. If this continues with the correct decryption key, please file an internal radar." with title "OpenCore Legacy Patcher" with icon file "{icon_path}" """ ).run() return False except Exception as e: break return True logging.info("- PatcherSupportPkg resources missing, Patcher likely corrupted!!!") return False # Entry Function def start_patch(self): """ Entry function for the patching process """ logging.info("- Starting Patch Process") logging.info(f"- Determining Required Patch set for Darwin {self.constants.detected_os}") self.patch_set_dictionary = sys_patch_generate.GenerateRootPatchSets(self.computer.real_model, self.constants, self.hardware_details).patchset if self.patch_set_dictionary == {}: logging.info("- No Root Patches required for your machine!") return logging.info("- Verifying whether Root Patching possible") if sys_patch_detect.DetectRootPatch(self.computer.real_model, self.constants).verify_patch_allowed(print_errors=not self.constants.wxpython_variant) is True: logging.info("- Patcher is capable of patching") if self._check_files(): if self._mount_root_vol() is True: self._patch_root_vol() else: logging.info("- Recommend rebooting the machine and trying to patch again") def start_unpatch(self) -> None: """ Entry function for unpatching the root volume """ logging.info("- Starting Unpatch Process") if sys_patch_detect.DetectRootPatch(self.computer.real_model, self.constants).verify_patch_allowed(print_errors=True) is True: if self._mount_root_vol() is True: self._unpatch_root_vol() else: logging.info("- Recommend rebooting the machine and trying to patch again")