gxp: [Copybara Auto Merge] Merge branch 'zumapro-v' into 'android14-gs-pixel-6.1'

gcip: Makefile supports standalone IIF
Bug: 312161537
gcip: unittests: Makefile supports standalone IIF
Bug: 333862288
gcip: alloc helper: update comments
gcip: pass IP type when submit fence waiters
Bug: 336524176
GCIP_MAIN_REV_ID: 44cb27076738eb87d7d3d235aee7df7a672b8fa6
gxp: remove LPM mapping from core
Bug: 299037074
gxp: direct mode: Request forced debug dump in case of mailbox timeout
Bug: 328551772
gxp: direct mode: Request forced debug dump in case of core crash
Bug: 328551772 (repeat)
gxp: Append fw start and stop logic to support direct mode debug dump.
Bug: 328551772 (repeat)
gxp: Add extra members to vd struct for direct mode
Bug: 328551772 (repeat)
gxp: Add get/set methods for debug dump flags
Bug: 328551772 (repeat)
gxp: Update gxp-host-device-structs.h
Bug: 328551772 (repeat)
gxp: increase gxp version number to 1.27
Bug: 328197763
gxp: Makefile supports standalone IIF
Bug: 332216049
gxp: makefile fix syntax
ganymede: enable gsa support
Bug: 329815240
gxp: Add gxp_thermal_control() implementation
Bug: 328688605
gxp: Add handler for GCIP_KCI_CODE_THERMAL_CONTROL
Bug: 328688605 (repeat)
gxp: pass IP type to fence waited function
Bug: 330929140 (repeat)
gxp: gsx01 add BCI and DSU support
Bug: 336947583
gxp: unittests: Makefile supports standalone IIF
Bug: 333862288 (repeat)
gxp: pass waiter IP type
Bug: 336524176 (repeat)
gxp: remove redundant header inclusion
Bug: 312161537 (repeat)
gcip: add sanitizer_config field in image configs
gcip: image config: support 36-bit mappings
Bug: 325138254
GCIP_MAIN_REV_ID: 673083751510a2e0a60702b1a87d74182ff33355
gxp: Add debugfs nodes for min/max freq limits
Bug: 328688605 (repeat)
gxp: adopt standalone IIF
Bug: 332216049 (repeat)
gxp: reset shadow memory before booting MCU
Bug: 302441258
gxp: support using dynamic memory for MCU fw
Bug: 312642303
gxp: support ns mapping in image configs
gxp: Add KCI handler for setting frequency limits
Bug: 328688605 (repeat)
gxp: add trace points for uci command and respond
Bug: 328197763 (repeat)
gxp: use IOMMU granule instead of PAGE_SIZE
Bug: 333336858
gcip: image config: always pass secure shared mappings to IP driver
Bug: 325139482
gcip: expose gcip_iommu_domain_granule
Bug: 333336858 (repeat)
gcip: fix page size on decoding image cfg
Bug: 333336858 (repeat)
gcip: iif: rename header file to iif-signaler-submission-waiter.h
GCIP_MAIN_REV_ID: 2d5859f668763d7e5a5624e701815b73b593923f
Bug: 330284596
gcip: remove get_cmd_queue_head callback
Bug: 330646127
gcip: make wait_for_cmd_queue_not_full mandatory
Bug: 330646127 (repeat)
gcip: fix context comments in gcip_mailbox_ops
gcip: Remove iif unittests
Bug: 326539224
gcip: guard dma_resv_[un]lock with LINUX_VERSION_CODE
Bug: 329016610
gcip: introduce gcip_fence_array_merge_ikf
Bug: 328046412
gcip: image config parse iova map flags from firmware
Bug: 324254320
Bug: 325139482 (repeat)
gcip: iommu add flag for MMIO device memory mappings
Bug: 324254320 (repeat)
gcip: image config add mapping flags in IOVA
Bug: 324254320 (repeat)
Bug: 325139482 (repeat)
gcip: unittests: remove Kconfig and update Makefile
Bug: 322448134
gcip: guard dma_buf map attachment with dma_resv
Bug: 328199067
gcip: add gcip_dma_fence_merge_fences
Bug: 320401031
iif: add lock_class_key to submitted_signalers_lock
Bug: 327075979
Merge "Merge branch 'headers' into HEAD" into main
GCIP_MAIN_REV_ID: f2f385cce0d35680ebe7042cc5a75dffaa27b801

Signed-off-by: Aurora zuma pro automerger <[email protected]>
GitOrigin-RevId: b27296515aa3cfdc9a29c412b7e4c65ba59a4f89
Change-Id: Id9b66ffe91c548bef14e1eb31d3814299b5483be
diff --git a/Kconfig b/Kconfig
index 174b3e4..f8c7358 100644
--- a/Kconfig
+++ b/Kconfig
@@ -25,5 +25,4 @@
 	  To compile this driver as a module, choose M here. The module will be
 	  called "gxp".
 
-
 endmenu
diff --git a/Makefile b/Makefile
index e8be6aa..717f11a 100644
--- a/Makefile
+++ b/Makefile
@@ -6,6 +6,7 @@
 GXP_CHIP := CALLISTO
 CONFIG_$(GXP_CHIP) ?= m
 GCIP_DIR := gcip-kernel-driver/drivers/gcip
+CURRENT_DIR := $(dir $(abspath $(lastword $(MAKEFILE_LIST))))
 
 obj-$(CONFIG_$(GXP_CHIP)) += gxp.o
 
@@ -13,6 +14,7 @@
 		gxp-bpm.o \
 		gxp-client.o \
 		gxp-core-telemetry.o \
+		gxp-dci.o \
 		gxp-debug-dump.o \
 		gxp-dma-fence.o \
 		gxp-dma-iommu.o \
@@ -34,7 +36,6 @@
 		gxp-vd.o
 
 gxp-mcu-objs := \
-		gxp-dci.o \
 		gxp-kci.o \
 		gxp-mcu-firmware.o \
 		gxp-mcu-fs.o \
@@ -93,11 +94,12 @@
 GXP_PLATFORM ?= SILICON
 
 gxp-flags := -DCONFIG_GXP_$(GXP_PLATFORM) -DCONFIG_$(GXP_CHIP)=1 \
-	     -I$(M)/include -I$(M)/gcip-kernel-driver/include \
-	     -I$(srctree)/$(M)/include \
-	     -I$(srctree)/$(M)/gcip-kernel-driver/include \
-	     -I$(srctree)/drivers/gxp/include \
+	     -I$(CURRENT_DIR)/include -I$(CURRENT_DIR)/gcip-kernel-driver/include \
 	     -I$(KERNEL_SRC)/../private/google-modules/power/mitigation
+# TODO(b/336717718): Remove path of embedded IIF
+gxp-flags += -I$(CURRENT_DIR)/gcip-kernel-driver/drivers/gcip/iif/include \
+	     -I$(KERNEL_SRC)/../private/google-modules/iif/include
+
 ccflags-y += $(EXTRA_CFLAGS) $(gxp-flags)
 # Flags needed for external modules.
 ccflags-y += -DCONFIG_GOOGLE_BCL
@@ -107,11 +109,18 @@
 ifneq ($(OUT_DIR),)
 # Access TPU driver's exported symbols.
 EXTRA_SYMBOLS += $(GMODULE_PATH)/edgetpu/$(EDGETPU_CHIP)/drivers/edgetpu/Module.symvers
+ifneq ($(wildcard $(GMODULE_PATH)/soc/gs/drivers/soc/google/gsa/Module.symvers),)
+EXTRA_SYMBOLS += $(GMODULE_PATH)/soc/gs/drivers/soc/google/gsa/Module.symvers
+endif
 
 ifneq ($(GXP_POWER_MITIGATION), false)
 EXTRA_SYMBOLS += $(GMODULE_PATH)/power/mitigation/Module.symvers
 endif
+
+ifneq ($(wildcard $(GMODULE_PATH)/iif/Module.symvers),)
+EXTRA_SYMBOLS += $(GMODULE_PATH)/iif/Module.symvers
 endif
+endif # OUT_DIR
 
 modules modules_install:
 	$(MAKE) -C $(KERNEL_SRC) M=$(M)/$(GCIP_DIR) gcip.o
diff --git a/callisto/iova.h b/callisto/iova.h
index c09637b..9dd83c3 100644
--- a/callisto/iova.h
+++ b/callisto/iova.h
@@ -27,5 +27,6 @@
 #define GXP_IREMAP_DATA_SIZE 0x1FF000 /* (2MB - 4KB) */
 /* Last 4K page of IREMAP_DATA */
 #define GXP_MCU_BOOT_MODE_OFFSET (GXP_IREMAP_DATA_OFFSET + GXP_IREMAP_DATA_SIZE)
+#define GXP_IREMAP_DYNAMIC_CODE_BASE 0x11000000
 
 #endif /* __CALLISTO_IOVA_H__ */
diff --git a/gcip-kernel-driver/drivers/gcip/Makefile b/gcip-kernel-driver/drivers/gcip/Makefile
index 6971e51..a57f5ec 100644
--- a/gcip-kernel-driver/drivers/gcip/Makefile
+++ b/gcip-kernel-driver/drivers/gcip/Makefile
@@ -32,11 +32,15 @@
 			   iif/iif-signaler-submission-waiter.o \
 			   iif/iif-sync-file.o
 
-CURRENT_DIR=$(dir $(abspath $(lastword $(MAKEFILE_LIST))))
+CURRENT_DIR := $(dir $(abspath $(lastword $(MAKEFILE_LIST))))
 
-ccflags-y += -I$(CURRENT_DIR)/../../include
+# TODO(b/336717718): Remove path of embedded IIF
+ccflags-y	+= -I$(CURRENT_DIR)/../../include \
+		   -I$(CURRENT_DIR)/iif/include \
+		   -I$(KERNEL_SRC)/../private/google-modules/iif/include
 
 
 
 modules modules_install clean:
-	$(MAKE) -C $(KERNEL_SRC) M=$(M) W=1 $(KBUILD_OPTIONS) $(@)
+	$(MAKE) -C $(KERNEL_SRC) M=$(M) W=1 $(KBUILD_OPTIONS) \
+	KBUILD_EXTRA_SYMBOLS="$(EXTRA_SYMBOLS)" $(@)
diff --git a/gcip-kernel-driver/drivers/gcip/gcip-dma-fence.c b/gcip-kernel-driver/drivers/gcip/gcip-dma-fence.c
index 8053f77..b27fc40 100644
--- a/gcip-kernel-driver/drivers/gcip/gcip-dma-fence.c
+++ b/gcip-kernel-driver/drivers/gcip/gcip-dma-fence.c
@@ -207,12 +207,32 @@
 	spin_unlock_irq(&gfence->lock);
 }
 
+struct dma_fence *gcip_dma_fence_merge_fences(int num_fences, struct dma_fence **fences)
+{
+	struct dma_fence *tmp;
+	struct dma_fence *array;
+	int i;
+
+	array = dma_fence_unwrap_merge(fences[0]);
+	if (!array)
+		return ERR_PTR(-ENOMEM);
+
+	for (i = 1; i < num_fences; i++) {
+		tmp = array;
+		array = dma_fence_unwrap_merge(tmp, fences[i]);
+		dma_fence_put(tmp);
+		if (!array)
+			return ERR_PTR(-ENOMEM);
+	}
+
+	return array;
+}
+
 struct dma_fence *gcip_dma_fence_merge_fds(int num_fences, int *fence_fds)
 {
 	struct dma_fence **fences;
-	struct dma_fence *tmp;
 	struct dma_fence *result;
-	int i = 0;
+	int i;
 
 	if (!num_fences)
 		return ERR_PTR(-EINVAL);
@@ -229,20 +249,7 @@
 		}
 	}
 
-	result = dma_fence_unwrap_merge(fences[0]);
-	if (!result) {
-		result = ERR_PTR(-ENOMEM);
-		goto out;
-	}
-	for (i = 1; i < num_fences; i++) {
-		tmp = result;
-		result = dma_fence_unwrap_merge(tmp, fences[i]);
-		dma_fence_put(tmp);
-		if (!result) {
-			result = ERR_PTR(-ENOMEM);
-			goto out;
-		}
-	}
+	result = gcip_dma_fence_merge_fences(num_fences, fences);
 
 out:
 	for (i = 0; i < num_fences; i++)
diff --git a/gcip-kernel-driver/drivers/gcip/gcip-fence-array.c b/gcip-kernel-driver/drivers/gcip/gcip-fence-array.c
index 5ef7546..567e3ac 100644
--- a/gcip-kernel-driver/drivers/gcip/gcip-fence-array.c
+++ b/gcip-kernel-driver/drivers/gcip/gcip-fence-array.c
@@ -5,9 +5,11 @@
  * Copyright (C) 2023 Google LLC
  */
 
+#include <linux/dma-fence.h>
 #include <linux/kref.h>
 #include <linux/slab.h>
 
+#include <gcip/gcip-dma-fence.h>
 #include <gcip/gcip-fence-array.h>
 #include <gcip/gcip-fence.h>
 
@@ -135,7 +137,7 @@
 		gcip_fence_signal(fence_array->fences[i], errno);
 }
 
-void gcip_fence_array_waited(struct gcip_fence_array *fence_array)
+void gcip_fence_array_waited(struct gcip_fence_array *fence_array, enum iif_ip_type ip)
 {
 	int i;
 
@@ -143,7 +145,7 @@
 		return;
 
 	for (i = 0; i < fence_array->size; i++)
-		gcip_fence_waited(fence_array->fences[i]);
+		gcip_fence_waited(fence_array->fences[i], ip);
 }
 
 void gcip_fence_array_submit_signaler(struct gcip_fence_array *fence_array)
@@ -157,7 +159,7 @@
 		gcip_fence_submit_signaler(fence_array->fences[i]);
 }
 
-void gcip_fence_array_submit_waiter(struct gcip_fence_array *fence_array)
+void gcip_fence_array_submit_waiter(struct gcip_fence_array *fence_array, enum iif_ip_type ip)
 {
 	int i;
 
@@ -165,11 +167,12 @@
 		return;
 
 	for (i = 0; i < fence_array->size; i++)
-		gcip_fence_submit_waiter(fence_array->fences[i]);
+		gcip_fence_submit_waiter(fence_array->fences[i], ip);
 }
 
 int gcip_fence_array_submit_waiter_and_signaler(struct gcip_fence_array *in_fences,
-						struct gcip_fence_array *out_fences)
+						struct gcip_fence_array *out_fences,
+						enum iif_ip_type ip)
 {
 	int i;
 
@@ -207,7 +210,7 @@
 
 	/* Submits a waiter to @in_fences. */
 	for (i = 0; in_fences && i < in_fences->size; i++)
-		gcip_fence_submit_waiter(in_fences->fences[i]);
+		gcip_fence_submit_waiter(in_fences->fences[i], ip);
 
 	return 0;
 }
@@ -256,3 +259,30 @@
 	return gcip_fence_wait_signaler_submission(fence_array->fences, fence_array->size, eventfd,
 						   remaining_signalers);
 }
+
+struct dma_fence *gcip_fence_array_merge_ikf(struct gcip_fence_array *fence_array)
+{
+	struct dma_fence **fences;
+	struct dma_fence *merged;
+	int i;
+
+	if (!fence_array || !fence_array->size || !fence_array->same_type ||
+	    fence_array->type != GCIP_IN_KERNEL_FENCE)
+		return NULL;
+
+	if (fence_array->size == 1)
+		return dma_fence_get(fence_array->fences[0]->fence.ikf);
+
+	fences = kcalloc(fence_array->size, sizeof(*fences), GFP_KERNEL);
+	if (!fences)
+		return ERR_PTR(-ENOMEM);
+
+	for (i = 0; i < fence_array->size; i++)
+		fences[i] = fence_array->fences[i]->fence.ikf;
+
+	merged = gcip_dma_fence_merge_fences(fence_array->size, fences);
+
+	kfree(fences);
+
+	return merged;
+}
diff --git a/gcip-kernel-driver/drivers/gcip/gcip-fence.c b/gcip-kernel-driver/drivers/gcip/gcip-fence.c
index 8548818..1538b80 100644
--- a/gcip-kernel-driver/drivers/gcip/gcip-fence.c
+++ b/gcip-kernel-driver/drivers/gcip/gcip-fence.c
@@ -11,9 +11,9 @@
 
 #include <gcip/gcip-dma-fence.h>
 #include <gcip/gcip-fence.h>
-#include <gcip/iif/iif-fence.h>
-#include <gcip/iif/iif-signaler-submission-watier.h>
-#include <gcip/iif/iif.h>
+#include <iif/iif-fence.h>
+#include <iif/iif-signaler-submission-waiter.h>
+#include <iif/iif.h>
 
 static struct gcip_fence *gcip_fence_alloc(enum gcip_fence_type type)
 {
@@ -175,10 +175,10 @@
 	return -EOPNOTSUPP;
 }
 
-int gcip_fence_submit_waiter(struct gcip_fence *fence)
+int gcip_fence_submit_waiter(struct gcip_fence *fence, enum iif_ip_type ip)
 {
 	if (fence->type == GCIP_INTER_IP_FENCE)
-		return iif_fence_submit_waiter(fence->fence.iif, IIF_IP_DSP);
+		return iif_fence_submit_waiter(fence->fence.iif, ip);
 	return -EOPNOTSUPP;
 }
 
@@ -196,10 +196,10 @@
 	}
 }
 
-void gcip_fence_waited(struct gcip_fence *fence)
+void gcip_fence_waited(struct gcip_fence *fence, enum iif_ip_type ip)
 {
 	if (fence->type == GCIP_INTER_IP_FENCE)
-		iif_fence_waited(fence->fence.iif);
+		iif_fence_waited(fence->fence.iif, ip);
 }
 
 /*
diff --git a/gcip-kernel-driver/drivers/gcip/gcip-image-config.c b/gcip-kernel-driver/drivers/gcip/gcip-image-config.c
index c345505..62cb6a8 100644
--- a/gcip-kernel-driver/drivers/gcip/gcip-image-config.c
+++ b/gcip-kernel-driver/drivers/gcip/gcip-image-config.c
@@ -12,6 +12,26 @@
 
 #include <gcip/gcip-image-config.h>
 
+/*
+ * Return true for a secure config mapping that is not shared.  The host IP driver doesn't need to
+ * process such mappings.
+ */
+static bool skip_secure_mapping(struct gcip_image_config *config, u32 map_flags)
+{
+	return gcip_image_config_is_secure(config) && !GCIP_IMAGE_CONFIG_MAP_SHARED(map_flags);
+}
+
+static dma_addr_t virt_address_to_dma(u32 virt_address)
+{
+	dma_addr_t daddr = virt_address & ~(GCIP_IMG_CFG_MAP_FLAGS_MASK);
+	u32 flags = virt_address & GCIP_IMG_CFG_MAP_FLAGS_MASK;
+
+	if (GCIP_IMAGE_CONFIG_MAP_36BIT(flags))
+		daddr <<= 4;
+
+	return daddr;
+}
+
 static int setup_iommu_mappings(struct gcip_image_config_parser *parser,
 				struct gcip_image_config *config)
 {
@@ -19,9 +39,13 @@
 	dma_addr_t daddr;
 	size_t size;
 	phys_addr_t paddr;
+	u32 map_flags;
 
 	for (i = 0; i < config->num_iommu_mappings; i++) {
-		daddr = config->iommu_mappings[i].virt_address;
+		daddr = virt_address_to_dma(config->iommu_mappings[i].virt_address);
+		map_flags = config->iommu_mappings[i].virt_address & GCIP_IMG_CFG_MAP_FLAGS_MASK;
+		if (skip_secure_mapping(config, map_flags))
+			continue;
 		if (unlikely(!daddr)) {
 			dev_warn(parser->dev, "Invalid config, device address is zero");
 			ret = -EIO;
@@ -37,7 +61,7 @@
 			ret = -EOVERFLOW;
 			goto err;
 		}
-		ret = parser->ops->map(parser->data, daddr, paddr, size,
+		ret = parser->ops->map(parser->data, daddr, paddr, size, map_flags,
 				       GCIP_IMAGE_CONFIG_FLAGS_SECURE);
 		if (ret) {
 			dev_err(parser->dev,
@@ -51,9 +75,13 @@
 
 err:
 	while (i--) {
-		daddr = config->iommu_mappings[i].virt_address;
+		daddr = config->iommu_mappings[i].virt_address & ~(GCIP_IMG_CFG_MAP_FLAGS_MASK);
+		map_flags = config->iommu_mappings[i].virt_address & GCIP_IMG_CFG_MAP_FLAGS_MASK;
+		if (skip_secure_mapping(config, map_flags))
+			continue;
 		size = gcip_config_to_size(config->iommu_mappings[i].image_config_value);
-		parser->ops->unmap(parser->data, daddr, size, GCIP_IMAGE_CONFIG_FLAGS_SECURE);
+		parser->ops->unmap(parser->data, daddr, size, map_flags,
+				   GCIP_IMAGE_CONFIG_FLAGS_SECURE);
 	}
 	return ret;
 }
@@ -63,14 +91,19 @@
 {
 	dma_addr_t daddr;
 	size_t size;
+	u32 map_flags;
 	int i;
 
 	for (i = config->num_iommu_mappings - 1; i >= 0; i--) {
-		daddr = config->iommu_mappings[i].virt_address;
+		daddr = virt_address_to_dma(config->iommu_mappings[i].virt_address);
+		map_flags = config->iommu_mappings[i].virt_address & GCIP_IMG_CFG_MAP_FLAGS_MASK;
+		if (skip_secure_mapping(config, map_flags))
+			continue;
 		size = gcip_config_to_size(config->iommu_mappings[i].image_config_value);
 		dev_dbg(parser->dev, "Image config removing IOMMU mapping: %pad size=%#lx", &daddr,
 			size);
-		parser->ops->unmap(parser->data, daddr, size, GCIP_IMAGE_CONFIG_FLAGS_SECURE);
+		parser->ops->unmap(parser->data, daddr, size, map_flags,
+				   GCIP_IMAGE_CONFIG_FLAGS_SECURE);
 	}
 }
 
@@ -96,7 +129,7 @@
 			ret = -EOVERFLOW;
 			goto err;
 		}
-		ret = parser->ops->map(parser->data, daddr, paddr, size, 0);
+		ret = parser->ops->map(parser->data, daddr, paddr, size, 0, 0);
 		if (ret)
 			goto err;
 		paddr += size;
@@ -108,7 +141,7 @@
 	while (i--) {
 		size = gcip_ns_config_to_size(config->ns_iommu_mappings[i]);
 		daddr = config->ns_iommu_mappings[i] & GCIP_IMG_CFG_ADDR_MASK;
-		parser->ops->unmap(parser->data, daddr, size, 0);
+		parser->ops->unmap(parser->data, daddr, size, 0, 0);
 	}
 	return ret;
 }
@@ -125,7 +158,7 @@
 		daddr = config->ns_iommu_mappings[i] & GCIP_IMG_CFG_ADDR_MASK;
 		dev_dbg(parser->dev, "Image config removing NS IOMMU mapping: %pad size=%#lx",
 			&daddr, size);
-		parser->ops->unmap(parser->data, daddr, size, 0);
+		parser->ops->unmap(parser->data, daddr, size, 0, 0);
 	}
 }
 
@@ -136,19 +169,16 @@
 
 	if (ret)
 		return ret;
-	if (gcip_image_config_is_ns(config)) {
-		ret = setup_iommu_mappings(parser, config);
-		if (ret)
-			clear_ns_iommu_mappings(parser, config);
-	}
+	ret = setup_iommu_mappings(parser, config);
+	if (ret)
+		clear_ns_iommu_mappings(parser, config);
 	return ret;
 }
 
 static void unmap_image_config(struct gcip_image_config_parser *parser,
 			       struct gcip_image_config *config)
 {
-	if (gcip_image_config_is_ns(config))
-		clear_iommu_mappings(parser, config);
+	clear_iommu_mappings(parser, config);
 	clear_ns_iommu_mappings(parser, config);
 }
 
diff --git a/gcip-kernel-driver/drivers/gcip/gcip-iommu.c b/gcip-kernel-driver/drivers/gcip/gcip-iommu.c
index 3b893d6..b285489 100644
--- a/gcip-kernel-driver/drivers/gcip/gcip-iommu.c
+++ b/gcip-kernel-driver/drivers/gcip/gcip-iommu.c
@@ -10,6 +10,7 @@
 #include <linux/dma-buf.h>
 #include <linux/dma-direction.h>
 #include <linux/dma-mapping.h>
+#include <linux/dma-resv.h>
 #include <linux/genalloc.h>
 #include <linux/iova.h>
 #include <linux/limits.h>
@@ -38,6 +39,7 @@
 #define GCIP_MAP_MASK_DMA_COHERENT GCIP_MAP_MASK(DMA_COHERENT)
 #define GCIP_MAP_MASK_DMA_ATTR GCIP_MAP_MASK(DMA_ATTR)
 #define GCIP_MAP_MASK_RESTRICT_IOVA GCIP_MAP_MASK(RESTRICT_IOVA)
+#define GCIP_MAP_MASK_MMIO GCIP_MAP_MASK(MMIO)
 
 #define GCIP_MAP_FLAGS_GET_VALUE(ATTR, flags) \
 	(((flags) & GCIP_MAP_MASK(ATTR)) >> (GCIP_MAP_FLAGS_##ATTR##_OFFSET))
@@ -45,6 +47,7 @@
 #define GCIP_MAP_FLAGS_GET_DMA_COHERENT(flags) GCIP_MAP_FLAGS_GET_VALUE(DMA_COHERENT, flags)
 #define GCIP_MAP_FLAGS_GET_DMA_ATTR(flags) GCIP_MAP_FLAGS_GET_VALUE(DMA_ATTR, flags)
 #define GCIP_MAP_FLAGS_GET_RESTRICT_IOVA(flags) GCIP_MAP_FLAGS_GET_VALUE(RESTRICT_IOVA, flags)
+#define GCIP_MAP_FLAGS_GET_MMIO(flags) GCIP_MAP_FLAGS_GET_VALUE(MMIO, flags)
 
 /* Restricted IOVA ceiling is for components with 32-bit DMA windows */
 #define GCIP_RESTRICT_IOVA_CEILING UINT_MAX
@@ -94,13 +97,6 @@
 	}
 }
 
-static inline size_t gcip_iommu_domain_granule(struct gcip_iommu_domain *domain)
-{
-	if (unlikely(domain->default_domain))
-		return PAGE_SIZE;
-	return domain->domain_pool->granule;
-}
-
 /*
  * Allocates an IOVA for the scatterlist and maps it to @domain.
  *
@@ -497,8 +493,16 @@
 				  mapping->gcip_map_flags, false);
 	}
 
+#if LINUX_VERSION_CODE >= KERNEL_VERSION(6, 2, 0)
+	dma_resv_lock(dmabuf_mapping->dma_buf->resv, NULL);
 	dma_buf_unmap_attachment(dmabuf_mapping->dma_buf_attachment, dmabuf_mapping->sgt_default,
 				 mapping->dir);
+	dma_resv_unlock(dmabuf_mapping->dma_buf->resv);
+#else
+	dma_buf_unmap_attachment(dmabuf_mapping->dma_buf_attachment, dmabuf_mapping->sgt_default,
+				 mapping->dir);
+#endif
+
 	dma_buf_detach(dmabuf_mapping->dma_buf, dmabuf_mapping->dma_buf_attachment);
 	dma_buf_put(dmabuf_mapping->dma_buf);
 	kfree(dmabuf_mapping);
@@ -1293,7 +1297,13 @@
 #endif
 
 	/* Map the attachment into the default domain. */
+#if LINUX_VERSION_CODE >= KERNEL_VERSION(6, 2, 0)
+	dma_resv_lock(dmabuf->resv, NULL);
 	sgt = dma_buf_map_attachment(attachment, dir);
+	dma_resv_unlock(dmabuf->resv);
+#else
+	sgt = dma_buf_map_attachment(attachment, dir);
+#endif
 	if (IS_ERR(sgt)) {
 		ret = PTR_ERR(sgt);
 		dev_err(dev, "Failed to get sgt from attachment (ret=%d, name=%s, size=%lu)\n", ret,
@@ -1311,7 +1321,13 @@
 	return mapping;
 
 err_map_dma_buf_sgt:
+#if LINUX_VERSION_CODE >= KERNEL_VERSION(6, 2, 0)
+	dma_resv_lock(dmabuf->resv, NULL);
 	dma_buf_unmap_attachment(attachment, sgt, dir);
+	dma_resv_unlock(dmabuf->resv);
+#else
+	dma_buf_unmap_attachment(attachment, sgt, dir);
+#endif
 err_map_attachment:
 	dma_buf_detach(dmabuf, attachment);
 	return ERR_PTR(ret);
@@ -1363,9 +1379,13 @@
 {
 	enum dma_data_direction dir = GCIP_MAP_FLAGS_GET_DMA_DIRECTION(gcip_map_flags);
 	bool coherent = GCIP_MAP_FLAGS_GET_DMA_COHERENT(gcip_map_flags);
+	bool mmio = GCIP_MAP_FLAGS_GET_MMIO(gcip_map_flags);
 	unsigned long attrs = GCIP_MAP_FLAGS_GET_DMA_ATTR(gcip_map_flags);
 	int prot = dma_info_to_prot(dir, coherent, attrs);
 
+	if (mmio)
+		prot |= IOMMU_MMIO;
+
 #if GCIP_IOMMU_MAP_HAS_GFP
 	return iommu_map(domain->domain, iova, paddr, size, prot, GFP_KERNEL);
 #else
diff --git a/gcip-kernel-driver/drivers/gcip/gcip-kci.c b/gcip-kernel-driver/drivers/gcip/gcip-kci.c
index f4e904b..40da93c 100644
--- a/gcip-kernel-driver/drivers/gcip/gcip-kci.c
+++ b/gcip-kernel-driver/drivers/gcip/gcip-kci.c
@@ -14,13 +14,6 @@
 #include <gcip/gcip-kci.h>
 #include <gcip/gcip-mailbox.h>
 
-static u32 gcip_kci_get_cmd_queue_head(struct gcip_mailbox *mailbox)
-{
-	struct gcip_kci *kci = gcip_mailbox_get_data(mailbox);
-
-	return kci->ops->get_cmd_queue_head(kci);
-}
-
 static u32 gcip_kci_get_cmd_queue_tail(struct gcip_mailbox *mailbox)
 {
 	struct gcip_kci *kci = gcip_mailbox_get_data(mailbox);
@@ -255,7 +248,6 @@
 }
 
 static const struct gcip_mailbox_ops gcip_mailbox_ops = {
-	.get_cmd_queue_head = gcip_kci_get_cmd_queue_head,
 	.get_cmd_queue_tail = gcip_kci_get_cmd_queue_tail,
 	.inc_cmd_queue_tail = gcip_kci_inc_cmd_queue_tail,
 	.acquire_cmd_queue_lock = gcip_kci_acquire_cmd_queue_lock,
diff --git a/gcip-kernel-driver/drivers/gcip/gcip-mailbox.c b/gcip-kernel-driver/drivers/gcip/gcip-mailbox.c
index 24849c9..a4b6b7d 100644
--- a/gcip-kernel-driver/drivers/gcip/gcip-mailbox.c
+++ b/gcip-kernel-driver/drivers/gcip/gcip-mailbox.c
@@ -21,7 +21,6 @@
 #define TEST_TRIGGER_TIMEOUT_RACE(...)
 #endif
 
-#define GET_CMD_QUEUE_HEAD() mailbox->ops->get_cmd_queue_head(mailbox)
 #define GET_CMD_QUEUE_TAIL() mailbox->ops->get_cmd_queue_tail(mailbox)
 #define INC_CMD_QUEUE_TAIL(inc) mailbox->ops->inc_cmd_queue_tail(mailbox, inc)
 #define ACQUIRE_CMD_QUEUE_LOCK(try, atomic)                                                        \
@@ -162,33 +161,16 @@
 				    gcip_mailbox_cmd_flags_t flags)
 {
 	int ret = 0;
-	u32 tail;
 	bool atomic = false;
 
 	ACQUIRE_CMD_QUEUE_LOCK(false, &atomic);
 
 	if (!(flags & GCIP_MAILBOX_CMD_FLAGS_SKIP_ASSIGN_SEQ))
 		SET_CMD_ELEM_SEQ(cmd, mailbox->cur_seq);
-	/*
-	 * The lock ensures mailbox cmd_queue_tail cannot be changed by other processes (this
-	 * method should be the only one to modify the value of tail), therefore we can remember
-	 * its value here and use it in the condition of wait_event() call.
-	 */
-	tail = GET_CMD_QUEUE_TAIL();
-
-	if (mailbox->ops->wait_for_cmd_queue_not_full) {
-		/* Wait until the cmd queue has a space for putting cmd. */
-		ret = mailbox->ops->wait_for_cmd_queue_not_full(mailbox);
-		if (ret)
-			goto out;
-	} else if (GET_CMD_QUEUE_HEAD() == (tail ^ mailbox->queue_wrap_bit)) {
-		/*
-		 * Default logic of checking the fullness of cmd_queue. If the cmd_queue is full,
-		 * it's up to the caller to retry.
-		 */
-		ret = -EAGAIN;
+	/* Wait until the cmd queue has a space for putting cmd. */
+	ret = mailbox->ops->wait_for_cmd_queue_not_full(mailbox);
+	if (ret)
 		goto out;
-	}
 
 	if (async_resp->resp) {
 		/* Adds @resp to the wait_list only if the cmd can be pushed successfully. */
@@ -198,9 +180,11 @@
 		if (ret)
 			goto out;
 	}
+
 	/* Size of cmd_queue is a multiple of mailbox->cmd_elem_size. */
-	memcpy(mailbox->cmd_queue + mailbox->cmd_elem_size *
-					    CIRC_QUEUE_REAL_INDEX(tail, mailbox->queue_wrap_bit),
+	memcpy(mailbox->cmd_queue +
+		       mailbox->cmd_elem_size *
+			       CIRC_QUEUE_REAL_INDEX(GET_CMD_QUEUE_TAIL(), mailbox->queue_wrap_bit),
 	       cmd, mailbox->cmd_elem_size);
 	INC_CMD_QUEUE_TAIL(1);
 	if (mailbox->ops->after_enqueue_cmd) {
@@ -492,9 +476,9 @@
 		return 0;
 	}
 
-	if (!ops->get_cmd_queue_head || !ops->get_cmd_queue_tail || !ops->inc_cmd_queue_tail ||
-	    !ops->acquire_cmd_queue_lock || !ops->release_cmd_queue_lock ||
-	    !ops->get_cmd_elem_seq || !ops->set_cmd_elem_seq || !ops->get_cmd_elem_code) {
+	if (!ops->get_cmd_queue_tail || !ops->inc_cmd_queue_tail || !ops->acquire_cmd_queue_lock ||
+	    !ops->release_cmd_queue_lock || !ops->get_cmd_elem_seq || !ops->set_cmd_elem_seq ||
+	    !ops->get_cmd_elem_code || !ops->wait_for_cmd_queue_not_full) {
 		dev_err(mailbox->dev, "Incomplete mailbox CMD queue ops.\n");
 		return -EINVAL;
 	}
diff --git a/gcip-kernel-driver/drivers/gcip/gcip-usage-stats.c b/gcip-kernel-driver/drivers/gcip/gcip-usage-stats.c
index 5c50705..90c0468 100644
--- a/gcip-kernel-driver/drivers/gcip/gcip-usage-stats.c
+++ b/gcip-kernel-driver/drivers/gcip/gcip-usage-stats.c
@@ -721,7 +721,7 @@
 		mutex_unlock(&ustats->dvfs_freqs_lock);
 		dvfs_freqs_num = ustats->ops->get_default_dvfs_freqs_num(ustats->data);
 		for (i = 0; i < dvfs_freqs_num; i++)
-			written += scnprintf(buf + written, PAGE_SIZE - written, "%.*s%d", i, " ",
+			written += scnprintf(buf + written, PAGE_SIZE - written, "%.*s%u", i, " ",
 					     ustats->ops->get_default_dvfs_freq(i, ustats->data));
 	} else {
 		dvfs_freqs_num = ustats->dvfs_freqs_num;
diff --git a/gcip-kernel-driver/drivers/gcip/iif/iif-fence-table.c b/gcip-kernel-driver/drivers/gcip/iif/iif-fence-table.c
index dcea5df..8af7ccb 100644
--- a/gcip-kernel-driver/drivers/gcip/iif/iif-fence-table.c
+++ b/gcip-kernel-driver/drivers/gcip/iif/iif-fence-table.c
@@ -1,8 +1,15 @@
 // SPDX-License-Identifier: GPL-2.0-only
 /*
- * GCIP-integrated IIF driver fence table.
+ * Interface to utilize IIF fence tables, the wait table and the signal table. Both tables will have
+ * one entry per fence ID.
  *
- * Copyright (C) 2023 Google LLC
+ * - Wait table: Describes which IPs are waiting on each fence. This table will be written by the
+ *               kernel driver only.
+ *
+ * - Signal table: Describes how many signals are remaining to unblock each fence. This table will
+ *                 be initialized by the kernel driver and each signaler IP will update it.
+ *
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #define pr_fmt(fmt) "iif: " fmt
@@ -12,7 +19,7 @@
 #include <linux/of.h>
 #include <linux/slab.h>
 
-#include <gcip/iif/iif-fence-table.h>
+#include <iif/iif-fence-table.h>
 
 #define IIF_FENCE_WAIT_TABLE_PROP_NAME "iif-fence-wait-table-region"
 #define IIF_FENCE_SIGNAL_TABLE_PROP_NAME "iif-fence-signal-table-region"
diff --git a/gcip-kernel-driver/drivers/gcip/iif/iif-fence.c b/gcip-kernel-driver/drivers/gcip/iif/iif-fence.c
index 5834af5..9a2563c 100644
--- a/gcip-kernel-driver/drivers/gcip/iif/iif-fence.c
+++ b/gcip-kernel-driver/drivers/gcip/iif/iif-fence.c
@@ -1,27 +1,34 @@
 // SPDX-License-Identifier: GPL-2.0-only
 /*
- * GCIP-integrated IIF driver fence.
+ * The inter-IP fence.
  *
- * Copyright (C) 2023 Google LLC
+ * The actual functionality (waiting and signaling) won't be done by the kernel driver. The main
+ * role of it is creating fences with assigning fence IDs, initializing the fence table and managing
+ * the life cycle of them.
+ *
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #define pr_fmt(fmt) "iif: " fmt
 
 #include <linux/atomic.h>
 #include <linux/container_of.h>
+#include <linux/export.h>
 #include <linux/file.h>
 #include <linux/fs.h>
 #include <linux/idr.h>
 #include <linux/kref.h>
+#include <linux/lockdep.h>
+#include <linux/lockdep_types.h>
 #include <linux/spinlock.h>
 #include <linux/types.h>
 #include <linux/version.h>
 
-#include <gcip/iif/iif-fence-table.h>
-#include <gcip/iif/iif-fence.h>
-#include <gcip/iif/iif-manager.h>
-#include <gcip/iif/iif-sync-file.h>
-#include <gcip/iif/iif.h>
+#include <iif/iif-fence-table.h>
+#include <iif/iif-fence.h>
+#include <iif/iif-manager.h>
+#include <iif/iif-sync-file.h>
+#include <iif/iif.h>
 
 /*
  * Returns the number of remaining signalers to be submitted. Returns 0 if all signalers are
@@ -86,21 +93,24 @@
 /*
  * Signals @fence.
  *
+ * The function returns the number of remaining signals to unblock the fence on success. Otherwise,
+ * returns a negative errno.
+ *
+ * If the function returns 0, it means that the fence has been unblocked and the caller is expected
+ * to call the `iif_fence_notify_poll_cb` function to notify all others waiting on the fence.
+ *
  * If @complete is true, it will make @fence have been signaled by all signalers. This must be used
- * only when @fence is going to be released before all signalers signal the fence and let the drive
- * side notice that there was some problem by triggering registered callbacks.
+ * only when @fence is going to be released before all signalers signal the fence.
  *
  * Caller must hold @fence->signaled_signalers_lock.
  */
-static void iif_fence_signal_locked(struct iif_fence *fence, bool complete)
+static int iif_fence_signal_locked(struct iif_fence *fence, bool complete)
 {
-	struct iif_fence_poll_cb *cur, *tmp;
-
 	lockdep_assert_held(&fence->signaled_signalers_lock);
 
 	if (iif_fence_is_signaled_locked(fence)) {
-		pr_warn("The fence is already signaled, id=%u", fence->id);
-		return;
+		pr_err("The fence is already signaled, id=%u", fence->id);
+		return -EBUSY;
 	}
 
 	if (!complete)
@@ -108,12 +118,22 @@
 	else
 		fence->signaled_signalers = fence->total_signalers;
 
-	/* All signalers have signaled the fence. */
-	if (iif_fence_is_signaled_locked(fence)) {
-		list_for_each_entry_safe(cur, tmp, &fence->poll_cb_list, node) {
-			list_del_init(&cur->node);
-			cur->func(fence, cur);
-		}
+	return fence->total_signalers - fence->signaled_signalers;
+}
+
+/*
+ * Notifies the poll callbacks registered to @fence.
+ *
+ * This function must be called only when @fence is signaled so that @fence->signaled_signalers_lock
+ * doesn't have to be held.
+ */
+static void iif_fence_notify_poll_cb(struct iif_fence *fence)
+{
+	struct iif_fence_poll_cb *cur, *tmp;
+
+	list_for_each_entry_safe(cur, tmp, &fence->poll_cb_list, node) {
+		list_del_init(&cur->node);
+		cur->func(fence, cur);
 	}
 }
 
@@ -176,11 +196,27 @@
 
 	if (!list_empty(&fence->poll_cb_list) && !iif_fence_is_signaled_locked(fence)) {
 		iif_fence_set_signal_error_locked(fence, -EDEADLK);
+		/*
+		 * We don't have to check the return value here since theoretically it will never
+		 * fail.
+		 */
 		iif_fence_signal_locked(fence, true);
 	}
 
 	spin_unlock_irqrestore(&fence->signaled_signalers_lock, flags);
 
+	/*
+	 * It is always safe to call this function.
+	 * - If the if-clause above was executed, it means that the fence has been signaled and it
+	 *   is good to call this function.
+	 * - If @fence->poll_cb_list was empty, this function call will be NO-OP.
+	 * - If `iif_fence_is_signaled_locked(fence)` was true, it means that the fence was already
+	 *   signaled and it is good to call it. (In this case, all callbacks should be called when
+	 *   the fence was signaled and @fence->poll_cb_list should be already empty. It means that
+	 *   the function call will be NO-OP theoretically.)
+	 */
+	iif_fence_notify_poll_cb(fence);
+
 	/* Checks whether there is remaining all_signaler_submitted callback. */
 	iif_fence_submitted_signalers_lock(fence);
 
@@ -200,16 +236,30 @@
 	 */
 	iif_fence_retire(fence);
 
+#if IS_ENABLED(CONFIG_DEBUG_SPINLOCK)
+	lockdep_unregister_key(&fence->submitted_signalers_key);
+#endif /* IS_ENABLED(CONFIG_DEBUG_SPINLOCK) */
+
 	if (fence->ops && fence->ops->on_release)
 		fence->ops->on_release(fence);
 }
 
+/*
+ * This callback will be registered to @fence when the fence is initialized as its signaler as AP.
+ * Also, it will be called when @fence has been unblocked.
+ */
+static void iif_fence_ap_poll_callback(struct iif_fence *fence, struct iif_fence_poll_cb *cb)
+{
+	iif_manager_broadcast_fence_unblocked_by_ap(fence->mgr, fence);
+}
+
 int iif_fence_init(struct iif_manager *mgr, struct iif_fence *fence,
 		   const struct iif_fence_ops *ops, enum iif_ip_type signaler_ip,
 		   uint16_t total_signalers)
 {
 	unsigned int id_min = signaler_ip * IIF_NUM_FENCES_PER_IP;
 	unsigned int id_max = id_min + IIF_NUM_FENCES_PER_IP - 1;
+	int ret = 0;
 
 	fence->id = ida_alloc_range(&mgr->idp, id_min, id_max, GFP_KERNEL);
 	if (fence->id < 0)
@@ -224,7 +274,14 @@
 	fence->ops = ops;
 	fence->state = IIF_FENCE_STATE_INITIALIZED;
 	kref_init(&fence->kref);
+#if IS_ENABLED(CONFIG_DEBUG_SPINLOCK)
+	lockdep_register_key(&fence->submitted_signalers_key);
+	__raw_spin_lock_init(spinlock_check(&fence->submitted_signalers_lock),
+			     "&fence->submitted_signalers_lock", &fence->submitted_signalers_key,
+			     LD_WAIT_CONFIG);
+#else
 	spin_lock_init(&fence->submitted_signalers_lock);
+#endif /* IS_ENABLED(CONFIG_DEBUG_SPINLOCK) */
 	spin_lock_init(&fence->signaled_signalers_lock);
 	spin_lock_init(&fence->outstanding_waiters_lock);
 	iif_fence_table_init_fence_entry(&mgr->fence_table, fence->id, total_signalers);
@@ -232,7 +289,11 @@
 	INIT_LIST_HEAD(&fence->all_signaler_submitted_cb_list);
 	atomic_set(&fence->num_sync_file, 0);
 
-	return 0;
+	if (signaler_ip == IIF_IP_AP)
+		ret = iif_fence_add_poll_callback(fence, &fence->ap_poll_cb,
+						  iif_fence_ap_poll_callback);
+
+	return ret;
 }
 
 int iif_fence_install_fd(struct iif_fence *fence)
@@ -333,13 +394,25 @@
 	return 0;
 }
 
-void iif_fence_signal(struct iif_fence *fence)
+int iif_fence_signal(struct iif_fence *fence)
 {
 	unsigned long flags;
+	int ret;
 
 	spin_lock_irqsave(&fence->signaled_signalers_lock, flags);
-	iif_fence_signal_locked(fence, false);
+	ret = iif_fence_signal_locked(fence, false);
 	spin_unlock_irqrestore(&fence->signaled_signalers_lock, flags);
+
+	if (!ret) {
+		/*
+		 * If @fence has been signaled, it is safe to execute all registered poll callbacks
+		 * without holding @fence->signaled_signalers_lock since the drivers can't register
+		 * callbacks anymore.
+		 */
+		iif_fence_notify_poll_cb(fence);
+	}
+
+	return ret;
 }
 
 void iif_fence_set_signal_error(struct iif_fence *fence, int error)
@@ -378,7 +451,7 @@
 	return signaled;
 }
 
-void iif_fence_waited(struct iif_fence *fence)
+void iif_fence_waited(struct iif_fence *fence, enum iif_ip_type ip)
 {
 	unsigned long flags;
 
diff --git a/gcip-kernel-driver/drivers/gcip/iif/iif-manager.c b/gcip-kernel-driver/drivers/gcip/iif/iif-manager.c
index 253c36f..cc84870 100644
--- a/gcip-kernel-driver/drivers/gcip/iif/iif-manager.c
+++ b/gcip-kernel-driver/drivers/gcip/iif/iif-manager.c
@@ -1,19 +1,24 @@
 // SPDX-License-Identifier: GPL-2.0-only
 /*
- * GCIP-integrated IIF driver manager.
+ * The manager of inter-IP fences.
  *
- * Copyright (C) 2023 Google LLC
+ * It manages the pool of fence IDs. The IIF driver device will initialize a manager and each IP
+ * driver will fetch the manager from the IIF device.
+ *
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #include <linux/container_of.h>
+#include <linux/export.h>
 #include <linux/idr.h>
 #include <linux/kref.h>
 #include <linux/of.h>
 #include <linux/slab.h>
 #include <linux/version.h>
 
-#include <gcip/iif/iif-fence-table.h>
-#include <gcip/iif/iif-manager.h>
+#include <iif/iif-fence-table.h>
+#include <iif/iif-fence.h>
+#include <iif/iif-manager.h>
 
 static void iif_manager_destroy(struct kref *kref)
 {
@@ -54,3 +59,37 @@
 {
 	kref_put(&mgr->kref, iif_manager_destroy);
 }
+
+int iif_manager_register_ops(struct iif_manager *mgr, enum iif_ip_type ip,
+			     const struct iif_manager_ops *ops, void *data)
+{
+	if (!ops || !ops->fence_unblocked_by_ap)
+		return -EINVAL;
+
+	mgr->ops[ip] = ops;
+	mgr->data[ip] = data;
+
+	return 0;
+}
+
+int iif_manager_acquire_block_wakelock(struct iif_manager *mgr, enum iif_ip_type ip)
+{
+	if (mgr->ops[ip] && mgr->ops[ip]->acquire_block_wakelock)
+		return mgr->ops[ip]->acquire_block_wakelock(mgr->data[ip]);
+	return 0;
+}
+
+void iif_manager_release_block_wakelock(struct iif_manager *mgr, enum iif_ip_type ip)
+{
+	if (mgr->ops[ip] && mgr->ops[ip]->release_block_wakelock)
+		mgr->ops[ip]->release_block_wakelock(mgr->data[ip]);
+}
+
+void iif_manager_broadcast_fence_unblocked_by_ap(struct iif_manager *mgr, struct iif_fence *fence)
+{
+	enum iif_ip_type ip;
+	unsigned int tmp;
+
+	for_each_waiting_ip(&mgr->fence_table, fence->id, ip, tmp)
+		mgr->ops[ip]->fence_unblocked_by_ap(fence, mgr->data[ip]);
+}
diff --git a/gcip-kernel-driver/drivers/gcip/iif/iif-signaler-submission-waiter.c b/gcip-kernel-driver/drivers/gcip/iif/iif-signaler-submission-waiter.c
index 98edc18..110cd9a 100644
--- a/gcip-kernel-driver/drivers/gcip/iif/iif-signaler-submission-waiter.c
+++ b/gcip-kernel-driver/drivers/gcip/iif/iif-signaler-submission-waiter.c
@@ -6,13 +6,14 @@
  */
 
 #include <linux/eventfd.h>
+#include <linux/export.h>
 #include <linux/kref.h>
 #include <linux/list.h>
 #include <linux/slab.h>
 #include <linux/spinlock.h>
 
-#include <gcip/iif/iif-fence.h>
-#include <gcip/iif/iif-signaler-submission-watier.h>
+#include <iif/iif-fence.h>
+#include <iif/iif-signaler-submission-waiter.h>
 
 static struct iif_signaler_submission_waiter *
 iif_signaler_submission_waiter_alloc(unsigned int eventfd, int pending_fences)
@@ -73,7 +74,7 @@
 
 	/*
 	 * The `iif_all_signaler_submission_waiter_cancel` function will delete @cb from
-	 * @watier->cb_list, decrement the refcount of @waiter and release @cb instead.
+	 * @waiter->cb_list, decrement the refcount of @waiter and release @cb instead.
 	 */
 	if (waiter->cancel) {
 		spin_unlock_irqrestore(&waiter->lock, flags);
diff --git a/gcip-kernel-driver/drivers/gcip/iif/iif-sync-file.c b/gcip-kernel-driver/drivers/gcip/iif/iif-sync-file.c
index f4982d3..3a1f88f 100644
--- a/gcip-kernel-driver/drivers/gcip/iif/iif-sync-file.c
+++ b/gcip-kernel-driver/drivers/gcip/iif/iif-sync-file.c
@@ -1,8 +1,12 @@
 // SPDX-License-Identifier: GPL-2.0-only
 /*
- * GCIP-integrated IIF driver sync file.
+ * IIF driver sync file.
  *
- * Copyright (C) 2023 Google LLC
+ * To export fences to the userspace, the driver will allocate a sync file to a fence and will
+ * return its file descriptor to the user. The user can distinguish fences with it. The driver will
+ * convert the file descriptor to the corresponding fence ID and will pass it to the IP.
+ *
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #include <linux/anon_inodes.h>
@@ -15,8 +19,8 @@
 #include <linux/types.h>
 #include <linux/uaccess.h>
 
-#include <gcip/iif/iif-fence.h>
-#include <gcip/iif/iif-sync-file.h>
+#include <iif/iif-fence.h>
+#include <iif/iif-sync-file.h>
 
 static void iif_sync_file_fence_signaled(struct iif_fence *fence, struct iif_fence_poll_cb *poll_cb)
 {
diff --git a/gcip-kernel-driver/include/gcip/iif/iif-fence-table.h b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-fence-table.h
similarity index 61%
rename from gcip-kernel-driver/include/gcip/iif/iif-fence-table.h
rename to gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-fence-table.h
index 0f10577..dfa52a3 100644
--- a/gcip-kernel-driver/include/gcip/iif/iif-fence-table.h
+++ b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-fence-table.h
@@ -1,17 +1,39 @@
 /* SPDX-License-Identifier: GPL-2.0-only */
 /*
- * GCIP-integrated IIF driver fence table.
+ * Interface to utilize IIF fence tables, the wait table and the signal table. Both tables will have
+ * one entry per fence ID.
  *
- * Copyright (C) 2023 Google LLC
+ * - Wait table: Describes which IPs are waiting on each fence. This table will be written by the
+ *               kernel driver only.
+ *
+ * - Signal table: Describes how many signals are remaining to unblock each fence. This table will
+ *                 be initialized by the kernel driver and each signaler IP will update it.
+ *
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #ifndef __IIF_IIF_FENCE_TABLE_H__
 #define __IIF_IIF_FENCE_TABLE_H__
 
+#include <linux/bitops.h>
 #include <linux/of.h>
 #include <linux/types.h>
 
-#include <gcip/iif/iif.h>
+#include <iif/iif.h>
+
+#define IIF_CLEAR_LSB(b) ((b) & ((b) - 1))
+
+/*
+ * Iterates the type of IPs waiting on the fence of @fence_id.
+ *
+ * fence_table (input): Pointer to the fence table.
+ * fence_id (input): ID of the fence to get IPs waiting on it.
+ * ip (output): Type of IP waiting on the fence (enum iif_ip_type).
+ * tmp (output): Temporary variable to iterate the wait table entry (int).
+ */
+#define for_each_waiting_ip(fence_table, fence_id, waiting_ip, tmp)                               \
+	for (tmp = (fence_table)->wait_table[fence_id].waiting_ips, waiting_ip = __ffs(tmp); tmp; \
+	     tmp = IIF_CLEAR_LSB(tmp), waiting_ip = __ffs(tmp))
 
 /* Entry of the wait table. */
 struct iif_wait_table_entry {
diff --git a/gcip-kernel-driver/include/gcip/iif/iif-fence.h b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-fence.h
similarity index 92%
rename from gcip-kernel-driver/include/gcip/iif/iif-fence.h
rename to gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-fence.h
index befd745..67ee061 100644
--- a/gcip-kernel-driver/include/gcip/iif/iif-fence.h
+++ b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-fence.h
@@ -1,19 +1,24 @@
 /* SPDX-License-Identifier: GPL-2.0-only */
 /*
- * GCIP-integrated IIF driver fence.
+ * The inter-IP fence.
  *
- * Copyright (C) 2023 Google LLC
+ * The actual functionality (waiting and signaling) won't be done by the kernel driver. The main
+ * role of it is creating fences with assigning fence IDs, initializing the fence table and managing
+ * the life cycle of them.
+ *
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #ifndef __IIF_IIF_FENCE_H__
 #define __IIF_IIF_FENCE_H__
 
 #include <linux/kref.h>
+#include <linux/lockdep_types.h>
 #include <linux/spinlock.h>
 #include <linux/types.h>
 
-#include <gcip/iif/iif-manager.h>
-#include <gcip/iif/iif.h>
+#include <iif/iif-manager.h>
+#include <iif/iif.h>
 
 struct iif_fence;
 struct iif_fence_ops;
@@ -50,6 +55,33 @@
 	IIF_FENCE_STATE_RETIRED,
 };
 
+/*
+ * Contains the callback function which will be called when all signalers have signaled the fence.
+ *
+ * The callback can be registered to the fence by the `iif_fence_add_poll_callback` function.
+ */
+struct iif_fence_poll_cb {
+	/* Node to be added to the list. */
+	struct list_head node;
+	/* Actual callback function to be called. */
+	iif_fence_poll_cb_t func;
+};
+
+/*
+ * Contains the callback function which will be called when all signalers have been submitted.
+ *
+ * The callback will be registered to the fence when the `iif_fence_submit_waiter` function fails
+ * in the submission.
+ */
+struct iif_fence_all_signaler_submitted_cb {
+	/* Node to be added to the list. */
+	struct list_head node;
+	/* Actual callback function to be called. */
+	iif_fence_all_signaler_submitted_cb_t func;
+	/* The number of remaining signalers to be submitted. */
+	int remaining_signalers;
+};
+
 /* The fence object. */
 struct iif_fence {
 	/* IIF manager. */
@@ -67,6 +99,9 @@
 	 * @all_signaler_submitted_error.
 	 */
 	spinlock_t submitted_signalers_lock;
+#if IS_ENABLED(CONFIG_DEBUG_SPINLOCK)
+	struct lock_class_key submitted_signalers_key;
+#endif /* IS_ENABLED(CONFIG_DEBUG_SPINLOCK) */
 	/* The interrupt state before holding @submitted_signalers_lock. */
 	unsigned long submitted_signalers_lock_flags;
 	/* The number of signaled signalers. */
@@ -93,6 +128,8 @@
 	int all_signaler_submitted_error;
 	/* The number of sync_file(s) bound to the fence. */
 	atomic_t num_sync_file;
+	/* The callback called if the fence's signaler is AP and the fence is signaled. */
+	struct iif_fence_poll_cb ap_poll_cb;
 };
 
 /* Operators of `struct iif_fence`. */
@@ -108,33 +145,6 @@
 };
 
 /*
- * Contains the callback function which will be called when all signalers have signaled the fence.
- *
- * The callback can be registered to the fence by the `iif_fence_add_poll_callback` function.
- */
-struct iif_fence_poll_cb {
-	/* Node to be added to the list. */
-	struct list_head node;
-	/* Actual callback function to be called. */
-	iif_fence_poll_cb_t func;
-};
-
-/*
- * Contains the callback function which will be called when all signalers have been submitted.
- *
- * The callback will be registered to the fence when the `iif_fence_submit_waiter` function fails
- * in the submission.
- */
-struct iif_fence_all_signaler_submitted_cb {
-	/* Node to be added to the list. */
-	struct list_head node;
-	/* Actual callback function to be called. */
-	iif_fence_all_signaler_submitted_cb_t func;
-	/* The number of remaining signalers to be submitted. */
-	int remaining_signalers;
-};
-
-/*
  * Initializes @fence which will be signaled by @signaler_ip IP. @total_signalers is the number of
  * signalers which must be submitted to the fence. Its initial reference count is 1.
  *
@@ -199,7 +209,7 @@
 int iif_fence_submit_waiter(struct iif_fence *fence, enum iif_ip_type ip);
 
 /* Signals @fence. If all signalers have signaled, it will notify polling FDs. */
-void iif_fence_signal(struct iif_fence *fence);
+int iif_fence_signal(struct iif_fence *fence);
 
 /*
  * Sets @fence->signal_error to let the user know that @fence has been signaled with an error.
@@ -226,8 +236,8 @@
  */
 bool iif_fence_is_signaled(struct iif_fence *fence);
 
-/* Notifies the driver that a waiter finished waiting on @fence. */
-void iif_fence_waited(struct iif_fence *fence);
+/* Notifies the driver that a waiter of @ip finished waiting on @fence. */
+void iif_fence_waited(struct iif_fence *fence, enum iif_ip_type ip);
 
 /*
  * Registers a callback which will be called when all signalers of @fence signaled. Once the
diff --git a/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-manager.h b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-manager.h
new file mode 100644
index 0000000..95c994c
--- /dev/null
+++ b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-manager.h
@@ -0,0 +1,157 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/*
+ * The manager of inter-IP fences.
+ *
+ * It manages the pool of fence IDs. The IIF driver device will initialize a manager and each IP
+ * driver will fetch the manager from the IIF device.
+ *
+ * Copyright (C) 2023-2024 Google LLC
+ */
+
+#ifndef __IIF_IIF_MANAGER_H__
+#define __IIF_IIF_MANAGER_H__
+
+#include <linux/cdev.h>
+#include <linux/device.h>
+#include <linux/idr.h>
+#include <linux/kref.h>
+#include <linux/of.h>
+#include <linux/types.h>
+
+#include <iif/iif-fence-table.h>
+
+struct iif_fence;
+
+/* Operators. */
+struct iif_manager_ops {
+	/* Following callbacks are required. */
+	/*
+	 * Called when @fence is unblocked by AP.
+	 *
+	 * This callback is meaningful only when @fence's signaler type is AP. When the runtime
+	 * signals @fence as many as the number of total signalers which was decided when the fence
+	 * was created and the fence has been unblocked eventually, this callback will be called if
+	 * the IP registered this operator is waiting on the fence to be unblocked.
+	 *
+	 * If this callback is called, the IP drivers should notify their IP that the fence has been
+	 * unblocked (i.e., send @fence->id to IP) so that they will check whether they can start
+	 * processing commands which were pended by the fence. Also, the IP drivers should check
+	 * whether the fence was signaled with an error or not by checking @fence->signal_error and
+	 * decide what to do if there was an error.
+	 *
+	 * This callback returns void since the IIF driver has nothing can do when the IP driver
+	 * fails to notify their IP. It is the responsibility of the IP driver side if that happens.
+	 *
+	 * Note that if IP drivers want to poll the fence unblock in general cases, they should
+	 * utilize the `iif_fence_{add,remove}_poll_callback` functions instead.
+	 *
+	 * Context: Normal.
+	 */
+	void (*fence_unblocked_by_ap)(struct iif_fence *fence, void *data);
+
+	/* Following callbacks are optional. */
+	/*
+	 * Acquires the block wakelock of IP.
+	 *
+	 * The callback will be executed when IP submits a waiter (i.e., `iif_fence_submit_waiter`
+	 * is called).
+	 *
+	 * This callback is required if the waiter IPx doesn't allow to be notified by the signaler
+	 * IPy when the IPx block is not powered on. If somehow the driver of waiter IPx submits a
+	 * waiter command to the firmware late and the signaler IPy processes its command early by
+	 * the race, IPy may try to notify IPx which is not powered on yet and it may cause a bug if
+	 * the IPx spec doesn't allow that. Therefore, if the IPx implements this callback, the IIF
+	 * driver will try to acquire the IPx block wakelock before submitting a waiter of IPx.
+	 *
+	 * Context: Depends on in which context the IPx calls the `iif_fence_submit_waiter`
+	 *          function.
+	 */
+	int (*acquire_block_wakelock)(void *data);
+	/*
+	 * Releases the block wakelock of IP.
+	 *
+	 * If both of the waiter IPx and the signaler IPy finishes waiting on / signaling the fence,
+	 * (i.e., both `iif_fence_waited` and `iif_fence_signal` are called) the block wakelock of
+	 * IPx will be released.
+	 *
+	 * If somehow IPx driver calls `iif_fence_waited` before the fence is signaled (For example,
+	 * IPy is not responding and IPx has processed its command as timeout.), releasing the block
+	 * wakelock of IPx will be pended until IPy processes its command normally or as timeout and
+	 * IPy driver eventually calls `iif_fence_signal`.
+	 *
+	 * Context: Depends on in which context the IPx calls the `iif_fence_waited` function or
+	 *          the IPy calls the `iif_fence_signal` function. Since this callback will be
+	 *          implemented by the IPx and IPx may not be aware of in which context IPy calls
+	 *          the `iif_fence_signal` function, it's recommended to assume the context is any
+	 *          including IRQ context.
+	 */
+	void (*release_block_wakelock)(void *data);
+};
+
+/*
+ * The structure overall data required by IIF driver such as fence table.
+ *
+ * Until we have stand-alone IIF driver, one of the IP drivers will initializes a manager by
+ * the `iif_init` function and every IP driver will share it.
+ */
+struct iif_manager {
+	/* Reference count of this instance. */
+	struct kref kref;
+	/* Fence ID pool. */
+	struct ida idp;
+	/* Fence table shared with the firmware. */
+	struct iif_fence_table fence_table;
+	/* Operators per IP. */
+	const struct iif_manager_ops *ops[IIF_IP_RESERVED];
+	/* User-data per IP. */
+	void *data[IIF_IP_RESERVED];
+	/* Platform bus device. */
+	struct device *dev;
+	/* Char device structure. */
+	struct cdev char_dev;
+	/* Char device number. */
+	dev_t char_dev_no;
+};
+
+/*
+ * Initializes IIF driver and returns its manager. Its initial reference count is 1. It will map
+ * the fence table by parsing the device tree via @np.
+ *
+ * The returned manager will be destroyed when its reference count becomes 0 by `iif_manager_put`
+ * function.
+ */
+struct iif_manager *iif_manager_init(const struct device_node *np);
+
+/* Increases the reference count of @mgr. */
+struct iif_manager *iif_manager_get(struct iif_manager *mgr);
+
+/* Decreases the reference count of @mgr and if it becomes 0, releases @mgr. */
+void iif_manager_put(struct iif_manager *mgr);
+
+/*
+ * Registers operators of @ip.
+ *
+ * Note that @ops must not be released until @ip won't be utilized as signaler or waiter anymore.
+ */
+int iif_manager_register_ops(struct iif_manager *mgr, enum iif_ip_type ip,
+			     const struct iif_manager_ops *ops, void *data);
+
+/*
+ * Acquires the block wakelock of @ip.
+ *
+ * Returns 0 on success or @ip hasn't defined the `acquire_block_wakelock` operator. Otherwise,
+ * returns a negative errno.
+ */
+int iif_manager_acquire_block_wakelock(struct iif_manager *mgr, enum iif_ip_type ip);
+
+/* Releases the block wakelock of @ip. */
+void iif_manager_release_block_wakelock(struct iif_manager *mgr, enum iif_ip_type ip);
+
+/*
+ * Notifies @fence has been unblocked by AP to IPs waiting on the fence.
+ *
+ * This function will be called if @fence's signaler is AP and the fence has been unblocked.
+ */
+void iif_manager_broadcast_fence_unblocked_by_ap(struct iif_manager *mgr, struct iif_fence *fence);
+
+#endif /* __IIF_IIF_MANAGER_H__ */
diff --git a/gcip-kernel-driver/include/gcip/iif/iif-signaler-submission-watier.h b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-signaler-submission-waiter.h
similarity index 98%
rename from gcip-kernel-driver/include/gcip/iif/iif-signaler-submission-watier.h
rename to gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-signaler-submission-waiter.h
index 6692b80..3d53972 100644
--- a/gcip-kernel-driver/include/gcip/iif/iif-signaler-submission-watier.h
+++ b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-signaler-submission-waiter.h
@@ -8,7 +8,7 @@
 #ifndef __IIF_IIF_SIGNALER_SUBMISSION_WAITER_H__
 #define __IIF_IIF_SIGNALER_SUBMISSION_WAITER_H__
 
-#include <gcip/iif/iif-fence.h>
+#include <iif/iif-fence.h>
 
 #define IIF_NO_REGISTER_EVENTFD (~0u)
 
diff --git a/gcip-kernel-driver/include/gcip/iif/iif-sync-file.h b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-sync-file.h
similarity index 75%
rename from gcip-kernel-driver/include/gcip/iif/iif-sync-file.h
rename to gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-sync-file.h
index a79f25b..7b4711c 100644
--- a/gcip-kernel-driver/include/gcip/iif/iif-sync-file.h
+++ b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif-sync-file.h
@@ -1,8 +1,12 @@
 /* SPDX-License-Identifier: GPL-2.0-only */
 /*
- * GCIP-integrated IIF driver sync file.
+ * IIF driver sync file.
  *
- * Copyright (C) 2023 Google LLC
+ * To export fences to the userspace, the driver will allocate a sync file to a fence and will
+ * return its file descriptor to the user. The user can distinguish fences with it. The driver will
+ * convert the file descriptor to the corresponding fence ID and will pass it to the IP.
+ *
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #ifndef __IIF_IIF_SYNC_FILE_H__
@@ -11,7 +15,7 @@
 #include <linux/file.h>
 #include <linux/wait.h>
 
-#include <gcip/iif/iif-fence.h>
+#include <iif/iif-fence.h>
 
 #define IIF_SYNC_FILE_FLAGS_POLL_ENABLED 0
 
diff --git a/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif.h b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif.h
new file mode 100644
index 0000000..b71b69c
--- /dev/null
+++ b/gcip-kernel-driver/drivers/gcip/iif/include/iif/iif.h
@@ -0,0 +1,148 @@
+/* SPDX-License-Identifier: GPL-2.0-only */
+/*
+ * Defines the interface of the IIF driver.
+ *
+ * Copyright (C) 2023 Google LLC
+ */
+
+#ifndef __IIF_IIF_H__
+#define __IIF_IIF_H__
+
+#include <linux/ioctl.h>
+#include <linux/types.h>
+
+/* Interface Version. */
+#define IIF_INTERFACE_VERSION_MAJOR 1
+#define IIF_INTERFACE_VERSION_MINOR 0
+
+#define IIF_IOCTL_BASE 'i'
+
+/* The ioctl number for the fence FDs will start from here. */
+#define IIF_FENCE_IOCTL_NUM_BASE 0x80
+
+/*
+ * The max number of fences can be created per IP.
+ * Increasing this value needs to increase the size of fence table.
+ */
+#define IIF_NUM_FENCES_PER_IP 1024
+
+/* The maximum number of fences can be passed to one ioctl request. */
+#define IIF_MAX_NUM_FENCES 64
+
+/*
+ * Type of IPs.
+ *
+ * The order of IP must be matched with the firmware side because the fence ID will be assigned
+ * according to the IP type.
+ */
+enum iif_ip_type {
+	IIF_IP_DSP,
+	IIF_IP_TPU,
+	IIF_IP_GPU,
+	IIF_IP_AP,
+	IIF_IP_NUM,
+
+	/* Reserve the number of IP type to expand the fence table easily in the future. */
+	IIF_IP_RESERVED = 16,
+};
+
+/*
+ * ioctls for /dev/iif.
+ */
+
+struct iif_create_fence_ioctl {
+	/*
+	 * Input:
+	 * The type of the fence signaler IP. (See enum iif_ip_type)
+	 */
+	__u8 signaler_ip;
+	/*
+	 * Input:
+	 * The number of the signalers.
+	 */
+	__u16 total_signalers;
+	/*
+	 * Output:
+	 * The file descriptor of the created fence.
+	 */
+	__s32 fence;
+};
+
+/* Create an IIF fence. */
+#define IIF_CREATE_FENCE _IOWR(IIF_IOCTL_BASE, 0, struct iif_create_fence_ioctl)
+
+/*
+ * The ioctl won't register @eventfd and will simply return the number of
+ * remaining signalers of each fence.
+ */
+#define IIF_FENCE_REMAINING_SIGNALERS_NO_REGISTER_EVENTFD (~0u)
+
+struct iif_fence_remaining_signalers_ioctl {
+	/*
+	 * Input:
+	 * User-space pointer to an int array of inter-IP fence file descriptors
+	 * to check whether there are remaining signalers to be submitted or
+	 * not.
+	 */
+	__u64 fences;
+	/*
+	 * Input:
+	 * The number of fences in `fence_array`.
+	 * If > IIF_MAX_NUM_FENCES, the ioctl will fail with errno == EINVAL.
+	 */
+	__u32 fences_count;
+	/*
+	 * Input:
+	 * The eventfd which will be triggered if there were fence(s) which
+	 * haven't finished the signaler submission yet when the ioctl is called
+	 * and when they eventually have finished the submission. Note that if
+	 * all fences already finished the submission (i.e., all values in the
+	 * returned @remaining_signalers are 0), this eventfd will be ignored.
+	 *
+	 * Note that if `IIF_FENCE_REMAINING_SIGNALERS_NO_REGISTER_EVENTFD` is
+	 * passed, this ioctl will simply return the number of remaining
+	 * signalers of each fence to @remaining_signalers.
+	 */
+	__u32 eventfd;
+	/*
+	 * Output:
+	 * User-space pointer to an int array where the driver will write the
+	 * number of remaining signalers to be submitted per fence. The order
+	 * will be the same with @fences.
+	 */
+	__u64 remaining_signalers;
+};
+
+/*
+ * Check whether there are remaining signalers to be submitted to fences.
+ * If all signalers have been submitted, the runtime is expected to send waiter
+ * commands right away. Otherwise, it will listen the eventfd to wait signaler
+ * submission to be finished.
+ */
+#define IIF_FENCE_REMAINING_SIGNALERS \
+	_IOWR(IIF_IOCTL_BASE, 1, struct iif_fence_remaining_signalers_ioctl)
+
+/*
+ * ioctls for inter-IP fence FDs.
+ */
+
+struct iif_fence_get_information_ioctl {
+	/* The type of the signaler IP. (enum iif_ip_type) */
+	__u8 signaler_ip;
+	/* The number of total signalers. */
+	__u16 total_signalers;
+	/* The number of submitted signalers. */
+	__u16 submitted_signalers;
+	/* The number of signaled signalers. */
+	__u16 signaled_signalers;
+	/* The number of outstanding waiters. */
+	__u16 outstanding_waiters;
+	/* Reserved. */
+	__u8 reserved[7];
+};
+
+/* Returns the fence information. */
+#define IIF_FENCE_GET_INFORMATION \
+	_IOR(IIF_IOCTL_BASE, IIF_FENCE_IOCTL_NUM_BASE, struct iif_fence_get_information_ioctl)
+
+#endif /* __IIF_IIF_H__ */
diff --git a/gcip-kernel-driver/include/gcip/gcip-alloc-helper.h b/gcip-kernel-driver/include/gcip/gcip-alloc-helper.h
index 17208bf..4f61135 100644
--- a/gcip-kernel-driver/include/gcip/gcip-alloc-helper.h
+++ b/gcip-kernel-driver/include/gcip/gcip-alloc-helper.h
@@ -13,12 +13,11 @@
 #include <linux/types.h>
 
 /*
- * The actual return value from the alloc_noncontiguous function.
- * The user should only care about @sgt. @pages is used internally for freeing memory.
+ * Used internally by the alloc_noncontiguous functions.
  */
 struct gcip_sgt_handle {
-	struct sg_table sgt;
-	void *mem;
+	struct sg_table sgt;	/* SG table for vmalloc'ed physical pages */
+	void *mem;		/* kernel VA of vmalloc area */
 };
 
 /*
@@ -28,10 +27,11 @@
  * @size: Total size in bytes. Will be page aligned.
  * @gfp: The GFP flag for malloc internal structures.
  *
- * Returns the SG table represents the non-contiguous region.
+ * Returns an SG table for the non-contiguous pages.
  * Returns NULL on any error.
  */
 struct sg_table *gcip_alloc_noncontiguous(struct device *dev, size_t size, gfp_t gfp);
+
 /* Frees the memory allocated by gcip_alloc_noncontiguous. */
 void gcip_free_noncontiguous(struct sg_table *sgt);
 
diff --git a/gcip-kernel-driver/include/gcip/gcip-dma-fence.h b/gcip-kernel-driver/include/gcip/gcip-dma-fence.h
index 87c7467..774bc98 100644
--- a/gcip-kernel-driver/include/gcip/gcip-dma-fence.h
+++ b/gcip-kernel-driver/include/gcip/gcip-dma-fence.h
@@ -161,7 +161,7 @@
 void gcip_dma_fence_show(struct gcip_dma_fence *gfence, struct seq_file *s);
 
 /**
- * Gets and merges an array of DMA fences from their FDs.
+ * gcip_dma_fence_merge_fences() -  Merges an array of DMA fences with dma_fence_unwrap_merge().
  *
  * Creates a dma_fence_array from all of the provided fences and returns a dma_fence representing
  * that array. If any of the provided fences are also arrays, the resulting array will include
@@ -171,12 +171,19 @@
  *
  * The returned fence must be released with `dma_fence_put()`.
  *
+ * Returns a pointer to the fence on success. Otherwise a negative errno as an ERR_PTR.
+ */
+struct dma_fence *gcip_dma_fence_merge_fences(int num_fences, struct dma_fence **fences);
+
+/**
+ * gcip_dma_fence_merge_fds() - Gets and merges an array of DMA fences from their FDs.
+ *
+ * Creates a dma_fence_array from all of the provided fences.
  * It is OK if @fence_fds do not refer to gcip_dma_fences.
  *
  * Returns a pointer to the fence on success. Otherwise a negative errno as an ERR_PTR.
  *  - If unable to allocate sufficient memory, returns ERR_PTR(-ENOMEM)
  *  - If any of @fence_fds are invalid, returns ERR_PTR(-ENOENT)
- *  - If unable to merge the fences, returns NULL
  */
 struct dma_fence *gcip_dma_fence_merge_fds(int num_fences, int *fence_fds);
 
diff --git a/gcip-kernel-driver/include/gcip/gcip-fence-array.h b/gcip-kernel-driver/include/gcip/gcip-fence-array.h
index afcc7ea..6b6dbfb 100644
--- a/gcip-kernel-driver/include/gcip/gcip-fence-array.h
+++ b/gcip-kernel-driver/include/gcip/gcip-fence-array.h
@@ -8,6 +8,7 @@
 #ifndef __GCIP_FENCE_ARRAY_H__
 #define __GCIP_FENCE_ARRAY_H__
 
+#include <linux/dma-fence.h>
 #include <linux/kref.h>
 
 #include <gcip/gcip-fence.h>
@@ -54,21 +55,22 @@
 void gcip_fence_array_signal(struct gcip_fence_array *fence_array, int errno);
 
 /*
- * Notifies the fences in @fence_array that a command which waited on them has finished their work.
+ * Notifies the fences in @fence_array that a command of @ip which waited on them has finished their
+ * work.
  *
  * This function is only meaningful when a fence is GCIP_INTER_IP_FENCE.
  */
-void gcip_fence_array_waited(struct gcip_fence_array *fence_array);
+void gcip_fence_array_waited(struct gcip_fence_array *fence_array, enum iif_ip_type ip);
 
 /* Submits a signaler to the fences in @fence_array. */
 void gcip_fence_array_submit_signaler(struct gcip_fence_array *fence_array);
 
-/* Submits a waiter to the fences in @fence_array. */
-void gcip_fence_array_submit_waiter(struct gcip_fence_array *fence_array);
+/* Submits a waiter of @ip to the fences in @fence_array. */
+void gcip_fence_array_submit_waiter(struct gcip_fence_array *fence_array, enum iif_ip_type ip);
 
 /*
- * Submits a waiter to each fence in @in_fences and a signaler to each fence in @out_fences. Either
- * @in_fences or @out_fences is allowed to be NULL.
+ * Submits a waiter of @ip to each fence in @in_fences and a signaler to each fence in @out_fences.
+ * Either @in_fences or @out_fences is allowed to be NULL.
  *
  * For the waiter submission, if at least one fence of @in_fences haven't finished the signaler
  * submission, this function will fail and return -EAGAIN.
@@ -82,7 +84,8 @@
  * Otherwise, returns 0 on success.
  */
 int gcip_fence_array_submit_waiter_and_signaler(struct gcip_fence_array *in_fences,
-						struct gcip_fence_array *out_fences);
+						struct gcip_fence_array *out_fences,
+						enum iif_ip_type ip);
 
 /*
  * Allocates and returns the array of inter-IP fence IDs. The number of IIFs in @fence_array will
@@ -108,4 +111,22 @@
 int gcip_fence_array_wait_signaler_submission(struct gcip_fence_array *fence_array,
 					      unsigned int eventfd, int *remaining_signalers);
 
+/*
+ * Creates a merged dma_fence_array object of the underlying DMA fences of @fence_array.
+ *
+ * Note that this function is meaningful only when the fence type of all fences of @fence_array are
+ * the same (i.e., @fence_array->same_type is true) and the type is DMA fence (i.e., fence_array->
+ * type is GCIP_IN_KERNEL_FENCE). If that's not the case or there is no fence in @fence_array, the
+ * function will return NULL.
+ *
+ * If there is only one DMA fence in @fence_array, the function will return the DMA fence itself,
+ * not a base fence of merged dma_fence_array.
+ *
+ * Returns the representing fence of the merged DMA fences on success or NULL if @fence_array is not
+ * meaningful to be merged. Otherwise, returns an errno pointer.
+ *
+ * The returned fence must be released with `dma_fence_put()`.
+ */
+struct dma_fence *gcip_fence_array_merge_ikf(struct gcip_fence_array *fence_array);
+
 #endif /* __GCIP_FENCE_ARRAY_H__ */
diff --git a/gcip-kernel-driver/include/gcip/gcip-fence.h b/gcip-kernel-driver/include/gcip/gcip-fence.h
index 5c0417a..004fa61 100644
--- a/gcip-kernel-driver/include/gcip/gcip-fence.h
+++ b/gcip-kernel-driver/include/gcip/gcip-fence.h
@@ -11,9 +11,9 @@
 #include <linux/dma-fence.h>
 #include <linux/kref.h>
 
-#include <gcip/iif/iif-fence.h>
-#include <gcip/iif/iif-manager.h>
-#include <gcip/iif/iif.h>
+#include <iif/iif-fence.h>
+#include <iif/iif-manager.h>
+#include <iif/iif.h>
 
 #define GCIP_FENCE_REMAINING_SIGNALERS_NO_REGISTER_EVENTFD (~0u)
 
@@ -90,7 +90,7 @@
 int gcip_fence_submit_signaler_locked(struct gcip_fence *fence);
 
 /*
- * Submits a waiter.
+ * Submits a waiter of @ip.
  * Note that the waiter submission will not be done when not all signalers have been submitted.
  *
  * This function is only meaningful when the fence type is GCIP_INTER_IP_FENCE and can be called in
@@ -100,7 +100,7 @@
  * has been succeeded when the function returns 0.) Otherwise, returns a negative errno if it fails
  * with other reasons.
  */
-int gcip_fence_submit_waiter(struct gcip_fence *fence);
+int gcip_fence_submit_waiter(struct gcip_fence *fence, enum iif_ip_type ip);
 
 /*
  * Signals @fence. If all signalers have signaled the fence, it will notify polling FDs.
@@ -110,11 +110,11 @@
 void gcip_fence_signal(struct gcip_fence *fence, int errno);
 
 /*
- * Notifies @fence that a command which waited the fence has finished their work.
+ * Notifies @fence that a command of @ip which waited the fence has finished their work.
  *
  * This function is only meaningful when the fence type is GCIP_INTER_IP_FENCE.
  */
-void gcip_fence_waited(struct gcip_fence *fence);
+void gcip_fence_waited(struct gcip_fence *fence, enum iif_ip_type ip);
 
 /*
  * Registers a callback which will be called when all signalers are submitted for @fence and
diff --git a/gcip-kernel-driver/include/gcip/gcip-image-config.h b/gcip-kernel-driver/include/gcip/gcip-image-config.h
index d6e66ad..bebae45 100644
--- a/gcip-kernel-driver/include/gcip/gcip-image-config.h
+++ b/gcip-kernel-driver/include/gcip/gcip-image-config.h
@@ -8,12 +8,12 @@
 #ifndef __GCIP_IMAGE_CONFIG_H__
 #define __GCIP_IMAGE_CONFIG_H__
 
-#include <asm/page.h>
 #include <linux/bits.h>
+#include <linux/sizes.h>
 #include <linux/types.h>
 
 #define GCIP_FW_NUM_VERSIONS 4
-#define GCIP_IMG_CFG_MAX_IOMMU_MAPPINGS 20
+#define GCIP_IMG_CFG_MAX_IOMMU_MAPPINGS 19
 #define GCIP_IMG_CFG_MAX_NS_IOMMU_MAPPINGS 5
 #define GCIP_IMG_CFG_MAX_PROTECTED_MEMORY_MAPPINGS 3
 
@@ -21,6 +21,30 @@
 #define GCIP_FW_PRIV_LEVEL_TZ 1
 #define GCIP_FW_PRIV_LEVEL_NS 2
 
+#define GCIP_FW_ASAN_ENABLED (1 << 0)
+#define GCIP_FW_UBSAN_ENABLED (1 << 1)
+
+#define GCIP_IMAGE_CONFIG_SANITIZER_SHIFT 2
+#define GCIP_IMAGE_CONFIG_SANITIZER_MASK (BIT(GCIP_IMAGE_CONFIG_SANITIZER_SHIFT) - 1)
+#define GCIP_IMAGE_CONFIG_SANITIZER_STATUS(cfg) (cfg & GCIP_IMAGE_CONFIG_SANITIZER_MASK)
+#define GCIP_IMAGE_CONFIG_SANITIZER_INDEX(cfg) (cfg >> GCIP_IMAGE_CONFIG_SANITIZER_SHIFT)
+
+/*
+ * Mapping flags in lower 12 bits (regardless of page size in use by AP kernel) of requested
+ * virt_address in iommu_mappings.  Must match definitions used by firmware.
+ */
+#define GCIP_IMG_CFG_MAP_FLAGS_MASK		(SZ_4K - 1)
+
+/* The entry is for a CSR/device region and must be mapped with IOMMU_MMIO flag. */
+#define GCIP_IMAGE_CONFIG_MAP_MMIO_BIT		(BIT(0))
+#define GCIP_IMAGE_CONFIG_MAP_MMIO(flags)	(flags & GCIP_IMAGE_CONFIG_MAP_MMIO_BIT)
+/* The mapping must be replicated for each PASID/context. */
+#define GCIP_IMAGE_CONFIG_MAP_SHARED_BIT	(BIT(1))
+#define GCIP_IMAGE_CONFIG_MAP_SHARED(flags)	(flags & GCIP_IMAGE_CONFIG_MAP_SHARED_BIT)
+/* The mapping uses a 36-bit IOVA. The incoming value needs to be shifted left 4 bits */
+#define GCIP_IMAGE_CONFIG_MAP_36BIT_BIT		(BIT(3))
+#define GCIP_IMAGE_CONFIG_MAP_36BIT(flags)	(flags & GCIP_IMAGE_CONFIG_MAP_36BIT_BIT)
+
 /*
  * The image configuration attached to the signed firmware.
  */
@@ -35,7 +59,7 @@
 	__u32 remapped_region_size;
 	__u32 num_iommu_mappings;
 	struct {
-		/* Device virtual address */
+		/* Device virtual address requested, with map flags in lower 12 bits */
 		__u32 virt_address;
 		/*
 		 * Encodes a 12-bit aligned address and the corresponding size
@@ -44,6 +68,8 @@
 		 */
 		__u32 image_config_value;
 	} iommu_mappings[GCIP_IMG_CFG_MAX_IOMMU_MAPPINGS];
+	__u32 reserved;
+	__u32 sanitizer_config;
 	__u32 protected_memory_regions[GCIP_IMG_CFG_MAX_PROTECTED_MEMORY_MAPPINGS];
 	__u32 secure_telemetry_region_start;
 	__u32 remapped_data_start;
@@ -60,21 +86,24 @@
 	 *
 	 * It is ensured that there is no overflow on @paddr + @size before calling this function.
 	 *
-	 * @flags is a bit-field with the following attributes:
-	 *   [0:0] - Security. 1 for secure and 0 for non-secure.
+	 * @cfg_map_flags holds GCIP_IMAGE_CONFIG_MAP_* flags masked from an iommu_mappings_entry,
+	 *                or zero for ns_iommu_mappings (GCIP_IMAGE_CONFIG_FLAGS_SECURE == 0)
+	 * @cfg_op_flags is a bit-field with the following attributes:
+	 *   [0:0] - Secure: 1 = handle secure iommu_mappings[] entry, 0 = ns_iommu_mappings[]
 	 *   [31:1] - Reserved.
 	 *
 	 * Returns 0 on success. Otherwise a negative errno.
 	 * Mandatory.
 	 */
 	int (*map)(void *data, dma_addr_t daddr, phys_addr_t paddr, size_t size,
-		   unsigned int flags);
+		   unsigned int cfg_map_flags, unsigned int cfg_op_flags);
 	/*
 	 * Removes the IOMMU mapping previously added by @map.
 	 *
 	 * Mandatory.
 	 */
-	void (*unmap)(void *data, dma_addr_t daddr, size_t size, unsigned int flags);
+	void (*unmap)(void *data, dma_addr_t daddr, size_t size, unsigned int cfg_map_flags,
+		      unsigned int cfg_op_flags);
 };
 
 struct gcip_image_config_parser {
@@ -183,4 +212,12 @@
 	return config->privilege_level == GCIP_FW_PRIV_LEVEL_NS;
 }
 
+/*
+ * Returns whether the privilege level specified by @config is secure.
+ */
+static inline bool gcip_image_config_is_secure(struct gcip_image_config *config)
+{
+	return config->privilege_level != GCIP_FW_PRIV_LEVEL_NS;
+}
+
 #endif /* __GCIP_IMAGE_CONFIG_H__ */
diff --git a/gcip-kernel-driver/include/gcip/gcip-iommu.h b/gcip-kernel-driver/include/gcip/gcip-iommu.h
index 425691d..96dd824 100644
--- a/gcip-kernel-driver/include/gcip/gcip-iommu.h
+++ b/gcip-kernel-driver/include/gcip/gcip-iommu.h
@@ -62,6 +62,12 @@
 #define GCIP_MAP_FLAGS_RESTRICT_IOVA_TO_FLAGS(restrict) \
 	((u64)(restrict) << GCIP_MAP_FLAGS_RESTRICT_IOVA_OFFSET)
 
+#define GCIP_MAP_FLAGS_MMIO_OFFSET \
+	(GCIP_MAP_FLAGS_RESTRICT_IOVA_OFFSET + GCIP_MAP_FLAGS_RESTRICT_IOVA_BIT_SIZE)
+#define GCIP_MAP_FLAGS_MMIO_BIT_SIZE 1
+#define GCIP_MAP_FLAGS_MMIO_TO_FLAGS(mmio) \
+	((u64)(mmio) << GCIP_MAP_FLAGS_RESTRICT_IOVA_OFFSET)
+
 /* Helper macros to easily create the mapping direction flags. */
 #define GCIP_MAP_FLAGS_DMA_RW GCIP_MAP_FLAGS_DMA_DIRECTION_TO_FLAGS(DMA_BIDIRECTIONAL)
 #define GCIP_MAP_FLAGS_DMA_RO GCIP_MAP_FLAGS_DMA_DIRECTION_TO_FLAGS(DMA_TO_DEVICE)
@@ -80,7 +86,9 @@
  *               (See [REDACTED]
  *   [13:13] - RESTRICT_IOVA:
  *               Restrict the IOVA assignment to 32 bit address window.
- *   [63:14] - RESERVED
+ *   [14:14] - MMIO:
+ *               Mapping is for device memory, use IOMMU_MMIO flag.
+ *   [63:15] - RESERVED
  *               Set RESERVED bits to 0 to ensure backwards compatibility.
  *
  * One can use gcip_iommu_encode_gcip_map_flags or `GCIP_MAP_FLAGS_DMA_*_TO_FLAGS` macros to
@@ -539,6 +547,13 @@
 	mapping->data = data;
 }
 
+static inline size_t gcip_iommu_domain_granule(struct gcip_iommu_domain *domain)
+{
+	if (unlikely(domain->default_domain))
+		return PAGE_SIZE;
+	return domain->domain_pool->granule;
+}
+
 /**
  * gcip_iommu_map() - Maps the desired mappings to the domain.
  * @domain: The GCIP domain to be mapped to.
diff --git a/gcip-kernel-driver/include/gcip/gcip-mailbox.h b/gcip-kernel-driver/include/gcip/gcip-mailbox.h
index 1088303..13a0ea3 100644
--- a/gcip-kernel-driver/include/gcip/gcip-mailbox.h
+++ b/gcip-kernel-driver/include/gcip/gcip-mailbox.h
@@ -137,18 +137,15 @@
 struct gcip_mailbox_ops {
 	/* Mandatory. */
 	/*
-	 * Gets the head of mailbox command queue.
-	 * Context: normal.
-	 */
-	u32 (*get_cmd_queue_head)(struct gcip_mailbox *mailbox);
-	/*
 	 * Gets the tail of mailbox command queue.
-	 * Context: normal.
+	 *
+	 * Context: cmd_queue_lock.
 	 */
 	u32 (*get_cmd_queue_tail)(struct gcip_mailbox *mailbox);
 	/*
 	 * Increases the tail of mailbox command queue by @inc.
-	 * Context: normal.
+	 *
+	 * Context: cmd_queue_lock.
 	 */
 	void (*inc_cmd_queue_tail)(struct gcip_mailbox *mailbox, u32 inc);
 	/*
@@ -167,43 +164,61 @@
 	int (*acquire_cmd_queue_lock)(struct gcip_mailbox *mailbox, bool try, bool *atomic);
 	/*
 	 * Releases the lock of cmd_queue which is acquired by calling `acquire_cmd_queue_lock`.
-	 * Context: normal.
+	 *
+	 * Context: cmd_queue_lock.
 	 */
 	void (*release_cmd_queue_lock)(struct gcip_mailbox *mailbox);
 	/*
 	 * Gets the sequence number of @cmd queue element.
-	 * Context: normal.
+	 *
+	 * Context: cmd_queue_lock.
 	 */
 	u64 (*get_cmd_elem_seq)(struct gcip_mailbox *mailbox, void *cmd);
 	/*
 	 * Sets the sequence number of @cmd queue element.
-	 * Context: normal.
+	 *
+	 * Context: cmd_queue_lock.
 	 */
 	void (*set_cmd_elem_seq)(struct gcip_mailbox *mailbox, void *cmd, u64 seq);
 	/*
 	 * Gets the code of @cmd queue element.
+	 *
 	 * Context: normal.
 	 */
 	u32 (*get_cmd_elem_code)(struct gcip_mailbox *mailbox, void *cmd);
+	/*
+	 * Waits for the cmd queue of @mailbox has a available space for putting the command. If
+	 * the queue has a space, returns 0. Otherwise, returns error as non-zero value. It depends
+	 * on the implementation details, but it is okay to return right away with error when the
+	 * queue is full. If this callback returns an error, `gcip_mailbox_send_cmd` function or
+	 * `gcip_mailbox_put_cmd` function will return that error too.
+	 *
+	 * Context: cmd_queue_lock.
+	 */
+	int (*wait_for_cmd_queue_not_full)(struct gcip_mailbox *mailbox);
 
 	/*
 	 * Gets the size of mailbox response queue.
+	 *
 	 * Context: normal.
 	 */
 	u32 (*get_resp_queue_size)(struct gcip_mailbox *mailbox);
 	/*
 	 * Gets the head of mailbox response queue.
-	 * Context: normal and in_interrupt().
+	 *
+	 * Context: resp_queue_lock.
 	 */
 	u32 (*get_resp_queue_head)(struct gcip_mailbox *mailbox);
 	/*
 	 * Gets the tail of mailbox response queue.
-	 * Context: normal and in_interrupt().
+	 *
+	 * Context: resp_queue_lock.
 	 */
 	u32 (*get_resp_queue_tail)(struct gcip_mailbox *mailbox);
 	/*
 	 * Increases the head of mailbox response queue by @inc.
-	 * Context: normal and in_interrupt().
+	 *
+	 * Context: resp_queue_lock.
 	 */
 	void (*inc_resp_queue_head)(struct gcip_mailbox *mailbox, u32 inc);
 	/*
@@ -224,17 +239,20 @@
 	int (*acquire_resp_queue_lock)(struct gcip_mailbox *mailbox, bool try, bool *atomic);
 	/*
 	 * Releases the lock of resp_queue which is acquired by calling `acquire_resp_queue_lock`.
-	 * Context: normal and in_interrupt().
+	 *
+	 * Context: resp_queue_lock.
 	 */
 	void (*release_resp_queue_lock)(struct gcip_mailbox *mailbox);
 	/*
 	 * Gets the sequence number of @resp queue element.
-	 * Context: normal and in_interrupt().
+	 *
+	 * Context: wait_list_lock.
 	 */
 	u64 (*get_resp_elem_seq)(struct gcip_mailbox *mailbox, void *resp);
 	/*
 	 * Sets the sequence number of @resp queue element.
-	 * Context: normal and in_interrupt().
+	 *
+	 * Context: cmd_queue_lock.
 	 */
 	void (*set_resp_elem_seq)(struct gcip_mailbox *mailbox, void *resp, u64 seq);
 
@@ -249,6 +267,7 @@
 	 * The lock can be a mutex lock or a spin lock. However, if @irqsave is considered and
 	 * "_irqsave" is used, it must be spin lock only.
 	 * The lock will be released by calling `release_wait_list_lock` callback.
+	 *
 	 * Context: normal and in_interrupt().
 	 */
 	void (*acquire_wait_list_lock)(struct gcip_mailbox *mailbox, bool irqsave,
@@ -257,28 +276,14 @@
 	 * Releases the lock of wait_list which is acquired by calling `acquire_wait_list_lock`.
 	 * If @irqsave is true, restores @flags from `acquire_wait_list_lock` to the irq state.
 	 * Or it can be ignored, if @irqsave was not considered in the `acquire_wait_list_lock`.
-	 * Context: normal and in_interrupt().
+	 *
+	 * Context: wait_list_lock.
 	 */
 	void (*release_wait_list_lock)(struct gcip_mailbox *mailbox, bool irqrestore,
 				       unsigned long flags);
 
 	/* Optional. */
 	/*
-	 * Waits for the cmd queue of @mailbox has a available space for putting the command. If
-	 * the queue has a space, returns 0. Otherwise, returns error as non-zero value. It depends
-	 * on the implementation details, but it is okay to return right away with error when the
-	 * queue is full. If this callback returns an error, `gcip_mailbox_send_cmd` function or
-	 * `gcip_mailbox_put_cmd` function will return that error too. This callback is called with
-	 * the `cmd_queue_lock` being held.
-	 *
-	 * Note: if this callback is NULL, it will simply check the fullness of cmd_queue and
-	 * return -EAGAIN error right away if it is full. Please refer the implementation of the
-	 * `gcip_mailbox_enqueue_cmd` function.
-	 *
-	 * Context: normal.
-	 */
-	int (*wait_for_cmd_queue_not_full)(struct gcip_mailbox *mailbox);
-	/*
 	 * This callback will be called before putting the @resp into @mailbox->wait_list and
 	 * putting @cmd of @resp into the command queue. After this callback returns, the consumer
 	 * is able to start processing it and the mailbox is going to wait for it. Therefore, this
@@ -290,15 +295,15 @@
 	 *
 	 * If @resp is synchronous, @awaiter will be NULL.
 	 *
-	 * Context: normal.
+	 * Context: cmd_queue_lock.
 	 */
 	int (*before_enqueue_wait_list)(struct gcip_mailbox *mailbox, void *resp,
 					struct gcip_mailbox_resp_awaiter *awaiter);
 	/*
 	 * This callback will be called after putting the @cmd to the command queue. It can be used
 	 * for triggering the doorbell. Returns 0 on success, or returns error code otherwise.
-	 * This is called with the `cmd_queue_lock` being held.
-	 * Context: normal.
+	 *
+	 * Context: cmd_queue_lock.
 	 */
 	int (*after_enqueue_cmd)(struct gcip_mailbox *mailbox, void *cmd);
 	/*
@@ -306,6 +311,7 @@
 	 * a signal to break up waiting consuming the response queue. This is called without
 	 * holding any locks.
 	 * - @num_resps: the number of fetched responses.
+	 *
 	 * Context: normal and in_interrupt().
 	 */
 	void (*after_fetch_resps)(struct gcip_mailbox *mailbox, u32 num_resps);
@@ -313,6 +319,7 @@
 	 * Before handling each fetched responses, this callback will be called. If this callback
 	 * is not defined or returns true, the mailbox will handle the @resp normally. If the @resp
 	 * should not be handled, returns false. This is called without holding any locks.
+	 *
 	 * Context: normal and in_interrupt().
 	 */
 	bool (*before_handle_resp)(struct gcip_mailbox *mailbox, const void *resp);
@@ -321,6 +328,7 @@
 	 * chip implementation. However, @awaiter should be released by calling the
 	 * `gcip_mailbox_release_awaiter` function when the kernel driver doesn't need
 	 * @awaiter anymore.
+	 *
 	 * Context: normal and in_interrupt().
 	 */
 	void (*handle_awaiter_arrived)(struct gcip_mailbox *mailbox,
@@ -330,6 +338,7 @@
 	 * implementation. However, @awaiter should be released by calling the
 	 * `gcip_mailbox_release_awaiter` function when the kernel driver doesn't need
 	 * @awaiter anymore. This is called without holding any locks.
+	 *
 	 * Context: normal and in_interrupt().
 	 */
 	void (*handle_awaiter_timedout)(struct gcip_mailbox *mailbox,
@@ -339,20 +348,22 @@
 	 * The @awaiter should be marked as unprocessable to make it not to be processed by
 	 * the `handle_awaiter_arrived` or `handle_awaiter_timedout` callbacks in race
 	 * conditions. Don't have to release @awaiter of this function by calling the
-	 * `gcip_mailbox_release_awaiter` function. It will be released internally. This is
-	 * called with the `wait_list_lock` being held.
-	 * Context: normal.
+	 * `gcip_mailbox_release_awaiter` function. It will be released internally.
+	 *
+	 * Context: wait_list_lock.
 	 */
 	void (*flush_awaiter)(struct gcip_mailbox *mailbox,
 			      struct gcip_mailbox_resp_awaiter *awaiter);
 	/*
 	 * Releases the @data which was passed to the `gcip_mailbox_put_cmd` function. This is
 	 * called without holding any locks.
+	 *
 	 * Context: normal and in_interrupt().
 	 */
 	void (*release_awaiter_data)(void *data);
 	/*
 	 * Checks if the block is off.
+	 *
 	 * Context: in_interrupt()
 	 */
 	bool (*is_block_off)(struct gcip_mailbox *mailbox);
@@ -362,11 +373,13 @@
 	 * be fetched from @cmd, @resp or @data passed to the `gcip_mailbox_put_cmd` function.
 	 * Therefore, this callback passes all of them not only @cmd. This can be called without
 	 * holding any locks.
+	 *
 	 * Context: normal.
 	 */
 	u32 (*get_cmd_timeout)(struct gcip_mailbox *mailbox, void *cmd, void *resp, void *data);
 	/*
 	 * Called when a command fails to be sent.
+	 *
 	 * Context: normal.
 	 */
 	void (*on_error)(struct gcip_mailbox *mailbox, int err);
diff --git a/gcip-kernel-driver/include/gcip/gcip-usage-stats.h b/gcip-kernel-driver/include/gcip/gcip-usage-stats.h
index a4e0cb4..4c400eb 100644
--- a/gcip-kernel-driver/include/gcip/gcip-usage-stats.h
+++ b/gcip-kernel-driver/include/gcip/gcip-usage-stats.h
@@ -60,7 +60,7 @@
 #define GCIP_USAGE_STATS_METRIC_SIZE_V1 20
 
 /* Max number of frequencies to support. */
-#define GCIP_USAGE_STATS_MAX_DVFS_FREQ_NUM 10
+#define GCIP_USAGE_STATS_MAX_DVFS_FREQ_NUM 25
 
 struct gcip_usage_stats_attr;
 
@@ -457,7 +457,7 @@
 	 * @idx will not exceed the number of default DVFS frequencies which is returned by the
 	 * `get_default_dvfs_freqs_num` operator.
 	 */
-	int (*get_default_dvfs_freq)(int idx, void *data);
+	unsigned int (*get_default_dvfs_freq)(int idx, void *data);
 };
 
 /* Structure manages the information of usage stats and device attributes. */
diff --git a/gcip-kernel-driver/include/gcip/iif/iif-manager.h b/gcip-kernel-driver/include/gcip/iif/iif-manager.h
deleted file mode 100644
index 3432e09..0000000
--- a/gcip-kernel-driver/include/gcip/iif/iif-manager.h
+++ /dev/null
@@ -1,47 +0,0 @@
-/* SPDX-License-Identifier: GPL-2.0-only */
-/*
- * GCIP-integrated IIF driver manager.
- *
- * Copyright (C) 2023 Google LLC
- */
-
-#ifndef __IIF_IIF_MANAGER_H__
-#define __IIF_IIF_MANAGER_H__
-
-#include <linux/idr.h>
-#include <linux/kref.h>
-#include <linux/of.h>
-
-#include <gcip/iif/iif-fence-table.h>
-
-/*
- * The structure overall data required by IIF driver such as fence table.
- *
- * Until we have stand-alone IIF driver, one of the IP drivers will initializes a manager by
- * the `iif_init` function and every IP driver will share it.
- */
-struct iif_manager {
-	/* Reference count of this instance. */
-	struct kref kref;
-	/* Fence ID pool. */
-	struct ida idp;
-	/* Fence table shared with the firmware. */
-	struct iif_fence_table fence_table;
-};
-
-/*
- * Initializes IIF driver and returns its manager. Its initial reference count is 1. It will map
- * the fence table by parsing the device tree via @np.
- *
- * The returned manager will be destroyed when its reference count becomes 0 by `iif_manager_put`
- * function.
- */
-struct iif_manager *iif_manager_init(const struct device_node *np);
-
-/* Increases the reference count of @mgr. */
-struct iif_manager *iif_manager_get(struct iif_manager *mgr);
-
-/* Decreases the reference count of @mgr and if it becomes 0, releases @mgr. */
-void iif_manager_put(struct iif_manager *mgr);
-
-#endif /* __IIF_IIF_MANAGER_H__ */
diff --git a/gcip-kernel-driver/include/gcip/iif/iif.h b/gcip-kernel-driver/include/gcip/iif/iif.h
deleted file mode 100644
index 0ccdb72..0000000
--- a/gcip-kernel-driver/include/gcip/iif/iif.h
+++ /dev/null
@@ -1,73 +0,0 @@
-/* SPDX-License-Identifier: GPL-2.0-only */
-/*
- * Defines the interface of the IIF driver.
- *
- * Copyright (C) 2023 Google LLC
- */
-
-#ifndef __IIF_IIF_H__
-#define __IIF_IIF_H__
-
-#include <linux/ioctl.h>
-#include <linux/types.h>
-
-/* Interface Version. */
-#define IIF_INTERFACE_VERSION_MAJOR 1
-#define IIF_INTERFACE_VERSION_MINOR 0
-
-#define IIF_IOCTL_BASE 'i'
-
-/* The ioctl number for the fence FDs will start from here. */
-#define IIF_FENCE_IOCTL_NUM_BASE 0x80
-
-/*
- * The max number of fences can be created per IP.
- * Increasing this value needs to increase the size of fence table.
- */
-#define IIF_NUM_FENCES_PER_IP 1024
-
-/*
- * Type of IPs.
- *
- * The order of IP must be matched with the firmware side because the fence ID will be assigned
- * according to the IP type.
- */
-enum iif_ip_type {
-	IIF_IP_DSP,
-	IIF_IP_TPU,
-	IIF_IP_GPU,
-	IIF_IP_NUM,
-
-	/* Reserve the number of IP type to expand the fence table easily in the future. */
-	IIF_IP_RESERVED = 16,
-};
-
-/*
- * ioctls for /dev/iif.
- * TODO(b/312161537): introduce ioctls once we have a standalone IIF driver.
- */
-
-struct iif_fence_get_information_ioctl {
-	/* The type of the signaler IP. (enum iif_ip_type) */
-	__u8 signaler_ip;
-	/* The number of total signalers. */
-	__u16 total_signalers;
-	/* The number of submitted signalers. */
-	__u16 submitted_signalers;
-	/* The number of signaled signalers. */
-	__u16 signaled_signalers;
-	/* The number of outstanding waiters. */
-	__u16 outstanding_waiters;
-	/* Reserved. */
-	__u8 reserved[7];
-};
-
-/*
- * ioctls for inter-IP fence FDs.
- */
-
-/* Returns the fence information. */
-#define IIF_FENCE_GET_INFORMATION \
-	_IOR(IIF_IOCTL_BASE, IIF_FENCE_IOCTL_NUM_BASE, struct iif_fence_get_information_ioctl)
-
-#endif /* __IIF_IIF_H__ */
diff --git a/gxp-client.c b/gxp-client.c
index a2a3dd0..dea2f61 100644
--- a/gxp-client.c
+++ b/gxp-client.c
@@ -141,6 +141,7 @@
 	down_write(&client->semaphore);
 
 	if (client->vd && client->vd->state != GXP_VD_OFF) {
+		gxp_vd_check_and_wait_for_debug_dump(client->vd);
 		down_write(&gxp->vd_semaphore);
 		gxp_vd_stop(client->vd);
 		up_write(&gxp->vd_semaphore);
@@ -413,6 +414,8 @@
 	if (client->vd->state == GXP_VD_UNAVAILABLE)
 		return;
 
+	gxp_vd_check_and_wait_for_debug_dump(client->vd);
+
 	down_write(&gxp->vd_semaphore);
 	gxp_vd_suspend(client->vd);
 	up_write(&gxp->vd_semaphore);
diff --git a/gxp-common-platform.c b/gxp-common-platform.c
index a2a5443..0524fe1 100644
--- a/gxp-common-platform.c
+++ b/gxp-common-platform.c
@@ -31,18 +31,19 @@
 #include "gxp-client.h"
 #include "gxp-config.h"
 #include "gxp-core-telemetry.h"
+#include "gxp-dci.h"
 #include "gxp-debug-dump.h"
 #include "gxp-dma-fence.h"
 #include "gxp-dma.h"
 #include "gxp-dmabuf.h"
 #include "gxp-domain-pool.h"
-#include "gxp-firmware.h"
 #include "gxp-firmware-data.h"
 #include "gxp-firmware-loader.h"
+#include "gxp-firmware.h"
 #include "gxp-internal.h"
 #include "gxp-lpm.h"
-#include "gxp-mailbox.h"
 #include "gxp-mailbox-driver.h"
+#include "gxp-mailbox.h"
 #include "gxp-mapping.h"
 #include "gxp-notification.h"
 #include "gxp-pm.h"
@@ -55,11 +56,6 @@
 #include <soc/google/tpu-ext.h>
 #endif
 
-#if GXP_USE_LEGACY_MAILBOX
-#include "gxp-mailbox-impl.h"
-#else
-#include "gxp-dci.h"
-#endif
 
 /* We will only have one gxp device */
 #define GXP_DEV_COUNT 1
@@ -590,206 +586,6 @@
 	return 0;
 }
 
-static int gxp_ioctl_etm_trace_start_command(struct gxp_client *client,
-					     struct gxp_etm_trace_start_ioctl __user *argp)
-{
-	struct gxp_dev *gxp = client->gxp;
-	struct gxp_etm_trace_start_ioctl ibuf;
-	int phys_core;
-	int ret = 0;
-
-	if (copy_from_user(&ibuf, argp, sizeof(ibuf)))
-		return -EFAULT;
-
-	ibuf.trace_ram_enable &= ETM_TRACE_LSB_MASK;
-	ibuf.atb_enable &= ETM_TRACE_LSB_MASK;
-
-	if (!ibuf.trace_ram_enable && !ibuf.atb_enable)
-		return -EINVAL;
-
-	if (!(ibuf.sync_msg_period == 0 ||
-	    (ibuf.sync_msg_period <= ETM_TRACE_SYNC_MSG_PERIOD_MAX &&
-	     ibuf.sync_msg_period >= ETM_TRACE_SYNC_MSG_PERIOD_MIN &&
-	     is_power_of_2(ibuf.sync_msg_period))))
-		return -EINVAL;
-
-	if (ibuf.pc_match_mask_length > ETM_TRACE_PC_MATCH_MASK_LEN_MAX)
-		return -EINVAL;
-
-	/* Caller must hold VIRTUAL_DEVICE wakelock */
-	down_read(&client->semaphore);
-
-	if (!check_client_has_available_vd_wakelock(
-		    client, "GXP_ETM_TRACE_START_COMMAND")) {
-		ret = -ENODEV;
-		goto out_unlock_client_semaphore;
-	}
-
-	down_read(&gxp->vd_semaphore);
-
-	phys_core =
-		gxp_vd_virt_core_to_phys_core(client->vd, ibuf.virtual_core_id);
-	if (phys_core < 0) {
-		dev_err(gxp->dev, "Trace start failed: Invalid virtual core id (%u)\n",
-			ibuf.virtual_core_id);
-		ret = -EINVAL;
-		goto out;
-	}
-
-out:
-	up_read(&gxp->vd_semaphore);
-out_unlock_client_semaphore:
-	up_read(&client->semaphore);
-
-	return ret;
-}
-
-static int gxp_ioctl_etm_trace_sw_stop_command(struct gxp_client *client, __u16 __user *argp)
-{
-	struct gxp_dev *gxp = client->gxp;
-	u16 virtual_core_id;
-	int phys_core;
-	int ret = 0;
-
-	if (copy_from_user(&virtual_core_id, argp, sizeof(virtual_core_id)))
-		return -EFAULT;
-
-	/* Caller must hold VIRTUAL_DEVICE wakelock */
-	down_read(&client->semaphore);
-
-	if (!check_client_has_available_vd_wakelock(
-		    client, "GXP_ETM_TRACE_SW_STOP_COMMAND")) {
-		ret = -ENODEV;
-		goto out_unlock_client_semaphore;
-	}
-
-	down_read(&gxp->vd_semaphore);
-
-	phys_core = gxp_vd_virt_core_to_phys_core(client->vd, virtual_core_id);
-	if (phys_core < 0) {
-		dev_err(gxp->dev, "Trace stop via software trigger failed: Invalid virtual core id (%u)\n",
-			virtual_core_id);
-		ret = -EINVAL;
-		goto out;
-	}
-out:
-	up_read(&gxp->vd_semaphore);
-out_unlock_client_semaphore:
-	up_read(&client->semaphore);
-
-	return ret;
-}
-
-static int gxp_ioctl_etm_trace_cleanup_command(struct gxp_client *client, __u16 __user *argp)
-{
-	struct gxp_dev *gxp = client->gxp;
-	u16 virtual_core_id;
-	int phys_core;
-	int ret = 0;
-
-	if (copy_from_user(&virtual_core_id, argp, sizeof(virtual_core_id)))
-		return -EFAULT;
-
-	/* Caller must hold VIRTUAL_DEVICE wakelock */
-	down_read(&client->semaphore);
-
-	if (!check_client_has_available_vd_wakelock(
-		    client, "GXP_ETM_TRACE_CLEANUP_COMMAND")) {
-		ret = -ENODEV;
-		goto out_unlock_client_semaphore;
-	}
-
-	down_read(&gxp->vd_semaphore);
-
-	phys_core = gxp_vd_virt_core_to_phys_core(client->vd, virtual_core_id);
-	if (phys_core < 0) {
-		dev_err(gxp->dev, "Trace cleanup failed: Invalid virtual core id (%u)\n",
-			virtual_core_id);
-		ret = -EINVAL;
-		goto out;
-	}
-out:
-	up_read(&gxp->vd_semaphore);
-out_unlock_client_semaphore:
-	up_read(&client->semaphore);
-
-	return ret;
-}
-
-static int gxp_ioctl_etm_get_trace_info_command(struct gxp_client *client,
-						struct gxp_etm_get_trace_info_ioctl __user *argp)
-{
-	struct gxp_dev *gxp = client->gxp;
-	struct gxp_etm_get_trace_info_ioctl ibuf;
-	int phys_core;
-	u32 *trace_header;
-	u32 *trace_data;
-	int ret = 0;
-
-	if (copy_from_user(&ibuf, argp, sizeof(ibuf)))
-		return -EFAULT;
-
-	if (ibuf.type > 1)
-		return -EINVAL;
-
-	/* Caller must hold VIRTUAL_DEVICE wakelock */
-	down_read(&client->semaphore);
-
-	if (!check_client_has_available_vd_wakelock(
-		    client, "GXP_ETM_GET_TRACE_INFO_COMMAND")) {
-		ret = -ENODEV;
-		goto out_unlock_client_semaphore;
-	}
-
-	down_read(&gxp->vd_semaphore);
-
-	phys_core = gxp_vd_virt_core_to_phys_core(client->vd, ibuf.virtual_core_id);
-	if (phys_core < 0) {
-		dev_err(gxp->dev, "Get trace info failed: Invalid virtual core id (%u)\n",
-			ibuf.virtual_core_id);
-		ret = -EINVAL;
-		goto out;
-	}
-
-	trace_header = kzalloc(GXP_TRACE_HEADER_SIZE, GFP_KERNEL);
-	if (!trace_header) {
-		ret = -ENOMEM;
-		goto out;
-	}
-
-	trace_data = kzalloc(GXP_TRACE_RAM_SIZE, GFP_KERNEL);
-	if (!trace_data) {
-		ret = -ENOMEM;
-		goto out_free_header;
-	}
-
-	if (copy_to_user((void __user *)ibuf.trace_header_addr, trace_header,
-			 GXP_TRACE_HEADER_SIZE)) {
-		ret = -EFAULT;
-		goto out_free_data;
-	}
-
-	if (ibuf.type == 1) {
-		if (copy_to_user((void __user *)ibuf.trace_data_addr,
-				 trace_data, GXP_TRACE_RAM_SIZE)) {
-			ret = -EFAULT;
-			goto out_free_data;
-		}
-	}
-
-out_free_data:
-	kfree(trace_data);
-out_free_header:
-	kfree(trace_header);
-
-out:
-	up_read(&gxp->vd_semaphore);
-out_unlock_client_semaphore:
-	up_read(&client->semaphore);
-
-	return ret;
-}
-
 #if HAS_TPU_EXT
 
 /*
@@ -1117,7 +913,7 @@
 	 * We intentionally don't call `gcip_pm_*` functions while holding @client->semaphore.
 	 *
 	 * As the `gcip_pm_put` function cancels KCI works synchronously and the KCI works may hold
-	 * @client->semaphore in some logics such as MCU FW crash handler, it can cause deadlock
+	 * @client->semaphore in some logic such as MCU FW crash handler, it can cause deadlock
 	 * issues potentially if we call `gcip_pm_put` after holding @client->semaphore.
 	 *
 	 * Therefore, we decided to decouple calling the `gcip_pm_put` function from holding
@@ -1733,18 +1529,6 @@
 	case GXP_ALLOCATE_VIRTUAL_DEVICE:
 		ret = gxp_ioctl_allocate_vd(client, argp);
 		break;
-	case GXP_ETM_TRACE_START_COMMAND:
-		ret = gxp_ioctl_etm_trace_start_command(client, argp);
-		break;
-	case GXP_ETM_TRACE_SW_STOP_COMMAND:
-		ret = gxp_ioctl_etm_trace_sw_stop_command(client, argp);
-		break;
-	case GXP_ETM_TRACE_CLEANUP_COMMAND:
-		ret = gxp_ioctl_etm_trace_cleanup_command(client, argp);
-		break;
-	case GXP_ETM_GET_TRACE_INFO_COMMAND:
-		ret = gxp_ioctl_etm_get_trace_info_command(client, argp);
-		break;
 	case GXP_MAP_TPU_MBX_QUEUE:
 		ret = gxp_ioctl_map_tpu_mbx_queue(client, argp);
 		break;
@@ -2236,19 +2020,14 @@
 		dev_err(dev, "Failed to create mailbox manager: %d\n", ret);
 		goto err_dma_exit;
 	}
-	if (gxp_is_direct_mode(gxp)) {
-#if GXP_USE_LEGACY_MAILBOX
-		gxp_mailbox_init(gxp->mailbox_mgr);
-#else
+	if (gxp_is_direct_mode(gxp))
 		gxp_dci_init(gxp->mailbox_mgr);
-#endif
-	}
 
 #if IS_ENABLED(CONFIG_SUBSYSTEM_COREDUMP)
 	ret = gxp_debug_dump_init(gxp, &gxp_sscd_dev, &gxp_sscd_pdata);
 #else
 	ret = gxp_debug_dump_init(gxp, NULL, NULL);
-#endif  // !CONFIG_SUBSYSTEM_COREDUMP
+#endif  /* !CONFIG_SUBSYSTEM_COREDUMP */
 	if (ret)
 		dev_warn(dev, "Failed to initialize debug dump\n");
 
@@ -2330,11 +2109,6 @@
 		if (ret)
 			goto err_dma_fence_destroy;
 	}
-	/*
-	 * We only know where the system config region is after after_probe is
-	 * done so this can't be called earlier.
-	 */
-	gxp_fw_data_populate_system_config(gxp);
 
 	ret = gxp_device_add(gxp);
 	if (ret)
diff --git a/gxp-config.h b/gxp-config.h
index 56a84bf..de6bdfc 100644
--- a/gxp-config.h
+++ b/gxp-config.h
@@ -32,10 +32,6 @@
 
 #define SYNC_BARRIER_COUNT 16
 
-#ifndef GXP_USE_LEGACY_MAILBOX
-#define GXP_USE_LEGACY_MAILBOX 0
-#endif
-
 #ifndef GXP_HAS_LAP
 #define GXP_HAS_LAP 1
 #endif
diff --git a/gxp-core-telemetry.c b/gxp-core-telemetry.c
index 096eb53..2e8339d 100644
--- a/gxp-core-telemetry.c
+++ b/gxp-core-telemetry.c
@@ -394,7 +394,7 @@
 	unsigned long orig_pgoff = vma->vm_pgoff;
 
 	if (!phys) {
-		dev_err(gxp->dev, "Secure buffer is not mapped.\n");
+		dev_warn(gxp->dev, "Secure buffer is not mapped.\n");
 		return -ENODATA;
 	}
 
diff --git a/gxp-dci.c b/gxp-dci.c
index 5f658f8..29da8d2 100644
--- a/gxp-dci.c
+++ b/gxp-dci.c
@@ -2,7 +2,7 @@
 /*
  * Implementation of DCI (Direct Command Interface) using mailbox.
  *
- * Copyright (C) 2022 Google LLC
+ * Copyright (C) 2022-2024 Google LLC
  */
 
 #include <linux/kthread.h>
@@ -11,6 +11,7 @@
 #include <uapi/linux/sched/types.h>
 
 #include "gxp-dci.h"
+#include "gxp-debug-dump.h"
 #include "gxp-dma.h"
 #include "gxp-mailbox-driver.h"
 #include "gxp-pm.h"
@@ -281,13 +282,14 @@
 	spin_unlock_irqrestore(async_resp->dest_queue_lock, flags);
 }
 
-static void
-gxp_dci_handle_awaiter_timedout(struct gcip_mailbox *mailbox,
-				struct gcip_mailbox_resp_awaiter *awaiter)
+static void gxp_dci_handle_awaiter_timedout(struct gcip_mailbox *mailbox,
+					    struct gcip_mailbox_resp_awaiter *awaiter)
 {
 	struct gxp_mailbox *mbx = mailbox->data;
 	struct gxp_dci_async_response *async_resp = awaiter->data;
 	struct gxp_dci_response *resp = &async_resp->resp;
+	struct gxp_dev *gxp = mbx->gxp;
+	struct gxp_virtual_device *vd;
 	unsigned long flags;
 
 	/*
@@ -304,8 +306,21 @@
 		list_add_tail(&async_resp->list_entry, async_resp->dest_queue);
 		spin_unlock_irqrestore(async_resp->dest_queue_lock, flags);
 
-		gxp_pm_update_requested_power_states(
-			mbx->gxp, async_resp->requested_states, off_states);
+		/* Get hold of the virtual device. */
+		down_read(&gxp->vd_semaphore);
+		if (gxp->core_to_vd[mbx->core_id]) {
+			vd = gxp->core_to_vd[mbx->core_id];
+			/*
+			 * Response timeout most probably would happen because of core stall. Check
+			 * if forced debug dump can be requested to the participating cores for the
+			 * current vd.
+			 */
+			gxp_debug_dump_send_forced_debug_dump_request(gxp, vd);
+		}
+		up_read(&gxp->vd_semaphore);
+
+		gxp_pm_update_requested_power_states(mbx->gxp, async_resp->requested_states,
+						     off_states);
 
 		if (async_resp->eventfd) {
 			gxp_eventfd_signal(async_resp->eventfd);
@@ -337,7 +352,6 @@
 }
 
 static const struct gcip_mailbox_ops gxp_dci_gcip_mbx_ops = {
-	.get_cmd_queue_head = gxp_mailbox_gcip_ops_get_cmd_queue_head,
 	.get_cmd_queue_tail = gxp_mailbox_gcip_ops_get_cmd_queue_tail,
 	.inc_cmd_queue_tail = gxp_mailbox_gcip_ops_inc_cmd_queue_tail,
 	.acquire_cmd_queue_lock = gxp_mailbox_gcip_ops_acquire_cmd_queue_lock,
@@ -460,25 +474,31 @@
 		.cmd_elem_size = sizeof(struct gxp_dci_command),
 		.resp_elem_size = sizeof(struct gxp_dci_response),
 	};
+	struct gxp_mailbox *mbx;
 
 	dci = kzalloc(sizeof(*dci), GFP_KERNEL);
 	if (!dci)
 		return ERR_PTR(-ENOMEM);
 	mbx_args.data = dci;
 
-	dci->mbx = gxp_mailbox_alloc(mgr, vd, virt_core, core_id, &mbx_args);
-	if (IS_ERR(dci->mbx))
+	mbx = gxp_mailbox_alloc(mgr, vd, virt_core, core_id, &mbx_args);
+	if (IS_ERR(mbx)) {
 		kfree(dci);
-	else
-		gxp_mailbox_generate_device_interrupt(dci->mbx, BIT(0));
+		return mbx;
+	}
+	dci->mbx = mbx;
+	gxp_mailbox_reinit(dci->mbx);
+	gxp_mailbox_generate_device_interrupt(dci->mbx, BIT(0));
 
-	return dci->mbx;
+	return mbx;
 }
 
 void gxp_dci_release(struct gxp_mailbox_manager *mgr,
 		     struct gxp_virtual_device *vd, uint virt_core,
 		     struct gxp_mailbox *mbx)
 {
+	/* Frees dci. */
+	kfree(mbx->data);
 	gxp_mailbox_release(mgr, vd, virt_core, mbx);
 }
 
diff --git a/gxp-dci.h b/gxp-dci.h
index 09eaed4..7367939 100644
--- a/gxp-dci.h
+++ b/gxp-dci.h
@@ -2,7 +2,7 @@
 /*
  * Implementation of DCI (Direct Command Interface) using mailbox.
  *
- * Copyright (C) 2022 Google LLC
+ * Copyright (C) 2022-2024 Google LLC
  */
 
 #ifndef __GXP_DCI_H__
diff --git a/gxp-debug-dump.c b/gxp-debug-dump.c
index 381438e..e19bbc3 100644
--- a/gxp-debug-dump.c
+++ b/gxp-debug-dump.c
@@ -23,16 +23,14 @@
 #include "gxp-debug-dump.h"
 #include "gxp-dma.h"
 #include "gxp-doorbell.h"
-#include "gxp-firmware.h"
 #include "gxp-firmware-data.h"
 #include "gxp-firmware-loader.h"
+#include "gxp-firmware.h"
 #include "gxp-host-device-structs.h"
 #include "gxp-internal.h"
 #include "gxp-lpm.h"
 #include "gxp-mailbox-driver.h"
 #include "gxp-mapping.h"
-#include "gxp-mcu.h"
-#include "gxp-mcu-telemetry.h"
 #include "gxp-notification.h"
 #include "gxp-pm.h"
 #include "gxp-vd.h"
@@ -41,6 +39,11 @@
 #include <linux/platform_data/sscoredump.h>
 #endif
 
+#if GXP_HAS_MCU
+#include "gxp-mcu-telemetry.h"
+#include "gxp-mcu.h"
+#endif /* GXP_HAS_MCU */
+
 #define SSCD_MSG_LENGTH 64
 
 #define GXP_SYNC_BARRIER_STRIDE (GXP_REG_SYNC_BARRIER_1 - GXP_REG_SYNC_BARRIER_0)
@@ -48,15 +51,6 @@
 #define DEBUG_DUMP_MEMORY_SIZE 0x400000 /* size in bytes */
 
 /*
- * The minimum wait time in millisecond to be enforced between two successive calls to the SSCD
- * module to prevent the overwrite of the previous generated core dump files. SSCD module generates
- * the files whose name are at second precision i.e.
- * crashinfo_<SUBSYSTEM_NAME>_<%Y-%m-%d_%H-%M-%S>.txt and
- * coredump_<SUBSYSTEM_NAME>_<%Y-%m-%d_%H-%M-%S>.bin.
- */
-#define SSCD_REPORT_WAIT_TIME (1000ULL)
-
-/*
  * CORE_FIRMWARE_RW_STRIDE & CORE_FIRMWARE_RW_ADDR must match with their
  * values defind in core firmware image config.
  */
@@ -613,6 +607,37 @@
 	core_dump_header->core_header.dump_available = 0;
 }
 
+void gxp_debug_dump_send_forced_debug_dump_request(struct gxp_dev *gxp,
+						   struct gxp_virtual_device *vd)
+{
+	uint phys_core;
+	uint generate_debug_dump;
+	uint debug_dump_generated;
+
+	lockdep_assert_held(&gxp->vd_semaphore);
+
+	for (phys_core = 0; phys_core < GXP_NUM_CORES; phys_core++) {
+		if (!(vd->core_list & BIT(phys_core)))
+			continue;
+
+		generate_debug_dump = gxp_firmware_get_generate_debug_dump(gxp, vd, phys_core);
+		debug_dump_generated = gxp_firmware_get_debug_dump_generated(gxp, vd, phys_core);
+		/*
+		 * If neither the core has generated the debug dump nor has been requested to
+		 * generate the forced debug dump.
+		 */
+		if (!debug_dump_generated && !generate_debug_dump) {
+			if (!gxp_lpm_is_powered(gxp, CORE_TO_PSM(phys_core))) {
+				dev_dbg(gxp->dev, "Core%u not powered on.\n", phys_core);
+				continue;
+			}
+			/* Send the interrupt to the core for requesting the forced debug dump. */
+			gxp_firmware_set_generate_debug_dump(gxp, vd, phys_core, 1);
+			gxp_notification_send(gxp, phys_core, CORE_NOTIF_GENERATE_DEBUG_DUMP);
+		}
+	}
+}
+
 /*
  * Caller must make sure that gxp->debug_dump_mgr->common_dump and
  * gxp->debug_dump_mgr->core_dump are not NULL.
@@ -790,10 +815,12 @@
 	uint core_id = debug_dump_work->core_id;
 	struct gxp_dev *gxp = debug_dump_work->gxp;
 	struct gxp_virtual_device *vd = NULL;
+	int old_core_dump_generated_list;
 
 	down_read(&gxp->vd_semaphore);
 	if (gxp->core_to_vd[core_id]) {
 		vd = gxp_vd_get(gxp->core_to_vd[core_id]);
+		gxp_debug_dump_send_forced_debug_dump_request(gxp, vd);
 	} else {
 		dev_warn(gxp->dev, "debug dump failed for null vd on core %d.", core_id);
 		up_read(&gxp->vd_semaphore);
@@ -810,6 +837,15 @@
 
 	gxp_generate_debug_dump(gxp, core_id, vd);
 
+	/* Update the debug dump processing status for the current core. */
+	old_core_dump_generated_list = atomic_fetch_or(BIT(core_id), &vd->core_dump_generated_list);
+	/*
+	 * Event the wait queue in case debug dump processing has been finished for all the
+	 * running cores for the vd.
+	 */
+	if ((old_core_dump_generated_list | BIT(core_id)) == vd->core_list)
+		wake_up(&vd->finished_dump_processing_waitq);
+
 	mutex_unlock(&vd->debug_dump_lock);
 	gxp_vd_put(vd);
 }
diff --git a/gxp-debug-dump.h b/gxp-debug-dump.h
index 5bcec03..e03155e 100644
--- a/gxp-debug-dump.h
+++ b/gxp-debug-dump.h
@@ -39,6 +39,15 @@
 	(GXP_NUM_COMMON_SEGMENTS + GXP_NUM_CORE_SEGMENTS + GXP_NUM_CORE_DATA_SEGMENTS + \
 	 GXP_NUM_BUFFER_MAPPINGS)
 
+/*
+ * The minimum wait time in millisecond to be enforced between two successive calls to the SSCD
+ * module to prevent the overwrite of the previous generated core dump files. SSCD module generates
+ * the files whose name are at second precision i.e.
+ * crashinfo_<SUBSYSTEM_NAME>_<%Y-%m-%d_%H-%M-%S>.txt and
+ * coredump_<SUBSYSTEM_NAME>_<%Y-%m-%d_%H-%M-%S>.bin.
+ */
+#define SSCD_REPORT_WAIT_TIME (1000ULL)
+
 #define GXP_Q7_ICACHE_SIZE 131072 /* I-cache size in bytes */
 #define GXP_Q7_ICACHE_LINESIZE 64 /* I-cache line size in bytes */
 #define GXP_Q7_ICACHE_WAYS 4
@@ -238,6 +247,22 @@
 void gxp_debug_dump_invalidate_segments(struct gxp_dev *gxp, uint32_t core_id);
 
 /**
+ * gxp_debug_dump_send_forced_debug_dump_request() - Sends the forced debug dump request to the
+ *                                                   running cores of the given vd.
+ * @gxp: The GXP device to obtain the handler for.
+ * @vd: vd whose cores to send the forced debug dump request.
+ *
+ * The caller must hold @gxp->vd_semaphore.
+ *
+ * In case of single core VD no forced debug dump request will be sent since there will be no other
+ * core. For multicore VD, forced debug dump request will be sent to other cores only during the
+ * first worker thread run. For successive worker thread runs since the forced request was already
+ * sent, it will not be sent again.
+ */
+void gxp_debug_dump_send_forced_debug_dump_request(struct gxp_dev *gxp,
+						   struct gxp_virtual_device *vd);
+
+/**
  * gxp_debug_dump_process_dump_mcu_mode() - Checks and process the debug dump
  *                                          for cores from core_list.
  * @gxp: The GXP device to obtain the handler for
diff --git a/gxp-dma-iommu.c b/gxp-dma-iommu.c
index 32c86f2..77537b9 100644
--- a/gxp-dma-iommu.c
+++ b/gxp-dma-iommu.c
@@ -38,29 +38,26 @@
 	switch (fault->type) {
 	case IOMMU_FAULT_DMA_UNRECOV:
 		dev_err(gxp->dev, "Unrecoverable IOMMU fault!\n");
+		dev_err(gxp->dev, "reason = %08X\n", fault->event.reason);
+		dev_err(gxp->dev, "flags = %08X\n", fault->event.flags);
+		dev_err(gxp->dev, "pasid = %08X\n", fault->event.pasid);
+		dev_err(gxp->dev, "perm = %08X\n", fault->event.perm);
+		dev_err(gxp->dev, "addr = %llX\n", fault->event.addr);
+		dev_err(gxp->dev, "fetch_addr = %llX\n", fault->event.fetch_addr);
 		break;
 	case IOMMU_FAULT_PAGE_REQ:
 		dev_err(gxp->dev, "IOMMU page request fault!\n");
+		dev_err(gxp->dev, "flags = %08X\n", fault->prm.flags);
+		dev_err(gxp->dev, "pasid = %08X\n", fault->prm.pasid);
+		dev_err(gxp->dev, "grpid = %08X\n", fault->prm.grpid);
+		dev_err(gxp->dev, "perm = %08X\n", fault->prm.perm);
+		dev_err(gxp->dev, "addr = %llX\n", fault->prm.addr);
 		break;
 	default:
 		dev_err(gxp->dev, "Unexpected IOMMU fault type (%d)\n",
 			fault->type);
-		return -EAGAIN;
 	}
 
-	/*
-	 * Normally the iommu driver should fill out the `event` struct for
-	 * unrecoverable errors, and the `prm` struct for page request faults.
-	 * The SysMMU driver, instead, always fills out the `event` struct.
-	 *
-	 * Note that the `fetch_addr` and `perm` fields are never filled out,
-	 * so we skip printing them.
-	 */
-	dev_err(gxp->dev, "reason = %08X\n", fault->event.reason);
-	dev_err(gxp->dev, "flags = %08X\n", fault->event.flags);
-	dev_err(gxp->dev, "pasid = %08X\n", fault->event.pasid);
-	dev_err(gxp->dev, "addr = %llX\n", fault->event.addr);
-
 	// Tell the IOMMU driver to carry on
 	return -EAGAIN;
 }
diff --git a/gxp-dmabuf.c b/gxp-dmabuf.c
index c51cd0b..a641ab7 100644
--- a/gxp-dmabuf.c
+++ b/gxp-dmabuf.c
@@ -25,12 +25,12 @@
 	dma_addr_t device_address = mapping->gcip_mapping->device_address;
 	size_t size = mapping->gcip_mapping->size;
 
-	trace_gxp_mapping_destroy_start(device_address, size);
+	trace_gxp_dmabuf_mapping_destroy_start(device_address, size);
 
 	gcip_iommu_mapping_unmap(mapping->gcip_mapping);
 	kfree(mapping);
 
-	trace_gxp_mapping_destroy_end(device_address, size);
+	trace_gxp_dmabuf_mapping_destroy_end(device_address, size);
 }
 
 struct gxp_mapping *gxp_dmabuf_map(struct gxp_dev *gxp, struct gcip_iommu_reserve_manager *mgr,
diff --git a/gxp-firmware-data.c b/gxp-firmware-data.c
index 4d1fd10..b5db087 100644
--- a/gxp-firmware-data.c
+++ b/gxp-firmware-data.c
@@ -38,16 +38,18 @@
 	 * is provided to all VDs and all cores. This is the R/W section.
 	 */
 	struct gxp_system_descriptor_rw *sys_desc_rw;
+	/* The size of system config region. */
+	u32 sys_cfg_size;
 };
 
 /*
- * Here assumes sys_cfg contains gxp_system_descriptor_ro in the first page and
- * gxp_system_descriptor_rw in the second page.
+ * Here assumes sys_cfg contains gxp_system_descriptor_ro in the first half and
+ * gxp_system_descriptor_rw in the second half.
  */
 static void set_system_cfg_region(struct gxp_dev *gxp, void *sys_cfg)
 {
 	struct gxp_system_descriptor_ro *des_ro = sys_cfg;
-	struct gxp_system_descriptor_rw *des_rw = sys_cfg + SZ_4K;
+	struct gxp_system_descriptor_rw *des_rw = sys_cfg + gxp->data_mgr->sys_cfg_size / 2;
 	struct gxp_core_telemetry_descriptor *descriptor =
 		&gxp->data_mgr->core_telemetry_desc;
 	struct telemetry_descriptor_ro *ro;
@@ -226,10 +228,11 @@
 	struct gxp_mapped_resource res = gxp_fw_data_resource(gxp);
 
 	/* Use the end of the shared region for system cfg. */
-	return res.vaddr + res.size - GXP_FW_DATA_SYSCFG_SIZE;
+	return res.vaddr + res.size - gxp->data_mgr->sys_cfg_size;
 }
 
-void gxp_fw_data_populate_system_config(struct gxp_dev *gxp)
+void gxp_fw_data_populate_system_config(struct gxp_dev *gxp, u32 sys_cfg_size)
 {
+	gxp->data_mgr->sys_cfg_size = sys_cfg_size;
 	set_system_cfg_region(gxp, gxp_fw_data_system_cfg(gxp));
 }
diff --git a/gxp-firmware-data.h b/gxp-firmware-data.h
index 028e610..e92f35b 100644
--- a/gxp-firmware-data.h
+++ b/gxp-firmware-data.h
@@ -15,8 +15,6 @@
 #include "gxp-internal.h"
 #include "gxp-vd.h"
 
-#define GXP_FW_DATA_SYSCFG_SIZE SZ_8K
-
 enum gxp_fw_data_protocol {
 	/* Use the per-VD configuration region. */
 	FW_DATA_PROTOCOL_PER_VD_CONFIG = 2,
@@ -109,13 +107,14 @@
 void *gxp_fw_data_system_cfg(struct gxp_dev *gxp);
 
 /**
- * gxp_fw_data_populate_system_config() - Populate settings onto firmware system
- *                                        config region.
- * @gxp: The GXP device
+ * gxp_fw_data_populate_system_config() - Populate settings onto firmware system config region.
+ * @gxp: The GXP device.
+ * @sys_cfg_size: The size of system config region.
  *
- * This function is expected to be called after "after_probe" in the probe
- * procedure since it uses gxp_fw_data_system_cfg().
+ * This function is expected to be called after firmware loaded to know the correct size of the
+ * system config.
+ * The size will be recorded in gxp->data_mgr.
  */
-void gxp_fw_data_populate_system_config(struct gxp_dev *gxp);
+void gxp_fw_data_populate_system_config(struct gxp_dev *gxp, u32 sys_cfg_size);
 
 #endif /* __GXP_FIRMWARE_DATA_H__ */
diff --git a/gxp-firmware-loader.c b/gxp-firmware-loader.c
index 1a145a3..6807484 100644
--- a/gxp-firmware-loader.c
+++ b/gxp-firmware-loader.c
@@ -12,6 +12,7 @@
 #include <gcip/gcip-image-config.h>
 
 #include "gxp-config.h"
+#include "gxp-firmware-data.h"
 #include "gxp-firmware-loader.h"
 #include "gxp-firmware.h"
 #include "gxp-internal.h"
@@ -246,13 +247,27 @@
 int gxp_firmware_loader_load_if_needed(struct gxp_dev *gxp)
 {
 	struct gxp_firmware_loader_manager *mgr = gxp->fw_loader_mgr;
+	resource_size_t addr_size;
 	int ret = 0;
 
 	mutex_lock(&mgr->lock);
 	if (mgr->is_loaded)
-		goto out;
+		goto out_unlock;
 	ret = gxp_firmware_loader_load_locked(gxp);
-out:
+	if (ret)
+		goto out_unlock;
+	/*
+	 * This should be done only after the firmware is successfully loaded because the size of
+	 * system config region is required.
+	 */
+	addr_size = mgr->core_img_cfg.iommu_mappings[SYS_CFG_REGION_IDX].image_config_value;
+	mutex_unlock(&mgr->lock);
+
+	gxp_fw_data_populate_system_config(gxp, gcip_config_to_size(addr_size));
+
+	return 0;
+
+out_unlock:
 	mutex_unlock(&mgr->lock);
 	return ret;
 }
diff --git a/gxp-firmware.c b/gxp-firmware.c
index f9049a0..9c0addc 100644
--- a/gxp-firmware.c
+++ b/gxp-firmware.c
@@ -334,6 +334,8 @@
 	core_cfg->top_access_ok = 0;
 	core_cfg->boot_status = GXP_BOOT_STATUS_NONE;
 	gxp_firmware_set_boot_mode(gxp, vd, core, GXP_BOOT_MODE_COLD_BOOT);
+	gxp_firmware_set_debug_dump_generated(gxp, vd, core, 0);
+	gxp_firmware_set_generate_debug_dump(gxp, vd, core, 0);
 }
 
 static int gxp_firmware_handshake(struct gxp_dev *gxp,
@@ -1008,7 +1010,7 @@
 	gxp_notification_unregister_handler(gxp, phys_core,
 					    HOST_NOTIF_CORE_TELEMETRY_STATUS);
 
-	if (gxp->mailbox_mgr->release_mailbox) {
+	if (gxp->mailbox_mgr->release_mailbox && gxp->mailbox_mgr->mailboxes[phys_core]) {
 		gxp->mailbox_mgr->release_mailbox(gxp->mailbox_mgr, vd, virt_core,
 						  gxp->mailbox_mgr->mailboxes[phys_core]);
 		dev_notice(gxp->dev, "Mailbox %u released\n", phys_core);
@@ -1150,8 +1152,43 @@
 	core_cfg->boot_status = status;
 }
 
-u32 gxp_firmware_get_boot_status(struct gxp_dev *gxp,
-				 struct gxp_virtual_device *vd, uint core)
+u32 gxp_firmware_get_generate_debug_dump(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					 uint core)
+{
+	struct gxp_host_control_region *core_cfg;
+
+	core_cfg = get_scratchpad_base(gxp, vd, core);
+	return core_cfg->generate_debug_dump;
+}
+
+void gxp_firmware_set_generate_debug_dump(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					  uint core, u32 generate_debug_dump)
+{
+	struct gxp_host_control_region *core_cfg;
+
+	core_cfg = get_scratchpad_base(gxp, vd, core);
+	core_cfg->generate_debug_dump = generate_debug_dump;
+}
+
+u32 gxp_firmware_get_debug_dump_generated(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					  uint core)
+{
+	struct gxp_host_control_region *core_cfg;
+
+	core_cfg = get_scratchpad_base(gxp, vd, core);
+	return core_cfg->debug_dump_generated;
+}
+
+void gxp_firmware_set_debug_dump_generated(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					   uint core, u32 debug_dump_generated)
+{
+	struct gxp_host_control_region *core_cfg;
+
+	core_cfg = get_scratchpad_base(gxp, vd, core);
+	core_cfg->debug_dump_generated = debug_dump_generated;
+}
+
+u32 gxp_firmware_get_boot_status(struct gxp_dev *gxp, struct gxp_virtual_device *vd, uint core)
 {
 	struct gxp_host_control_region *core_cfg;
 
diff --git a/gxp-firmware.h b/gxp-firmware.h
index 40ee22b..85436e9 100644
--- a/gxp-firmware.h
+++ b/gxp-firmware.h
@@ -22,8 +22,6 @@
 	CORE_CFG_REGION_IDX,
 	VD_CFG_REGION_IDX,
 	SYS_CFG_REGION_IDX,
-	/* TODO(b/299037074): Remove core's accesses to LPM. */
-	REMOTE_LPM_IDX = 7,
 };
 
 struct gxp_firmware_manager {
@@ -126,6 +124,34 @@
 				 struct gxp_virtual_device *vd, uint core);
 
 /*
+ * Returns the `generate_debug_dump` flag from the given virtual device's shared host-core region
+ * for the specified core.
+ */
+u32 gxp_firmware_get_generate_debug_dump(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					 uint core);
+
+/*
+ * Sets the `generate_debug_dump` flag for the given virtual device's shared host-core region
+ * for the specified core.
+ */
+void gxp_firmware_set_generate_debug_dump(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					  uint core, u32 generate_debug_dump);
+
+/*
+ * Returns the `debug_dump_generated` flag from the given virtual device's shared host-core region
+ * for the specified core.
+ */
+u32 gxp_firmware_get_debug_dump_generated(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					  uint core);
+
+/*
+ * Sets the `generate_debug_dump` flag from the given virtual device's shared host-core region
+ * for the specified core.
+ */
+void gxp_firmware_set_debug_dump_generated(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					   uint core, u32 debug_dump_generated);
+
+/*
  * Disable external interrupts to core.
  */
 void gxp_firmware_disable_ext_interrupts(struct gxp_dev *gxp, uint core);
diff --git a/gxp-gsa.h b/gxp-gsa.h
index 6fcfb6c..3ff544b 100644
--- a/gxp-gsa.h
+++ b/gxp-gsa.h
@@ -2,7 +2,7 @@
 /*
  * Wrapper for GSA related APIs.
  *
- * Copyright (C) 2023 Google LLC
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #ifndef __GXP_GSA_H__
diff --git a/gxp-gsx01-mailbox.c b/gxp-gsx01-mailbox.c
index db4c1aa..22d06f6 100644
--- a/gxp-gsx01-mailbox.c
+++ b/gxp-gsx01-mailbox.c
@@ -18,19 +18,14 @@
 
 #include "gxp-mailbox-driver.c"
 
-static u32 csr_read(struct gxp_mailbox *mailbox, uint reg_offset)
+static u32 gxp_mailbox_get_interrupt_status(struct gxp_mailbox *mailbox)
 {
-	return readl(mailbox->csr_reg_base + reg_offset);
-}
-
-static void csr_write(struct gxp_mailbox *mailbox, uint reg_offset, u32 value)
-{
-	writel(value, mailbox->csr_reg_base + reg_offset);
+	return gxp_mailbox_csr_read(mailbox, MBOX_INTMSR1_OFFSET);
 }
 
 void gxp_mailbox_reset_hw(struct gxp_mailbox *mailbox)
 {
-	csr_write(mailbox, MBOX_MCUCTLR_OFFSET, 1);
+	gxp_mailbox_csr_write(mailbox, MBOX_MCUCTLR_OFFSET, 1);
 }
 
 /* Interrupt to signal a response from the device to host */
@@ -78,23 +73,18 @@
 	 */
 	wmb();
 
-	csr_write(mailbox, MBOX_INTGR0_OFFSET, int_mask);
+	gxp_mailbox_csr_write(mailbox, MBOX_INTGR0_OFFSET, int_mask);
 }
 
 void gxp_mailbox_clear_interrupts(struct gxp_mailbox *mailbox, u32 intr_bits)
 {
-	csr_write(mailbox, MBOX_INTCR1_OFFSET, intr_bits);
+	gxp_mailbox_csr_write(mailbox, MBOX_INTCR1_OFFSET, intr_bits);
 }
 
 void gxp_mailbox_enable_interrupt(struct gxp_mailbox *mailbox)
 {
 }
 
-u32 gxp_mailbox_get_interrupt_status(struct gxp_mailbox *mailbox)
-{
-	return csr_read(mailbox, MBOX_INTMSR1_OFFSET);
-}
-
 int gxp_mailbox_wait_for_device_mailbox_init(struct gxp_mailbox *mailbox)
 {
 	return 0;
diff --git a/gxp-gsx01-ssmt.c b/gxp-gsx01-ssmt.c
index 190316d..4efdf78 100644
--- a/gxp-gsx01-ssmt.c
+++ b/gxp-gsx01-ssmt.c
@@ -2,7 +2,7 @@
 /*
  * GXP SSMT driver.
  *
- * Copyright (C) 2022 Google LLC
+ * Copyright (C) 2022-2024 Google LLC
  */
 
 #include <linux/platform_device.h>
diff --git a/gxp-gsx01-ssmt.h b/gxp-gsx01-ssmt.h
index 16eb82a..1fed70a 100644
--- a/gxp-gsx01-ssmt.h
+++ b/gxp-gsx01-ssmt.h
@@ -2,7 +2,7 @@
 /*
  * GXP SSMT driver.
  *
- * Copyright (C) 2022 Google LLC
+ * Copyright (C) 2022-2024 Google LLC
  */
 
 #ifndef __GXP_SSMT_H__
diff --git a/gxp-host-device-structs.h b/gxp-host-device-structs.h
index 8fe8a78..e2c38fa 100644
--- a/gxp-host-device-structs.h
+++ b/gxp-host-device-structs.h
@@ -10,20 +10,41 @@
  * Since the header is shared with the FW, it cannot rely on kernel-specific
  * headers or data structures.
  *
+ * Note: since the structures are shared across entities (cores, kernel, MCU),
+ * and they may change as the code is running (since they are used to build the
+ * system synchronization primitives), care should be taken to prevent the
+ * compiler from optimizing reads to such structures (for example, when polling
+ * on a value of a member in a struct; waiting for another core to change it).
+ * To achieve that, it's generally advised to access these structures as
+ * volatile to forbid the compiler from caching field values in CPU registers.
  */
 
 #ifndef __GXP_HOST_DEVICE_STRUCTURES_H__
 #define __GXP_HOST_DEVICE_STRUCTURES_H__
 
+/*
+ * The structure currently doesn't change its layout between the different
+ * HW generations; thus max core count is kept the same for all.
+ */
 #define MAX_NUM_CORES 4
 
-/* The number of physical doorbells and sync barriers allocated to each VD */
+/*
+ * The number of physical doorbells, sync barriers and timers allocated to each
+ * VD. The HW supports a total of 16 sync barriers; divided across 4 active VDs.
+ * Some are reserved for system use such as the UART sync barrier. It also
+ * supports a total of 32 doorbells; divided across 4 active VDs. Some are
+ * reserved for system use; such as the 4 doorbells for waking up cores.
+ * 8 timers are also supported by the HW. Some are reserved for system use; such
+ * as timers 0-3 and timer 7.
+ */
 #define GXP_NUM_DOORBELLS_PER_VD 7
 #define GXP_NUM_SYNC_BARRIERS_PER_VD 4
+#define GXP_NUM_TIMERS_PER_VD 1
 
 /* The first allowed doorbell and sync barrier to be used for VDs' usage */
 #define GXP_DOORBELLS_START 4 /* The first 4 are used for boot */
 #define GXP_SYNC_BARRIERS_START 1 /* The first 1 is used for UART */
+#define GXP_TIMERS_START 4 /* The first 4 are used for global and cores */
 
 /* Definitions for host->device boot mode requests */
 /*
@@ -89,12 +110,49 @@
 #define GXP_BOOT_STATUS_INVALID_MODE 4
 #define GXP_BOOT_STATUS_BOOTING 5
 #define GXP_BOOT_STATUS_BOOTING_FAILED 6
-#define GXP_BOOT_STATUS_SUSPENDING 7
-#define GXP_BOOT_STATUS_SUSPENDING_FAILED 8
-#define GXP_BOOT_STATUS_SUSPENDING_FAILED_ACTIVE_WL 9
-#define GXP_BOOT_STATUS_WAITING_FOR_WORKLOAD 10
-#define GXP_BOOT_STATUS_WAITING_FOR_DMA 11
-#define GXP_BOOT_STATUS_SHUTTING_DOWN 12
+#define GXP_BOOT_STATUS_SUSPEND_WAITING_FOR_WL 20
+#define GXP_BOOT_STATUS_SUSPEND_WAITING_FOR_DMA 21
+#define GXP_BOOT_STATUS_SUSPEND_SAVING_TCM 22
+#define GXP_BOOT_STATUS_SUSPEND_SAVING_TCM_FAILED 23
+#define GXP_BOOT_STATUS_SUSPEND_SAVING_TOP 24
+#define GXP_BOOT_STATUS_SUSPEND_SAVING_CORE 25
+#define GXP_BOOT_STATUS_SUSPEND_FAILED 26
+#define GXP_BOOT_STATUS_RESUME_RESTORING_CORE 40
+#define GXP_BOOT_STATUS_RESUME_RESTORING_CORE_FAILED 41
+#define GXP_BOOT_STATUS_RESUME_RESTORING_MISC 42
+#define GXP_BOOT_STATUS_RESUME_RESTORING_TCM 43
+#define GXP_BOOT_STATUS_RESUME_RESTORING_TOP 44
+#define GXP_BOOT_STATUS_RESUME_FAILED 44
+#define GXP_BOOT_STATUS_SHUTTING_DOWN 60
+
+/* Definitions for host->device warm up cache action requests */
+
+/* No boot action is needed. */
+#define GXP_WARM_UP_CACHE_NONE 0
+
+/* Call SystemImpl::HandleUserDispatch with empty command after resuming.  */
+#define GXP_WARM_UP_CACHE_CALL_HANDLE_USER_DISPATCH 1
+
+/* Timing info enums */
+/* Suspend */
+#define GXP_TIMING_MCU_INTENT_SUSPEND 0
+#define GXP_TIMING_DSP_ACK_SUSPEND 1
+#define GXP_TIMING_START_SUSPEND 2
+#define GXP_TIMING_CORE_QUIESCENCED 3
+#define GXP_TIMING_TCM_SAVED 4
+#define GXP_TIMING_TOP_SAVED 5
+#define GXP_TIMING_INTERNAL_SAVED 6
+#define GXP_TIMING_SUSPEND_COMPLETED 7
+
+/* Resume */
+#define GXP_TIMING_MCU_INTENT_RESUME 8
+#define GXP_TIMING_DSP_FIRST_INSTR 9
+#define GXP_TIMING_CACHE_RESET 10
+#define GXP_TIMING_INTERNAL_RESTORE_START 11
+#define GXP_TIMING_INTERNAL_RESTORED 12
+#define GXP_TIMING_TCM_RESTORED 13
+#define GXP_TIMING_TOP_RESTORED 14
+#define GXP_TIMING_RESUME_COMPLETED 15
 
 /* Bit masks for the status fields in the core telemetry structures. */
 /* The core telemetry buffers have been setup by the host. */
@@ -104,6 +162,9 @@
 /* There was an attempt to use the buffers but their content was invalid. */
 #define GXP_CORE_TELEMETRY_DEVICE_STATUS_SANITY_CHECK_FAILED (1 << 1)
 
+/* Mailbox command buffer descriptor invalid address for null command */
+#define GXP_DEVICE_ADDRESS_INVALID 0
+
 /*
  * A structure describing the core telemetry (logging and tracing) parameters
  * and buffers.
@@ -199,21 +260,74 @@
 	 */
 	uint32_t boot_status;
 
+	/* GXP kernel driver major version. Host is responsible for updating it. */
+	uint16_t gxp_kernel_driver_major_version;
+
+	/* GXP kernel driver minor version. Host is responsible for updating it. */
+	uint16_t gxp_kernel_driver_minor_version;
+
 	/* Reserved fields for future expansion */
-	uint32_t reserved_boot[12];
+	uint32_t reserved_boot[11];
 
 	/* To be used to communicate statistics for timing events during boot */
 	uint32_t timing_entries[16];
 
 	/* To be used to communicate crash events in case of failures */
-	uint32_t valid_crash_info;
+	uint32_t crash_handler_stage;
 	uint32_t crash_exccause;
 	uint32_t crash_excvaddr;
 	uint32_t crash_epc1;
-	uint32_t reserved_crash_info[12];
+	/* Written by host to request debug dump generation on core. */
+	uint32_t generate_debug_dump;
+	/* Written by core to notify the host about debug dump completion. */
+	uint32_t debug_dump_generated;
+	/* To be used by telemetry to check if IRQ needs to be sent to MCU/kernel. */
+	uint32_t telemetry_threshold_reached;
+
+	/* Debug keys to communicate runtime context */
+	/*
+	 * Android Process ID in runtime that triggered the workload on this DSP core.
+	 * It gets written by the runtime while preparing the workloads.
+	 */
+	uint16_t android_pid;
+	/*
+	 * Unique id of a GxpDevice within a process. It gets written by the runtime
+	 * while preparing the workloads.
+	 */
+	uint16_t device_id;
+	/*
+	 * Unique id associated with each run command of a device. It gets written by
+	 * the DSP firmware.
+	 */
+	uint32_t job_id;
+
+	/* Currently loaded active library details on DSP firmware */
+	/*
+	 * Base address of the currently loaded active library's code section on DSP
+	 * firmware.
+	 */
+	uint32_t code_section_base_addr;
+	/*
+	 * Base address of the currently loaded active library's data0 section on DSP
+	 * firmware.
+	 */
+	uint32_t data0_section_base_addr;
+	/*
+	 * Base address of the currently loaded active library's data1 section on DSP
+	 * firmware.
+	 */
+	uint32_t data1_section_base_addr;
+
+	uint32_t reserved_crash_info[4];
+
+	/*
+	 * Written to by the host to specify the action after resuming the core. See
+	 * the GXP_WARM_UP_* definitions for valid values.
+	 */
+	uint8_t warm_up_cache;
 
 	/* Reserved for more categories */
-	uint32_t reserved[16];
+	uint8_t reserved[63];
 
 	/*
 	 * The per-core job descriptor. This struct will be inspected by the FW
@@ -223,6 +337,22 @@
 };
 
 /*
+ * A structure describing the external state of the VD. This structure is read
+ * once by the FW upon the first cold boot and is never checked again.
+ */
+struct gxp_vd_descriptor {
+	/* The ID for this GXP application. */
+	uint32_t application_id;
+
+	/*
+	 * Whether or not this VD has been initialized by one of its cores.
+	 * This variable is protected by sync barrier at offset 0. Should be
+	 * initialized by the host to 0.
+	 */
+	uint32_t vd_is_initialized;
+};
+
+/*
  * A structure describing the telemetry (logging and tracing) parameters and
  * buffers; this describes R/O aspects of the telemetry buffers.
  */
@@ -247,22 +377,6 @@
 };
 
 /*
- * A structure describing the external state of the VD. This structure is read
- * once by the FW upon the first cold boot and is never checked again.
- */
-struct gxp_vd_descriptor {
-	/* The ID for this GXP application. */
-	uint32_t application_id;
-
-	/*
-	 * Whether or not this VD has been initialized by one of its cores.
-	 * This variable is protected by sync barrier at offset 0. Should be
-	 * initialized by the host to 0.
-	 */
-	uint32_t vd_is_initialized;
-};
-
-/*
  * A descriptor for data that is common to the entire system; usually accessed
  * by physical core. This region is mapped as R/O for all VDs. Should be
  * writable by the host (MCU/Kernel)
diff --git a/gxp-internal.h b/gxp-internal.h
index 35c0993..5985f27 100644
--- a/gxp-internal.h
+++ b/gxp-internal.h
@@ -23,9 +23,9 @@
 #include <linux/rwsem.h>
 #include <linux/spinlock.h>
 
-#include <gcip/iif/iif-manager.h>
 #include <gcip/gcip-resource-accessor.h>
 #include <gcip/gcip-thermal.h>
+#include <iif/iif-manager.h>
 
 #include "gxp-config.h"
 #include "gxp.h"
@@ -165,6 +165,7 @@
 
 	/* To manage IIF fences. */
 	struct iif_manager *iif_mgr;
+	struct device *iif_dev;
 
 	/* callbacks for chip-dependent implementations */
 
@@ -303,11 +304,8 @@
 	struct device_node *np;
 
 	np = of_parse_phandle(gxp->dev->of_node, phandle, 0);
-	if (IS_ERR_OR_NULL(np)) {
-		dev_err(gxp->dev, "Failed to find \"%s\" reserved memory\n",
-			phandle);
+	if (IS_ERR_OR_NULL(np))
 		return -ENODEV;
-	}
 
 	ret = of_address_to_resource(np, 0, r);
 	of_node_put(np);
diff --git a/gxp-kci.c b/gxp-kci.c
index 722722e..d86fb32 100644
--- a/gxp-kci.c
+++ b/gxp-kci.c
@@ -270,8 +270,7 @@
 	return -ENOMEM;
 }
 
-static void gxp_kci_release_resources(struct gxp_mailbox *mailbox,
-				      struct gxp_virtual_device *vd,
+static void gxp_kci_release_resources(struct gxp_mailbox *mailbox, struct gxp_virtual_device *vd,
 				      uint virt_core)
 {
 	struct gxp_kci *gkci = mailbox->data;
@@ -376,11 +375,7 @@
 {
 	struct gxp_mailbox *mailbox = gkci->mbx;
 
-	gxp_mailbox_write_descriptor(mailbox, mailbox->descriptor_buf.dsp_addr);
-	gxp_mailbox_reset(mailbox);
-	gxp_mailbox_enable_interrupt(mailbox);
-	gxp_mailbox_write_status(mailbox, 1);
-
+	gxp_mailbox_reinit(mailbox);
 	return 0;
 }
 
@@ -678,3 +673,26 @@
 	return gxp_kci_send_cmd_with_data(injection->kci_data, GCIP_KCI_CODE_FAULT_INJECTION,
 					  injection->opaque, sizeof(injection->opaque));
 }
+
+int gxp_kci_set_freq_limits(struct gxp_kci *gkci, u32 min_freq, u32 max_freq)
+{
+	struct gcip_kci_command_element cmd = {
+		.code = GCIP_KCI_CODE_SET_FREQ_LIMITS,
+		.dma = {
+			.size = min_freq,
+			.flags = max_freq,
+		},
+	};
+	return gxp_kci_send_cmd(gkci->mbx, &cmd);
+}
+
+int gxp_kci_thermal_control(struct gxp_kci *gkci, bool enable)
+{
+	struct gcip_kci_command_element cmd = {
+		.code = GCIP_KCI_CODE_THERMAL_CONTROL,
+		.dma = {
+			.flags = (u32)enable,
+		},
+	};
+	return gxp_kci_send_cmd(gkci->mbx, &cmd);
+}
diff --git a/gxp-kci.h b/gxp-kci.h
index 4bf413f..a355227 100644
--- a/gxp-kci.h
+++ b/gxp-kci.h
@@ -251,4 +251,24 @@
  */
 int gxp_kci_fault_injection(struct gcip_fault_inject *injection);
 
+/**
+ * gxp_kci_set_freq_limits() - Sends the frequency limits to be honoured to the firmware.
+ * @gkci: Reference to the GXP KCI mailbox struct.
+ * @min_freq: minimum frequency limit to be set.
+ * @max_freq: maximum frequency limit to be set.
+ *
+ * Return: 0 if the command is sent successfully.
+ */
+int gxp_kci_set_freq_limits(struct gxp_kci *gkci, u32 min_freq, u32 max_freq);
+
+/**
+ * gxp_kci_thermal_control() - Sends KCI command to enable/disable thermal throttling.
+ * @gkci: Reference to the GXP KCI mailbox struct.
+ * @enable: bool to indicate whether to enable or disable thermal throttling.
+ *
+ * Return: 0 if the command is sent successfully.
+ */
+
+int gxp_kci_thermal_control(struct gxp_kci *gkci, bool enable);
+
 #endif /* __GXP_KCI_H__ */
diff --git a/gxp-mailbox-driver.c b/gxp-mailbox-driver.c
index 7257114..a528695 100644
--- a/gxp-mailbox-driver.c
+++ b/gxp-mailbox-driver.c
@@ -12,7 +12,8 @@
 #include <linux/of_irq.h>
 #include <linux/spinlock.h>
 
-#include "gxp-config.h" /* GXP_USE_LEGACY_MAILBOX */
+#include <trace/events/gxp.h>
+
 #include "gxp-mailbox-driver.h"
 #include "gxp-mailbox-regs.h"
 #include "gxp-mailbox.h"
@@ -33,6 +34,8 @@
 {
 	struct gxp_mailbox *mailbox = (struct gxp_mailbox *)arg;
 
+	trace_gxp_uci_rsp_start(irq);
+
 	gxp_mailbox_chip_irq_handler(mailbox);
 	return IRQ_HANDLED;
 }
@@ -104,10 +107,9 @@
 	/* Nothing to cleanup */
 }
 
-void gxp_mailbox_driver_enable_interrupts(struct gxp_mailbox *mailbox)
+void gxp_mailbox_driver_register_interrupts(struct gxp_mailbox *mailbox)
 {
 	register_irq(mailbox);
-	gxp_mailbox_enable_interrupt(mailbox);
 }
 
 void gxp_mailbox_driver_disable_interrupts(struct gxp_mailbox *mailbox)
@@ -339,14 +341,6 @@
 	return gxp_mailbox_inc_resp_queue_head_nolock(mailbox, inc, wrap_bit);
 }
 
-#if !GXP_USE_LEGACY_MAILBOX
-u32 gxp_mailbox_gcip_ops_get_cmd_queue_head(struct gcip_mailbox *mailbox)
-{
-	struct gxp_mailbox *gxp_mbx = mailbox->data;
-
-	return gxp_mailbox_read_cmd_queue_head(gxp_mbx);
-}
-
 u32 gxp_mailbox_gcip_ops_get_cmd_queue_tail(struct gcip_mailbox *mailbox)
 {
 	struct gxp_mailbox *gxp_mbx = mailbox->data;
@@ -500,4 +494,3 @@
 
 	return gxp_pm_is_blk_down(gxp_mbx->gxp);
 }
-#endif /* !GXP_USE_LEGACY_MAILBOX */
diff --git a/gxp-mailbox-driver.h b/gxp-mailbox-driver.h
index 4349827..b883f7e 100644
--- a/gxp-mailbox-driver.h
+++ b/gxp-mailbox-driver.h
@@ -2,18 +2,19 @@
 /*
  * GXP mailbox driver.
  *
- * Copyright (C) 2020-2022 Google LLC
+ * Copyright (C) 2020-2024 Google LLC
  */
+
 #ifndef __GXP_MAILBOX_DRIVER_H__
 #define __GXP_MAILBOX_DRIVER_H__
 
+#include <linux/io.h>
+
+#include <gcip/gcip-mailbox.h>
+
 #include "gxp-config.h"
 #include "gxp-mailbox.h"
 
-#if !GXP_USE_LEGACY_MAILBOX
-#include <gcip/gcip-mailbox.h>
-#endif
-
 /* Utilities of circular queue operations */
 
 #define CIRCULAR_QUEUE_INDEX_MASK(wrap_bit) (wrap_bit - 1)
@@ -33,12 +34,22 @@
 void gxp_mailbox_driver_init(struct gxp_mailbox *mailbox);
 void gxp_mailbox_driver_exit(struct gxp_mailbox *mailbox);
 
-void gxp_mailbox_driver_enable_interrupts(struct gxp_mailbox *mailbox);
+void gxp_mailbox_driver_register_interrupts(struct gxp_mailbox *mailbox);
 void gxp_mailbox_driver_disable_interrupts(struct gxp_mailbox *mailbox);
 
 void __iomem *gxp_mailbox_get_csr_base(struct gxp_dev *gxp, uint index);
 void __iomem *gxp_mailbox_get_data_base(struct gxp_dev *gxp, uint index);
 
+static inline u32 gxp_mailbox_csr_read(struct gxp_mailbox *mailbox, uint reg_offset)
+{
+	return readl(mailbox->csr_reg_base + reg_offset);
+}
+
+static inline void gxp_mailbox_csr_write(struct gxp_mailbox *mailbox, uint reg_offset, u32 value)
+{
+	writel(value, mailbox->csr_reg_base + reg_offset);
+}
+
 void gxp_mailbox_reset_hw(struct gxp_mailbox *mailbox);
 /**
  * gxp_mailbox_generate_device_interrupt(): Trigger interrupt to device.
@@ -59,12 +70,6 @@
  * @mailbox: Mailbox for which to enable interrupts.
  */
 void gxp_mailbox_enable_interrupt(struct gxp_mailbox *mailbox);
-/**
- * gxp_mailbox_get_interrupt_status() - Retrieve the set interrupt bits coming
- *                                      to AP/Host.
- * @mailbox: Mailbox for which to get interrupt status.
- */
-u32 gxp_mailbox_get_interrupt_status(struct gxp_mailbox *mailbox);
 /* gxp_mailbox_wait_for_device_mailbox_init() - Wait for mailbox to get
  *                                              enabled/initialised by device.
  * @mailbox: Mailbox to get it enabled from device end.
@@ -165,12 +170,10 @@
 int gxp_mailbox_inc_resp_queue_head_locked(struct gxp_mailbox *mailbox, u32 inc,
 					   u32 wrap_bit);
 
-#if !GXP_USE_LEGACY_MAILBOX
 /*
  * Following functions are used when setting the operators of `struct gcip_mailbox_ops`.
  * To use these functions, @mailbox->data should be set as an instance of `struct gxp_mailbox`.
  */
-u32 gxp_mailbox_gcip_ops_get_cmd_queue_head(struct gcip_mailbox *mailbox);
 u32 gxp_mailbox_gcip_ops_get_cmd_queue_tail(struct gcip_mailbox *mailbox);
 void gxp_mailbox_gcip_ops_inc_cmd_queue_tail(struct gcip_mailbox *mailbox,
 					     u32 inc);
@@ -201,6 +204,5 @@
 void gxp_mailbox_gcip_ops_after_fetch_resps(struct gcip_mailbox *mailbox,
 					    u32 num_resps);
 bool gxp_mailbox_gcip_ops_is_block_off(struct gcip_mailbox *mailbox);
-#endif /* !GXP_USE_LEGACY_MAILBOX */
 
 #endif /* __GXP_MAILBOX_DRIVER_H__ */
diff --git a/gxp-mailbox.c b/gxp-mailbox.c
index 5afd820..daf8fdb 100644
--- a/gxp-mailbox.c
+++ b/gxp-mailbox.c
@@ -15,7 +15,9 @@
 #include <linux/spinlock.h>
 #include <uapi/linux/sched/types.h>
 
-#include "gxp-config.h" /* GXP_USE_LEGACY_MAILBOX */
+#include <gcip/gcip-mailbox.h>
+
+#include "gxp-config.h"
 #include "gxp-dma.h"
 #include "gxp-internal.h"
 #include "gxp-mailbox.h"
@@ -23,15 +25,12 @@
 #include "gxp-pm.h"
 #include "gxp.h"
 
-#if GXP_USE_LEGACY_MAILBOX
-#include "gxp-mailbox-impl.h"
-#else
-#include <gcip/gcip-mailbox.h>
+#if GXP_HAS_MCU
 #include <gcip/gcip-kci.h>
 
 #include "gxp-kci.h"
 #include "gxp-mcu-telemetry.h"
-#endif
+#endif /* GXP_HAS_MCU */
 
 /* Timeout of 1s by default */
 int gxp_mbx_timeout = 2000;
@@ -50,14 +49,12 @@
 	struct gxp_mailbox *mailbox =
 		container_of(work, struct gxp_mailbox, response_work);
 
-#if GXP_USE_LEGACY_MAILBOX
-	gxp_mailbox_consume_responses(mailbox);
-#else
 	if (gxp_is_direct_mode(mailbox->gxp))
 		gcip_mailbox_consume_responses_work(mailbox->mbx_impl.gcip_mbx);
+#if GXP_HAS_MCU
 	else if (mailbox->type == GXP_MBOX_TYPE_KCI)
 		gxp_mcu_telemetry_irq_handler(((struct gxp_kci *)mailbox->data)->mcu);
-#endif
+#endif /* GXP_HAS_MCU */
 }
 
 /*
@@ -198,7 +195,6 @@
 	kfree(mailbox);
 }
 
-#if !GXP_USE_LEGACY_MAILBOX
 static int init_gcip_mailbox(struct gxp_mailbox *mailbox)
 {
 	const struct gcip_mailbox_args args = {
@@ -243,6 +239,8 @@
 	mailbox->mbx_impl.gcip_mbx = NULL;
 }
 
+#if GXP_HAS_MCU
+
 static int init_gcip_kci(struct gxp_mailbox *mailbox)
 {
 	const struct gcip_kci_args args = {
@@ -285,44 +283,35 @@
 	kfree(gcip_kci);
 	mailbox->mbx_impl.gcip_kci = NULL;
 }
-#endif /* !GXP_USE_LEGACY_MAILBOX */
+
+#endif /* GXP_HAS_MCU */
 
 /*
  * Initializes @mailbox->mbx_impl to start waiting and consuming responses.
  * This will initializes GCIP mailbox modules according to the type of @mailbox.
  * - GENERAL: will initialize @mailbox->mbx_impl.gcip_mbx
  * - KCI: will initialize @mailbox->mbx_impl.kci_mbx
- *
- * Note: On `GXP_USE_LEGACY_MAILBOX`, it will initialize @mailbox itself as its
- * queuing logic is implemented in `gxp-mailbox-impl.c`.
  */
 static int init_mailbox_impl(struct gxp_mailbox *mailbox)
 {
 	int ret;
 
-#if GXP_USE_LEGACY_MAILBOX
-	if (mailbox->type != GXP_MBOX_TYPE_GENERAL)
-		return -EOPNOTSUPP;
-
-	ret = gxp_mailbox_init_consume_responses(mailbox);
-	if (ret)
-		return ret;
-#else
 	switch (mailbox->type) {
 	case GXP_MBOX_TYPE_GENERAL:
 		ret = init_gcip_mailbox(mailbox);
 		if (ret)
 			return ret;
 		break;
+#if GXP_HAS_MCU
 	case GXP_MBOX_TYPE_KCI:
 		ret = init_gcip_kci(mailbox);
 		if (ret)
 			return ret;
 		break;
+#endif /* GXP_HAS_MCU */
 	default:
 		return -EOPNOTSUPP;
 	}
-#endif /* GXP_USE_LEGACY_MAILBOX */
 
 	return 0;
 }
@@ -331,9 +320,6 @@
 {
 	int ret;
 
-	gxp_mailbox_write_descriptor(mailbox, mailbox->descriptor_buf.dsp_addr);
-	gxp_mailbox_reset(mailbox);
-
 	ret = init_mailbox_impl(mailbox);
 	if (ret)
 		return ret;
@@ -343,14 +329,18 @@
 	kthread_init_work(&mailbox->response_work,
 			  gxp_mailbox_consume_responses_work);
 
-	/* Only enable interrupts once everything has been setup */
-	gxp_mailbox_driver_enable_interrupts(mailbox);
-	/* Enable the mailbox */
-	gxp_mailbox_write_status(mailbox, 1);
-
+	gxp_mailbox_driver_register_interrupts(mailbox);
 	return 0;
 }
 
+void gxp_mailbox_reinit(struct gxp_mailbox *mailbox)
+{
+	gxp_mailbox_write_descriptor(mailbox, mailbox->descriptor_buf.dsp_addr);
+	gxp_mailbox_reset(mailbox);
+	gxp_mailbox_enable_interrupt(mailbox);
+	gxp_mailbox_write_status(mailbox, 1);
+}
+
 struct gxp_mailbox *gxp_mailbox_alloc(struct gxp_mailbox_manager *mgr,
 				      struct gxp_virtual_device *vd,
 				      uint virt_core, u8 core_id,
@@ -378,24 +368,21 @@
  * This releases GCIP mailbox modules according to the type of @mailbox.
  * - GENERAL: will release @mailbox->mbx_impl.gcip_mbx
  * - KCI: will release @mailbox->mbx_impl.kci_mbx
- *
- * Note: On `GXP_USE_LEGACY_MAILBOX`, it will release @mailbox itself as its
- * queuing logic is implemented in `gxp-mailbox-impl.c`.
  */
 static void release_mailbox_impl(struct gxp_mailbox *mailbox)
 {
-#if GXP_USE_LEGACY_MAILBOX
-	gxp_mailbox_release_consume_responses(mailbox);
-#else
 	switch (mailbox->type) {
 	case GXP_MBOX_TYPE_GENERAL:
 		release_gcip_mailbox(mailbox);
 		break;
+#if GXP_HAS_MCU
 	case GXP_MBOX_TYPE_KCI:
 		release_gcip_kci(mailbox);
 		break;
+#endif /* GXP_HAS_MCU */
+	default:
+		break;
 	}
-#endif
 }
 
 void gxp_mailbox_release(struct gxp_mailbox_manager *mgr,
@@ -480,16 +467,18 @@
 	return 0;
 }
 
-#if !GXP_USE_LEGACY_MAILBOX
 int gxp_mailbox_send_cmd(struct gxp_mailbox *mailbox, void *cmd, void *resp)
 {
 	switch (mailbox->type) {
 	case GXP_MBOX_TYPE_GENERAL:
 		return gcip_mailbox_send_cmd(mailbox->mbx_impl.gcip_mbx, cmd, resp, 0);
+#if GXP_HAS_MCU
 	case GXP_MBOX_TYPE_KCI:
 		return gcip_kci_send_cmd(mailbox->mbx_impl.gcip_kci, cmd);
+#endif /* GXP_HAS_MCU */
+	default:
+		return -EOPNOTSUPP;
 	}
-	return -EOPNOTSUPP;
 }
 
 struct gcip_mailbox_resp_awaiter *gxp_mailbox_put_cmd(struct gxp_mailbox *mailbox, void *cmd,
@@ -505,4 +494,3 @@
 	}
 	return ERR_PTR(-EOPNOTSUPP);
 }
-#endif /* !GXP_USE_LEGACY_MAILBOX */
diff --git a/gxp-mailbox.h b/gxp-mailbox.h
index 7e3dd58..5fe0069 100644
--- a/gxp-mailbox.h
+++ b/gxp-mailbox.h
@@ -10,17 +10,14 @@
 #include <linux/kthread.h>
 #include <linux/spinlock.h>
 
+#include <gcip/gcip-mailbox.h>
+
+#include "gxp-config.h"
 #include "gxp-client.h"
-#include "gxp-config.h" /* GXP_USE_LEGACY_MAILBOX */
 #include "gxp-dma.h"
 #include "gxp-internal.h"
 #include "gxp-mailbox-manager.h"
 
-#if !GXP_USE_LEGACY_MAILBOX
-#include <gcip/gcip-kci.h>
-#include <gcip/gcip-mailbox.h>
-#endif
-
 /* Pre-agreed values can be passed to gxp_mailbox_set_control(). */
 #define GXP_MBOX_CONTROL_MAGIC_POWER_DOWN (0xcafebabeu)
 
@@ -66,8 +63,6 @@
 enum gxp_mailbox_type {
 	/*
 	 * Mailbox will utilize `gcip-mailbox.h` internally.
-	 * (Note: On `GXP_USE_LEGACY_MAILBOX`, it utilizes `gxp-mailbox-impl.h`
-	 * instead.)
 	 * Mostly will be used for handling user commands.
 	 */
 	GXP_MBOX_TYPE_GENERAL = 0,
@@ -92,6 +87,8 @@
 	u32 resp_queue_size;
 };
 
+struct gcip_kci;
+struct gcip_kci_ops;
 struct gxp_mailbox;
 
 /*
@@ -135,7 +132,6 @@
 	void (*release_resources)(struct gxp_mailbox *mailbox,
 				  struct gxp_virtual_device *vd,
 				  uint virt_core);
-#if !GXP_USE_LEGACY_MAILBOX
 	/*
 	 * Operators which has dependency on the GCIP according to the type of mailbox.
 	 * - GXP_MBOX_TYPE_GENERAL: @gcip_ops.mbx must be defined.
@@ -145,7 +141,6 @@
 		const struct gcip_mailbox_ops *mbx;
 		const struct gcip_kci_ops *kci;
 	} gcip_ops;
-#endif
 };
 
 struct gxp_mailbox_args {
@@ -203,13 +198,6 @@
 	struct gxp_mailbox_ops *ops;
 	void *data; /* private data */
 
-#if GXP_USE_LEGACY_MAILBOX
-	u64 cur_seq;
-	/* add to this list if a command needs to wait for a response */
-	struct list_head wait_list;
-	/* queue for waiting for the wait_list to be consumed */
-	wait_queue_head_t wait_list_waitq;
-#else /* !GXP_USE_LEGACY_MAILBOX */
 	/*
 	 * Implementation of the mailbox according to the type.
 	 * - GXP_MBOX_TYPE_GENERAL: @gcip_mbx will be allocated.
@@ -219,7 +207,6 @@
 		struct gcip_mailbox *gcip_mbx;
 		struct gcip_kci *gcip_kci;
 	} mbx_impl;
-#endif /* GXP_USE_LEGACY_MAILBOX */
 };
 
 /* Mailbox APIs */
@@ -255,7 +242,12 @@
 int gxp_mailbox_unregister_interrupt_handler(struct gxp_mailbox *mailbox,
 					     u32 int_bit);
 
-#if !GXP_USE_LEGACY_MAILBOX
+/*
+ * Initialises the mailbox CSRs.
+ * Must be called only after mailbox is allocated via gxp_mailbox_alloc().
+ */
+void gxp_mailbox_reinit(struct gxp_mailbox *mailbox);
+
 /*
  * Executes command synchronously. If @resp is not NULL, the response will be returned to it.
  * See the `gcip_mailbox_send_cmd` of `gcip-mailbox.h` or `gcip_kci_send_cmd` of `gcip-kci.h`
@@ -272,6 +264,5 @@
 struct gcip_mailbox_resp_awaiter *gxp_mailbox_put_cmd(struct gxp_mailbox *mailbox, void *cmd,
 						      void *resp, void *data,
 						      gcip_mailbox_cmd_flags_t flags);
-#endif /* !GXP_USE_LEGACY_MAILBOX */
 
 #endif /* __GXP_MAILBOX_H__ */
diff --git a/gxp-mcu-firmware.c b/gxp-mcu-firmware.c
index 6d19080..78a566f 100644
--- a/gxp-mcu-firmware.c
+++ b/gxp-mcu-firmware.c
@@ -16,6 +16,7 @@
 #include <linux/string.h>
 #include <linux/workqueue.h>
 
+#include <gcip/gcip-alloc-helper.h>
 #include <gcip/gcip-common-image-header.h>
 #include <gcip/gcip-fault-injection.h>
 #include <gcip/gcip-image-config.h>
@@ -62,17 +63,33 @@
 /*
  * Programs instruction remap CSRs.
  */
-static void program_iremap_csr(struct gxp_dev *gxp,
-			       struct gxp_mapped_resource *buf)
+static int program_iremap_csr(struct gxp_dev *gxp, struct gxp_mapped_resource *buf)
 {
+	struct gxp_mcu_firmware *mcu_fw = gxp_mcu_firmware_of(gxp);
+	size_t size;
+
 	dev_info(gxp->dev, "Program instruction remap CSRs");
 	gxp_soc_set_iremap_context(gxp);
-	gxp_write_32(gxp, GXP_REG_CFGVECTABLE0, buf->daddr);
 
-	gxp_write_32(gxp, GXP_REG_IREMAP_LOW, buf->daddr);
-	gxp_write_32(gxp, GXP_REG_IREMAP_HIGH, buf->daddr + buf->size);
+	if (mcu_fw->dynamic_fw_buffer) {
+		if (buf->daddr + buf->size > GXP_IREMAP_DYNAMIC_CODE_BASE) {
+			dev_err(gxp->dev,
+				"Bad dynamic firmware base %x, carveout daddr: %pad size: %llx",
+				GXP_IREMAP_DYNAMIC_CODE_BASE, &buf->daddr, buf->size);
+			return -EINVAL;
+		}
+		gxp_write_32(gxp, GXP_REG_CFGVECTABLE0, GXP_IREMAP_DYNAMIC_CODE_BASE);
+		size = mcu_fw->dynamic_fw_buffer->size;
+		gxp_write_32(gxp, GXP_REG_IREMAP_LOW, buf->daddr);
+		gxp_write_32(gxp, GXP_REG_IREMAP_HIGH, GXP_IREMAP_DYNAMIC_CODE_BASE + size);
+	} else {
+		gxp_write_32(gxp, GXP_REG_CFGVECTABLE0, buf->daddr);
+		gxp_write_32(gxp, GXP_REG_IREMAP_LOW, buf->daddr);
+		gxp_write_32(gxp, GXP_REG_IREMAP_HIGH, buf->daddr + buf->size);
+	}
 	gxp_write_32(gxp, GXP_REG_IREMAP_TARGET, buf->daddr);
 	gxp_write_32(gxp, GXP_REG_IREMAP_ENABLE, 1);
+	return 0;
 }
 
 /*
@@ -260,6 +277,68 @@
 	return gxp_pg_by_recovery_boot(gxp, force);
 }
 
+static struct gxp_mcu_firmware_ns_buffer *map_ns_buffer(struct gxp_dev *gxp, dma_addr_t daddr,
+							size_t size)
+{
+	struct gxp_mcu_firmware_ns_buffer *ns_buffer;
+	u64 gcip_map_flags = GCIP_MAP_FLAGS_DMA_RW;
+
+	ns_buffer = kzalloc(sizeof(*ns_buffer), GFP_KERNEL);
+	if (!ns_buffer)
+		return ERR_PTR(-ENOMEM);
+
+	ns_buffer->daddr = daddr;
+	ns_buffer->size = size;
+	ns_buffer->sgt = gcip_alloc_noncontiguous(gxp->dev, size, GFP_KERNEL);
+	if (!ns_buffer->sgt) {
+		kfree(ns_buffer);
+		return ERR_PTR(-ENOMEM);
+	}
+	if (!gcip_iommu_domain_map_sgt_to_iova(gxp_iommu_get_domain_for_dev(gxp), ns_buffer->sgt,
+					       daddr, &gcip_map_flags)) {
+		dev_err(gxp->dev, "Failed to map NS buffer, daddr: %pad size: %#zx", &daddr, size);
+		gcip_free_noncontiguous(ns_buffer->sgt);
+		kfree(ns_buffer);
+		return ERR_PTR(-EBUSY);
+	}
+	return ns_buffer;
+}
+
+static void unmap_ns_buffer(struct gxp_dev *gxp, struct gxp_mcu_firmware_ns_buffer *ns_buffer)
+{
+	gcip_iommu_domain_unmap_sgt_from_iova(gxp_iommu_get_domain_for_dev(gxp), ns_buffer->sgt,
+					      GCIP_MAP_FLAGS_DMA_RW);
+	gcip_free_noncontiguous(ns_buffer->sgt);
+	kfree(ns_buffer);
+}
+
+static void reset_shadow_memory(struct gxp_dev *gxp)
+{
+	struct gxp_mcu_firmware *mcu_fw = gxp_mcu_firmware_of(gxp);
+	struct gxp_mcu_firmware_ns_buffer *shadow_buffer = NULL, *cur;
+	u32 shadow_index =
+		GCIP_IMAGE_CONFIG_SANITIZER_INDEX(mcu_fw->cfg_parser.last_config.sanitizer_config);
+	u32 count = 0;
+
+	mutex_lock(&mcu_fw->ns_buffer_list_lock);
+	list_for_each_entry(cur, &mcu_fw->ns_buffer_list, list) {
+		if (count == shadow_index) {
+			shadow_buffer = cur;
+			break;
+		}
+		count++;
+	}
+	mutex_unlock(&mcu_fw->ns_buffer_list_lock);
+
+	if (shadow_buffer) {
+		memset(gcip_noncontiguous_sgt_to_mem(shadow_buffer->sgt), 0, shadow_buffer->size);
+		gxp_dma_sync_sg_for_device(gxp, shadow_buffer->sgt->sgl,
+					   shadow_buffer->sgt->orig_nents, DMA_TO_DEVICE);
+	} else {
+		dev_warn(gxp->dev, "shadow buffer not found");
+	}
+}
+
 int gxp_mcu_firmware_load(struct gxp_dev *gxp, char *fw_name,
 			  const struct firmware **fw)
 {
@@ -298,13 +377,6 @@
 
 	size = (*fw)->size - GCIP_FW_HEADER_SIZE;
 
-	if (size > mcu_fw->image_buf.size) {
-		dev_err(dev, "firmware %s size %#zx exceeds buffer size %#llx",
-			fw_name, size, mcu_fw->image_buf.size);
-		ret = -ENOSPC;
-		goto err_release_firmware;
-	}
-
 	imgcfg = get_image_config_from_hdr(hdr);
 	if (!imgcfg) {
 		dev_err(dev, "Unsupported image header generation");
@@ -332,8 +404,32 @@
 	}
 	mcu_fw->is_secure = !gcip_image_config_is_ns(imgcfg);
 
-	memcpy(mcu_fw->image_buf.vaddr, (*fw)->data + GCIP_FW_HEADER_SIZE,
-	       size);
+	mcu_fw->sanitizer_status = GCIP_IMAGE_CONFIG_SANITIZER_STATUS(imgcfg->sanitizer_config);
+
+	if (size > mcu_fw->image_buf.size || (mcu_fw->sanitizer_status != 0)) {
+		if (mcu_fw->is_secure) {
+			dev_err(dev, "firmware %s size %#zx exceeds buffer size %#llx", fw_name,
+				size, mcu_fw->image_buf.size);
+			ret = -ENOSPC;
+			goto err_clear_config;
+		}
+
+		/*
+		 * In non-secure mode, we support allocating buffers to put MCU firmware image
+		 * instead of using carveouts.
+		 */
+		mcu_fw->dynamic_fw_buffer = map_ns_buffer(gxp, GXP_IREMAP_DYNAMIC_CODE_BASE, size);
+		if (IS_ERR(mcu_fw->dynamic_fw_buffer))
+			goto err_clear_config;
+		memcpy(gcip_noncontiguous_sgt_to_mem(mcu_fw->dynamic_fw_buffer->sgt),
+		       (*fw)->data + GCIP_FW_HEADER_SIZE, size);
+		gxp_dma_sync_sg_for_device(gxp, mcu_fw->dynamic_fw_buffer->sgt->sgl,
+					   mcu_fw->dynamic_fw_buffer->sgt->orig_nents,
+					   DMA_TO_DEVICE);
+	} else {
+		memcpy(mcu_fw->image_buf.vaddr, (*fw)->data + GCIP_FW_HEADER_SIZE, size);
+	}
+
 out:
 	mutex_unlock(&mcu_fw->lock);
 	return 0;
@@ -358,6 +454,10 @@
 		mutex_unlock(&mcu_fw->lock);
 		return;
 	}
+	if (mcu_fw->dynamic_fw_buffer) {
+		unmap_ns_buffer(gxp, mcu_fw->dynamic_fw_buffer);
+		mcu_fw->dynamic_fw_buffer = NULL;
+	}
 	gcip_image_config_clear(&mcu_fw->cfg_parser);
 	mcu_fw->status = GCIP_FW_INVALID;
 	mutex_unlock(&mcu_fw->lock);
@@ -388,7 +488,11 @@
 			return -EIO;
 		}
 	} else {
-		program_iremap_csr(gxp, &mcu_fw->image_buf);
+		ret = program_iremap_csr(gxp, &mcu_fw->image_buf);
+		if (ret) {
+			gxp_lpm_down(gxp, GXP_MCU_CORE_ID);
+			return ret;
+		}
 		/* Raise wakeup doorbell */
 		dev_dbg(gxp->dev, "Raising doorbell %d interrupt\n",
 			CORE_WAKEUP_DOORBELL(GXP_MCU_CORE_ID));
@@ -548,6 +652,9 @@
 
 	lockdep_assert_held(&mcu_fw->lock);
 
+	if (mcu_fw->sanitizer_status & GCIP_FW_ASAN_ENABLED)
+		reset_shadow_memory(gxp);
+
 	/*
 	 * Resets UCI/KCI CSRs to ensure that no unconsumed commands are carried over from the last
 	 * execution.
@@ -596,8 +703,10 @@
 	int ret;
 
 	ret = gxp_acquire_rmem_resource(gxp, &r, "gxp-mcu-fw-region");
-	if (ret)
+	if (ret) {
+		dev_err(gxp->dev, "Failed to find reserved memory for MCU FW: %d", ret);
 		return ret;
+	}
 	buf->size = resource_size(&r);
 	buf->paddr = r.start;
 	buf->daddr = GXP_IREMAP_CODE_BASE;
@@ -734,24 +843,71 @@
 	.attrs = dev_attrs,
 };
 
-static int image_config_map(void *data, dma_addr_t daddr, phys_addr_t paddr,
-			    size_t size, unsigned int flags)
+static int image_config_map_ns(struct gxp_dev *gxp, dma_addr_t daddr, size_t size)
 {
-	struct gxp_dev *gxp = data;
-	const bool ns = !(flags & GCIP_IMAGE_CONFIG_FLAGS_SECURE);
+	struct gxp_mcu_firmware *mcu_fw = gxp_mcu_firmware_of(gxp);
+	struct gxp_mcu_firmware_ns_buffer *ns_buffer;
 
-	if (ns) {
-		dev_err(gxp->dev, "image config NS mappings are not supported");
-		return -EINVAL;
-	}
+	ns_buffer = map_ns_buffer(gxp, daddr, size);
+	if (IS_ERR(ns_buffer))
+		return PTR_ERR(ns_buffer);
 
-	return gcip_iommu_map(gxp_iommu_get_domain_for_dev(gxp), daddr, paddr, size,
-			      GCIP_MAP_FLAGS_DMA_RW);
+	mutex_lock(&mcu_fw->ns_buffer_list_lock);
+	list_add_tail(&ns_buffer->list, &mcu_fw->ns_buffer_list);
+	mutex_unlock(&mcu_fw->ns_buffer_list_lock);
+	return 0;
 }
 
-static void image_config_unmap(void *data, dma_addr_t daddr, size_t size, unsigned int flags)
+static void image_config_unmap_ns(struct gxp_dev *gxp, dma_addr_t daddr, size_t size)
+{
+	struct gxp_mcu_firmware *mcu_fw = gxp_mcu_firmware_of(gxp);
+	struct gxp_mcu_firmware_ns_buffer *ns_buffer = NULL, *cur;
+
+	mutex_lock(&mcu_fw->ns_buffer_list_lock);
+	list_for_each_entry(cur, &mcu_fw->ns_buffer_list, list) {
+		if (cur->daddr == daddr && cur->size == size) {
+			ns_buffer = cur;
+			list_del(&cur->list);
+			break;
+		}
+	}
+	mutex_unlock(&mcu_fw->ns_buffer_list_lock);
+
+	if (ns_buffer) {
+		unmap_ns_buffer(gxp, cur);
+	} else {
+		dev_warn(gxp->dev, "Failed to find NS buffer, daddr: %pad size: %#zx", &daddr,
+			 size);
+	}
+}
+
+static int image_config_map(void *data, dma_addr_t daddr, phys_addr_t paddr, size_t size,
+			    unsigned int cfg_map_flags, unsigned int cfg_op_flags)
 {
 	struct gxp_dev *gxp = data;
+	const bool ns = !(cfg_op_flags & GCIP_IMAGE_CONFIG_FLAGS_SECURE);
+	u64 gcip_map_flags = GCIP_MAP_FLAGS_DMA_RW;
+
+	if (ns)
+		return image_config_map_ns(gxp, daddr, size);
+
+	if (GCIP_IMAGE_CONFIG_MAP_MMIO(cfg_map_flags))
+		gcip_map_flags |= GCIP_MAP_FLAGS_MMIO_TO_FLAGS(1);
+
+	return gcip_iommu_map(gxp_iommu_get_domain_for_dev(gxp), daddr, paddr, size,
+			      gcip_map_flags);
+}
+
+static void image_config_unmap(void *data, dma_addr_t daddr, size_t size,
+			       unsigned int cfg_map_flags, unsigned int cfg_op_flags)
+{
+	struct gxp_dev *gxp = data;
+	const bool ns = !(cfg_op_flags & GCIP_IMAGE_CONFIG_FLAGS_SECURE);
+
+	if (ns) {
+		image_config_unmap_ns(gxp, daddr, size);
+		return;
+	}
 
 	gcip_iommu_unmap(gxp_iommu_get_domain_for_dev(gxp), daddr, size);
 }
@@ -814,6 +970,8 @@
 	mcu_fw->status = GCIP_FW_INVALID;
 	mcu_fw->crash_cnt = 0;
 	mutex_init(&mcu_fw->lock);
+	INIT_LIST_HEAD(&mcu_fw->ns_buffer_list);
+	mutex_init(&mcu_fw->ns_buffer_list_lock);
 	INIT_WORK(&mcu_fw->fw_crash_handler_work, gxp_mcu_firmware_crash_handler_work);
 
 	ret = device_add_group(gxp->dev, &firmware_attr_group);
diff --git a/gxp-mcu-firmware.h b/gxp-mcu-firmware.h
index 739c38b..e8df752 100644
--- a/gxp-mcu-firmware.h
+++ b/gxp-mcu-firmware.h
@@ -19,6 +19,17 @@
 #define GXP_MCU_BOOT_MODE_NORMAL 0
 #define GXP_MCU_BOOT_MODE_RECOVERY 1
 
+struct gxp_mcu_firmware_ns_buffer {
+	/* SG table for NS firmware buffer mappings. */
+	struct sg_table *sgt;
+	/* DMA address of the NS firmware buffer. */
+	dma_addr_t daddr;
+	/* Size of the NS firmware buffer. */
+	size_t size;
+	/* List of NS firmware buffer mappings for the device. */
+	struct list_head list;
+};
+
 struct gxp_mcu_firmware {
 	struct gxp_dev *gxp;
 	/* resource for MCU firmware image */
@@ -35,6 +46,14 @@
 	struct work_struct fw_crash_handler_work;
 	/* The container of fault injection data. */
 	struct gcip_fault_inject *fault_inject;
+	/* List of all NS buffer mappings for the device. */
+	struct list_head ns_buffer_list;
+	/* Lock to protect @ns_buffer_list. */
+	struct mutex ns_buffer_list_lock;
+	/* The buffer of dynamic fw memory, which is only used in non-secure mode. */
+	struct gxp_mcu_firmware_ns_buffer *dynamic_fw_buffer;
+	/* The sanitizer enablement status for ASAN and UBSAN */
+	int sanitizer_status;
 };
 
 /*
diff --git a/gxp-mcu-fs.c b/gxp-mcu-fs.c
index 2c97e30..c2effcc 100644
--- a/gxp-mcu-fs.c
+++ b/gxp-mcu-fs.c
@@ -12,10 +12,13 @@
 #include <linux/rwsem.h>
 #include <linux/slab.h>
 
+#include <gcip/gcip-dma-fence.h>
 #include <gcip/gcip-fence-array.h>
 #include <gcip/gcip-fence.h>
 #include <gcip/gcip-telemetry.h>
 
+#include <trace/events/gxp.h>
+
 #include "gxp-client.h"
 #include "gxp-internal.h"
 #include "gxp-mcu-fs.h"
@@ -52,59 +55,6 @@
 	return 0;
 }
 
-/**
- * polled_dma_fence_get() - Generate a dma-fence to represent all the fan-in fences.
- * @in_fences: The pointer to the gcip_fence_array object containing fan-in fances.
- *
- * Use dma_fence_array to handle all the fan-in fences if @in_fences->size > 1.
- * The caller should hold the reference count of the fences in @in_fences to make sure they will not
- * be released during the process.
- * The output fence will acquired 1 reference count in this function either with dma_fence_get() or
- * dma_fence_array_create().
- *
- * Return: The pointer to the generated dma-fence or the error pointer on error.
- *         A NULL pointer is returned if no in-kernel fence is passed in.
- */
-static struct dma_fence *polled_dma_fence_get(struct gcip_fence_array *in_fences)
-{
-	static int array_seq;
-	const int size = in_fences->size;
-	struct dma_fence_array *fence_array;
-	struct dma_fence **in_dma_fences;
-	int i;
-
-	if (!in_fences->size || !in_fences->same_type || in_fences->type != GCIP_IN_KERNEL_FENCE)
-		return NULL;
-
-	/* TODO(b/320401031): Remove this constraint after dma-fence-unwrap adopted. */
-	/* dma-fence-array as in-fences is currently not supported. */
-	for (i = 0; i < size; i++) {
-		if (dma_fence_is_array(in_fences->fences[i]->fence.ikf))
-			return ERR_PTR(-EINVAL);
-	}
-
-	if (size == 1)
-		return dma_fence_get(in_fences->fences[0]->fence.ikf);
-
-	in_dma_fences = kcalloc(size, sizeof(*in_dma_fences), GFP_KERNEL);
-	if (!in_dma_fences)
-		return ERR_PTR(-ENOMEM);
-
-	for (i = 0; i < size; i++)
-		in_dma_fences[i] = dma_fence_get(in_fences->fences[i]->fence.ikf);
-
-	/* fence_array will take care of the life cycle of in_dma_fences.*/
-	fence_array = dma_fence_array_create(size, in_dma_fences, dma_fence_context_alloc(1),
-					     array_seq++, false);
-	if (!fence_array) {
-		kfree(in_dma_fences);
-		/* dma_fence_array_create only returns NULL on allocation failure. */
-		return ERR_PTR(-ENOMEM);
-	}
-
-	return &fence_array->base;
-}
-
 /*
  * Returns the number of fences. If the runtime passed an invalid fence, returns an errno
  * accordingly.
@@ -145,6 +95,8 @@
 
 	cmd_seq = gcip_mailbox_inc_seq_num(mcu->uci.mbx->mbx_impl.gcip_mbx, 1);
 
+	trace_gxp_uci_cmd_start(cmd_seq);
+
 	num_in_fences = get_num_fences(ibuf.in_fences);
 	if (num_in_fences < 0)
 		return num_in_fences;
@@ -163,7 +115,8 @@
 		return PTR_ERR(out_fences);
 	}
 
-	polled_dma_fence = polled_dma_fence_get(in_fences);
+	/* @polled_dma_fence can be NULL if @in_fences are not DMA fences or there are no fences. */
+	polled_dma_fence = gcip_fence_array_merge_ikf(in_fences);
 	if (IS_ERR(polled_dma_fence)) {
 		ret = PTR_ERR(polled_dma_fence);
 		goto out;
@@ -183,7 +136,7 @@
 
 err_put_fence:
 	/*
-	 * Put the reference count of the fence acqurired in polled_dma_fence_get.
+	 * Put the reference count of the fence acqurired in gcip_fence_array_merge_ikf.
 	 * If the fence is a dma_fence_array and the callback is failed to be added,
 	 * the whole object and the array it holds will be freed.
 	 * If it is a NULL pointer, it's still safe to call this function.
@@ -192,6 +145,9 @@
 out:
 	gcip_fence_array_put(out_fences);
 	gcip_fence_array_put(in_fences);
+
+	trace_gxp_uci_cmd_end(cmd_seq);
+
 	return ret;
 }
 
diff --git a/gxp-mcu-platform.c b/gxp-mcu-platform.c
index 8e55dcd..edc66e4 100644
--- a/gxp-mcu-platform.c
+++ b/gxp-mcu-platform.c
@@ -5,11 +5,15 @@
  * Copyright (C) 2022 Google LLC
  */
 
+#include <linux/device.h>
 #include <linux/interrupt.h>
 #include <linux/moduleparam.h>
 #include <linux/of_irq.h>
+#include <linux/of_platform.h>
+#include <linux/of.h>
+#include <linux/platform_device.h>
 
-#include <gcip/iif/iif-manager.h>
+#include <iif/iif-manager.h>
 
 #include "gxp-config.h"
 #include "gxp-internal.h"
@@ -271,6 +275,81 @@
 	return GXP_CHIP_ANY;
 }
 
+static void gxp_get_embedded_iif_mgr(struct gxp_dev *gxp)
+{
+	struct iif_manager *mgr;
+
+#if HAS_TPU_EXT
+	if (gxp->tpu_dev.dev) {
+		int ret = edgetpu_ext_driver_cmd(gxp->tpu_dev.dev, EDGETPU_EXTERNAL_CLIENT_TYPE_DSP,
+						 GET_IIF_MANAGER, NULL, &mgr);
+
+		if (!ret) {
+			dev_info(gxp->dev, "Use the IIF manager of TPU driver");
+			/* Note that we shouldn't call `iif_manager_get` here. */
+			gxp->iif_mgr = mgr;
+			return;
+		}
+	}
+#endif /* HAS_TPU_EXT */
+
+	dev_info(gxp->dev, "Try to get an embedded IIF manager");
+
+	mgr = iif_manager_init(gxp->dev->of_node);
+	if (IS_ERR(mgr)) {
+		dev_warn(gxp->dev, "Failed to init an embedded IIF manager: %ld", PTR_ERR(mgr));
+		return;
+	}
+
+	gxp->iif_mgr = mgr;
+}
+
+static void gxp_get_iif_mgr(struct gxp_dev *gxp)
+{
+	struct platform_device *pdev;
+	struct device_node *node;
+	struct iif_manager *mgr;
+
+	node = of_parse_phandle(gxp->dev->of_node, "iif-device", 0);
+	if (IS_ERR_OR_NULL(node)) {
+		dev_warn(gxp->dev, "There is no iif-device node in the device tree");
+		goto get_embed;
+	}
+
+	pdev = of_find_device_by_node(node);
+	of_node_put(node);
+	if (!pdev) {
+		dev_warn(gxp->dev, "Failed to find the IIF device");
+		goto get_embed;
+	}
+
+	mgr = platform_get_drvdata(pdev);
+	if (!mgr) {
+		dev_warn(gxp->dev, "Failed to get a manager from IIF device");
+		goto put_device;
+	}
+
+	dev_info(gxp->dev, "Use the IIF manager of IIF device");
+
+	/* We don't need to call `get_device` since `of_find_device_by_node` takes a refcount. */
+	gxp->iif_dev = &pdev->dev;
+	gxp->iif_mgr = iif_manager_get(mgr);
+	return;
+
+put_device:
+	put_device(&pdev->dev);
+get_embed:
+	gxp_get_embedded_iif_mgr(gxp);
+}
+
+static void gxp_put_iif_mgr(struct gxp_dev *gxp)
+{
+	if (gxp->iif_mgr)
+		iif_manager_put(gxp->iif_mgr);
+	/* NO-OP if `gxp->iif_dev` is NULL. */
+	put_device(gxp->iif_dev);
+}
+
 int gxp_mcu_platform_after_probe(struct gxp_dev *gxp)
 {
 	int ret;
@@ -282,12 +361,7 @@
 	if (ret)
 		return ret;
 
-	gxp->iif_mgr = iif_manager_init(gxp->dev->of_node);
-	if (IS_ERR(gxp->iif_mgr)) {
-		dev_err(gxp->dev, "Failed to init IIF manager: %ld", PTR_ERR(gxp->iif_mgr));
-		gxp->iif_mgr = NULL;
-	}
-
+	gxp_get_iif_mgr(gxp);
 	gxp_usage_stats_init(gxp);
 	return gxp_mcu_init(gxp, gxp_mcu_of(gxp));
 }
@@ -299,9 +373,7 @@
 
 	gxp_mcu_exit(gxp_mcu_of(gxp));
 	gxp_usage_stats_exit(gxp);
-
-	if (gxp->iif_mgr)
-		iif_manager_put(gxp->iif_mgr);
+	gxp_put_iif_mgr(gxp);
 }
 
 void gxp_mcu_dev_init(struct gxp_mcu_dev *mcu_dev)
diff --git a/gxp-pm.c b/gxp-pm.c
index d294e02..c59ef17 100644
--- a/gxp-pm.c
+++ b/gxp-pm.c
@@ -27,11 +27,20 @@
 #include "gxp-pm.h"
 #include "mobile-soc.h"
 
+#if GXP_HAS_MCU
+#include <gcip/gcip-kci.h>
+
+#include "gxp-kci.h"
+#include "gxp-mcu.h"
+#endif /* GXP_HAS_MCU */
+
 /* Don't attempt to touch the device when @busy_count equals this value. */
 #define BUSY_COUNT_OFF (~0ull)
 
 #define DEBUGFS_BLK_POWERSTATE "blk_powerstate"
 #define DEBUGFS_WAKELOCK "wakelock"
+#define DEBUGFS_MIN_FREQ "min_freq"
+#define DEBUGFS_MAX_FREQ "max_freq"
 
 #define SHUTDOWN_DELAY_US_MIN 200
 #define SHUTDOWN_DELAY_US_MAX 400
@@ -403,8 +412,7 @@
 		return -EINVAL;
 	}
 	if (gxp->power_mgr->curr_state == AUR_OFF) {
-		dev_err(gxp->dev,
-			"Cannot request power state when BLK is off\n");
+		dev_warn(gxp->dev, "Cannot request power state when BLK is off\n");
 		return -EBUSY;
 	}
 	if (state == AUR_OFF)
@@ -731,36 +739,118 @@
 	return ret;
 }
 
-static int gxp_pm_power_up(void *data)
+#if GXP_HAS_MCU
+static int gxp_pm_update_freq_limits_locked(struct gxp_dev *gxp)
 {
-	struct gxp_dev *gxp = data;
-	int ret = gxp_pm_blk_on(gxp);
+	struct gxp_power_manager *mgr = gxp->power_mgr;
+	struct gxp_kci *kci = &(gxp_mcu_of(gxp)->kci);
+	int ret;
 
+	lockdep_assert_held(&gxp->power_mgr->freq_limits_lock);
+
+	ret = gxp_kci_set_freq_limits(kci, mgr->min_freq_limit, mgr->max_freq_limit);
 	if (ret) {
-		dev_err(gxp->dev, "Failed to power on BLK_AUR (ret=%d)\n", ret);
-		return ret;
-	}
-
-	if (gxp->pm_after_blk_on) {
-		ret = gxp->pm_after_blk_on(gxp);
-		if (ret) {
-			gxp_pm_blk_off(gxp);
-			return ret;
+		dev_warn(gxp->dev, "Set frequency limit request failed with error %d.", ret);
+		if (ret == GCIP_KCI_ERROR_INVALID_ARGUMENT) {
+			dev_warn(gxp->dev,
+				 "Invalid values within debugfs frequency limits: [%u, %u]\n",
+				 mgr->min_freq_limit, mgr->max_freq_limit);
+			ret = -EINVAL;
+		} else {
+			ret = -EIO;
 		}
+		mgr->min_freq_limit = 0;
+		mgr->max_freq_limit = 0;
+	} else {
+		dev_info(gxp->dev, "BLK frequency to remain in [%u, %u]kHz frequency limit.\n",
+			 mgr->min_freq_limit, mgr->max_freq_limit);
 	}
 
+	return ret;
+}
+
+static int gxp_pm_set_freq_limit(struct gxp_dev *gxp, u32 val, u32 *limit_to_set)
+{
+	int ret = 0;
+
+	if (val == *limit_to_set)
+		return ret;
+	/*
+	 * Need to hold pm lock to prevent races with power up/down when checking block state and
+	 * sending the KCI command to update limits.
+	 *
+	 * Since power_up will also acquire freq_limits_lock to send initial limits, pm lock must be
+	 * held first to avoid lock inversion.
+	 */
+	gcip_pm_lock(gxp->power_mgr->pm);
+	mutex_lock(&gxp->power_mgr->freq_limits_lock);
+
+	*limit_to_set = val;
+	if (!gxp_pm_is_blk_down(gxp))
+		ret = gxp_pm_update_freq_limits_locked(gxp);
+
+	mutex_unlock(&gxp->power_mgr->freq_limits_lock);
+	gcip_pm_unlock(gxp->power_mgr->pm);
+	return ret;
+}
+
+static int debugfs_min_freq_limit_get(void *data, u64 *val)
+{
+	struct gxp_dev *gxp = (struct gxp_dev *)data;
+	struct gxp_power_manager *mgr = gxp->power_mgr;
+
+	mutex_lock(&mgr->freq_limits_lock);
+	*val = mgr->min_freq_limit;
+	mutex_unlock(&mgr->freq_limits_lock);
 	return 0;
 }
 
-static int gxp_pm_power_down(void *data)
+static int debugfs_min_freq_limit_set(void *data, u64 val)
 {
-	struct gxp_dev *gxp = data;
+	struct gxp_dev *gxp = (struct gxp_dev *)data;
+	struct gxp_power_manager *mgr = gxp->power_mgr;
 
-	if (gxp->pm_before_blk_off)
-		gxp->pm_before_blk_off(gxp);
-	return gxp_pm_blk_off(gxp);
+	if (val > UINT_MAX) {
+		dev_err(gxp->dev, "Requested debugfs min freq %llu must be <= %u (UINT_MAX)\n", val,
+			UINT_MAX);
+		return -EINVAL;
+	}
+
+	return gxp_pm_set_freq_limit(gxp, (u32)val, &mgr->min_freq_limit);
 }
 
+DEFINE_DEBUGFS_ATTRIBUTE(debugfs_min_freq_limit_fops, debugfs_min_freq_limit_get,
+			 debugfs_min_freq_limit_set, "%llu\n");
+
+static int debugfs_max_freq_limit_get(void *data, u64 *val)
+{
+	struct gxp_dev *gxp = (struct gxp_dev *)data;
+	struct gxp_power_manager *mgr = gxp->power_mgr;
+
+	mutex_lock(&mgr->freq_limits_lock);
+	*val = mgr->max_freq_limit;
+	mutex_unlock(&mgr->freq_limits_lock);
+	return 0;
+}
+
+static int debugfs_max_freq_limit_set(void *data, u64 val)
+{
+	struct gxp_dev *gxp = (struct gxp_dev *)data;
+	struct gxp_power_manager *mgr = gxp->power_mgr;
+
+	if (val > UINT_MAX) {
+		dev_err(gxp->dev, "Requested debugfs max freq %llu must be <= %u (UINT_MAX)\n", val,
+			UINT_MAX);
+		return -EINVAL;
+	}
+
+	return gxp_pm_set_freq_limit(gxp, (u32)val, &mgr->max_freq_limit);
+}
+
+DEFINE_DEBUGFS_ATTRIBUTE(debugfs_max_freq_limit_fops, debugfs_max_freq_limit_get,
+			 debugfs_max_freq_limit_set, "%llu\n");
+#endif /* GXP_HAS_MCU */
+
 static int debugfs_wakelock_set(void *data, u64 val)
 {
 	struct gxp_dev *gxp = (struct gxp_dev *)data;
@@ -848,6 +938,42 @@
 			 debugfs_blk_powerstate_get, debugfs_blk_powerstate_set,
 			 "%llx\n");
 
+static int gxp_pm_power_up(void *data)
+{
+	struct gxp_dev *gxp = data;
+	int ret = gxp_pm_blk_on(gxp);
+
+	if (ret) {
+		dev_err(gxp->dev, "Failed to power on BLK_AUR (ret=%d)\n", ret);
+		return ret;
+	}
+
+	if (gxp->pm_after_blk_on) {
+		ret = gxp->pm_after_blk_on(gxp);
+		if (ret) {
+			gxp_pm_blk_off(gxp);
+			return ret;
+		}
+	}
+
+#if GXP_HAS_MCU
+	mutex_lock(&gxp->power_mgr->freq_limits_lock);
+	if (gxp->power_mgr->min_freq_limit || gxp->power_mgr->max_freq_limit)
+		gxp_pm_update_freq_limits_locked(gxp);
+	mutex_unlock(&gxp->power_mgr->freq_limits_lock);
+#endif /* GXP_HAS_MCU */
+	return 0;
+}
+
+static int gxp_pm_power_down(void *data)
+{
+	struct gxp_dev *gxp = data;
+
+	if (gxp->pm_before_blk_off)
+		gxp->pm_before_blk_off(gxp);
+	return gxp_pm_blk_off(gxp);
+}
+
 static void gxp_pm_on_busy(struct gxp_dev *gxp)
 {
 	set_cmu_pll_aur_mux_state(gxp, AUR_CMU_MUX_NORMAL);
@@ -862,6 +988,12 @@
 	}
 }
 
+/**
+ * gxp_pm_parse_pmu_base() - Parse the pmu-aur-status from device tree.
+ * @gxp: The gxp_dev object.
+ *
+ * Parse the pmu status from property or fall back to try finding from reg instead.
+ */
 static void gxp_pm_parse_pmu_base(struct gxp_dev *gxp)
 {
 	u32 reg;
@@ -869,31 +1001,28 @@
 	struct platform_device *pdev =
 		container_of(gxp->dev, struct platform_device, dev);
 	struct device *dev = gxp->dev;
-
-	/* TODO(b/309801480): Remove after DT updated */
-	r = platform_get_resource_byname(pdev, IORESOURCE_MEM, "pmu_aur_status");
-	if (r) {
-		gxp->power_mgr->aur_status = devm_ioremap_resource(gxp->dev, r);
-		if (IS_ERR(gxp->power_mgr->aur_status)) {
-			dev_warn(gxp->dev, "Failed to map PMU register base, ret=%ld\n",
-				 PTR_ERR(gxp->power_mgr->aur_status));
-			gxp->power_mgr->aur_status = NULL;
-		}
-	}
+	void __iomem *aur_status = ERR_PTR(-ENOENT);
 
 	/* "pmu-aur-status" DT property takes precedence over reg entry */
 	if (of_find_property(dev->of_node, "pmu-aur-status", NULL) &&
-			!of_property_read_u32_index(dev->of_node, "pmu-aur-status", 0, &reg)) {
-		gxp->power_mgr->aur_status = devm_ioremap(dev, reg, 0x4);
-		if (IS_ERR(gxp->power_mgr->aur_status)) {
-			dev_warn(gxp->dev, "Failed to map PMU register base, ret=%ld\n",
-				 PTR_ERR(gxp->power_mgr->aur_status));
-			gxp->power_mgr->aur_status = NULL;
-		}
-	} else {
-		if (!r)
-			dev_warn(gxp->dev, "Failed to find PMU register base\n");
+	    !of_property_read_u32_index(dev->of_node, "pmu-aur-status", 0, &reg)) {
+		aur_status = devm_ioremap(dev, reg, 0x4);
 	}
+
+	if (!IS_ERR(aur_status))
+		goto out;
+
+	/* Fall back to try pasring from resource if the property not found. */
+	r = platform_get_resource_byname(pdev, IORESOURCE_MEM, "pmu_aur_status");
+	if (r)
+		aur_status = devm_ioremap_resource(gxp->dev, r);
+
+out:
+	if (IS_ERR(aur_status))
+		dev_warn(gxp->dev, "Failed to get PMU register base, ret=%ld\n",
+			 PTR_ERR(aur_status));
+
+	gxp->power_mgr->aur_status = IS_ERR(aur_status) ? NULL : aur_status;
 }
 
 int gxp_pm_init(struct gxp_dev *gxp)
@@ -906,6 +1035,7 @@
 		.power_down = gxp_pm_power_down,
 	};
 	uint i;
+	int ret;
 
 	mgr = devm_kzalloc(gxp->dev, sizeof(*mgr), GFP_KERNEL);
 	if (!mgr)
@@ -914,8 +1044,8 @@
 
 	mgr->pm = gcip_pm_create(&args);
 	if (IS_ERR(mgr->pm)) {
-		devm_kfree(gxp->dev, mgr);
-		return PTR_ERR(mgr->pm);
+		ret = PTR_ERR(mgr->pm);
+		goto err_free_mgr;
 	}
 
 	mutex_init(&mgr->pm_lock);
@@ -937,8 +1067,11 @@
 	}
 	mutex_init(&mgr->set_acpm_state_work_lock);
 	mutex_init(&mgr->req_pm_qos_work_lock);
-	gxp->power_mgr->wq =
-		create_singlethread_workqueue("gxp_power_work_queue");
+	gxp->power_mgr->wq = create_singlethread_workqueue("gxp_power_work_queue");
+	if (!gxp->power_mgr->wq) {
+		ret = -ENOMEM;
+		goto err_pm_destroy;
+	}
 	gxp->power_mgr->force_mux_normal_count = 0;
 	gxp->power_mgr->blk_switch_count = 0l;
 	spin_lock_init(&gxp->power_mgr->busy_lock);
@@ -951,12 +1084,24 @@
 	gxp_pm_chip_init(gxp);
 
 	gxp->debugfs_wakelock_held = false;
-	debugfs_create_file(DEBUGFS_WAKELOCK, 0200, gxp->d_entry, gxp,
-			    &debugfs_wakelock_fops);
+#if GXP_HAS_MCU
+	mutex_init(&mgr->freq_limits_lock);
+	debugfs_create_file(DEBUGFS_MIN_FREQ, 0600, gxp->d_entry, gxp,
+			    &debugfs_min_freq_limit_fops);
+	debugfs_create_file(DEBUGFS_MAX_FREQ, 0600, gxp->d_entry, gxp,
+			    &debugfs_max_freq_limit_fops);
+#endif /* GXP_HAS_MCU */
+	debugfs_create_file(DEBUGFS_WAKELOCK, 0200, gxp->d_entry, gxp, &debugfs_wakelock_fops);
 	debugfs_create_file(DEBUGFS_BLK_POWERSTATE, 0600, gxp->d_entry, gxp,
 			    &debugfs_blk_powerstate_fops);
 
 	return 0;
+
+err_pm_destroy:
+	gcip_pm_destroy(mgr->pm);
+err_free_mgr:
+	devm_kfree(gxp->dev, mgr);
+	return ret;
 }
 
 int gxp_pm_destroy(struct gxp_dev *gxp)
@@ -966,6 +1111,10 @@
 	if (IS_GXP_TEST && !mgr)
 		return 0;
 
+#if GXP_HAS_MCU
+	debugfs_remove(debugfs_lookup(DEBUGFS_MIN_FREQ, gxp->d_entry));
+	debugfs_remove(debugfs_lookup(DEBUGFS_MAX_FREQ, gxp->d_entry));
+#endif /* GXP_HAS_MCU */
 	debugfs_remove(debugfs_lookup(DEBUGFS_BLK_POWERSTATE, gxp->d_entry));
 	debugfs_remove(debugfs_lookup(DEBUGFS_WAKELOCK, gxp->d_entry));
 
diff --git a/gxp-pm.h b/gxp-pm.h
index 27a1398..b3a5fc7 100644
--- a/gxp-pm.h
+++ b/gxp-pm.h
@@ -124,6 +124,13 @@
 	bool curr_low_clkmux;
 	/* Last requested clock mux state */
 	bool last_scheduled_low_clkmux;
+	/*
+	 * Min/Max frequency limits, in kHz, requested via debugfs.
+	 * Protected by `freq_limits_lock`.
+	 */
+	struct mutex freq_limits_lock;
+	u32 min_freq_limit;
+	u32 max_freq_limit;
 	int curr_state;
 	int curr_memory_state; /* Note: this state will not be maintained in the MCU mode. */
 	const struct gxp_pm_ops *ops;
diff --git a/gxp-thermal.c b/gxp-thermal.c
index 9807b09..8c851fb 100644
--- a/gxp-thermal.c
+++ b/gxp-thermal.c
@@ -33,7 +33,7 @@
 	struct gxp_dev *gxp = data;
 	int ret = 0;
 
-	dev_warn(gxp->dev, "Received thermal throttling requests %lu", rate);
+	dev_warn(gxp->dev, "Received thermal throttling requests %lu.\n", rate);
 	if (!gxp_is_direct_mode(gxp)) {
 #if GXP_HAS_MCU
 		struct gxp_mcu *mcu = gxp_mcu_of(gxp);
@@ -58,7 +58,19 @@
 
 static int gxp_thermal_control(void *data, bool enable)
 {
-	return -EOPNOTSUPP;
+	int ret = -EOPNOTSUPP;
+#if GXP_HAS_MCU
+	struct gxp_dev *gxp = data;
+	struct gxp_mcu *mcu = gxp_mcu_of(gxp);
+
+	dev_warn(gxp->dev, "Received request to %s thermal throttling.\n",
+		 enable ? "enable" : "disable");
+	ret = gxp_kci_thermal_control(&mcu->kci, enable);
+	if (ret)
+		dev_err(gxp->dev, "Error on %s thermal throttling: %d.\n",
+			enable ? "enabling" : "disabling", ret);
+#endif /* GXP_HAS_MCU */
+	return ret;
 }
 
 int gxp_thermal_init(struct gxp_dev *gxp)
diff --git a/gxp-trace.c b/gxp-trace.c
index 4d0202d..ebe15c2 100644
--- a/gxp-trace.c
+++ b/gxp-trace.c
@@ -2,7 +2,7 @@
 /*
  * GXP ftrace support.
  *
- * Copyright (C) 2023 Google LLC
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #define CREATE_TRACE_POINTS
diff --git a/gxp-uci.c b/gxp-uci.c
index 9ea2e43..6d655a8 100644
--- a/gxp-uci.c
+++ b/gxp-uci.c
@@ -11,9 +11,11 @@
 #include <linux/slab.h>
 #include <uapi/linux/sched/types.h>
 
-#include <gcip/iif/iif.h>
 #include <gcip/gcip-fence-array.h>
 #include <gcip/gcip-fence.h>
+#include <iif/iif.h>
+
+#include <trace/events/gxp.h>
 
 #include "gxp-config.h"
 #include "gxp-internal.h"
@@ -253,7 +255,7 @@
 	}
 
 	ret = gcip_fence_array_submit_waiter_and_signaler(async_resp->in_fences,
-							  async_resp->out_fences);
+							  async_resp->out_fences, IIF_IP_DSP);
 	if (ret) {
 		dev_err(mailbox->dev, "Failed to submit waiter or signaler to fences, ret=%d", ret);
 		list_del_init(&async_resp->wait_list_entry);
@@ -294,10 +296,12 @@
 	spin_unlock_irqrestore(async_resp->queue_lock, flags);
 
 	gcip_fence_array_signal(async_resp->out_fences, (status != GXP_RESP_OK) ? -ETIMEDOUT : 0);
-	gcip_fence_array_waited(async_resp->in_fences);
+	gcip_fence_array_waited(async_resp->in_fences, IIF_IP_DSP);
 	if (async_resp->eventfd)
 		gxp_eventfd_signal(async_resp->eventfd);
 
+	trace_gxp_uci_rsp_end(async_resp->resp.seq);
+
 	wake_up(async_resp->dest_queue_waitq);
 }
 
@@ -356,7 +360,6 @@
 }
 
 static const struct gcip_mailbox_ops gxp_uci_gcip_mbx_ops = {
-	.get_cmd_queue_head = gxp_mailbox_gcip_ops_get_cmd_queue_head,
 	.get_cmd_queue_tail = gxp_mailbox_gcip_ops_get_cmd_queue_tail,
 	.inc_cmd_queue_tail = gxp_mailbox_gcip_ops_inc_cmd_queue_tail,
 	.acquire_cmd_queue_lock = gxp_mailbox_gcip_ops_acquire_cmd_queue_lock,
@@ -574,10 +577,7 @@
 {
 	struct gxp_mailbox *mailbox = uci->mbx;
 
-	gxp_mailbox_write_descriptor(mailbox, mailbox->descriptor_buf.dsp_addr);
-	gxp_mailbox_reset(uci->mbx);
-	gxp_mailbox_enable_interrupt(mailbox);
-	gxp_mailbox_write_status(mailbox, 1);
+	gxp_mailbox_reinit(mailbox);
 
 	return 0;
 }
diff --git a/gxp-usage-stats.c b/gxp-usage-stats.c
index 44d6960..1268c25 100644
--- a/gxp-usage-stats.c
+++ b/gxp-usage-stats.c
@@ -73,7 +73,7 @@
 	return AUR_NUM_POWER_STATE;
 }
 
-static int get_default_dvfs_freq(int idx, void *data)
+static unsigned int get_default_dvfs_freq(int idx, void *data)
 {
 	if (idx >= AUR_NUM_POWER_STATE)
 		return 0;
diff --git a/gxp-vd.c b/gxp-vd.c
index 81237d6..9509c28 100644
--- a/gxp-vd.c
+++ b/gxp-vd.c
@@ -183,11 +183,11 @@
 {
 	struct gxp_dev *gxp = vd->gxp;
 	int ret;
-	const size_t ro_size = GXP_FW_DATA_SYSCFG_SIZE / 2;
+	const size_t ro_size = res->size / 2;
 
 	if (res->daddr == 0)
 		return 0;
-	if (res->size != GXP_FW_DATA_SYSCFG_SIZE) {
+	if (!res->size || !IS_ALIGNED(res->size, gcip_iommu_domain_granule(vd->domain) * 2)) {
 		dev_err(gxp->dev, "invalid system cfg size: %#llx", res->size);
 		return -EINVAL;
 	}
@@ -239,8 +239,7 @@
  * To keep compatibility, if not both mapping[0, 1] present then this function
  * falls back to map the MCU-core shared region with hard-coded IOVA and size.
  */
-static int map_cfg_regions(struct gxp_virtual_device *vd,
-			   struct gcip_image_config *img_cfg)
+static int map_cfg_regions(struct gxp_virtual_device *vd, struct gcip_image_config *img_cfg)
 {
 	struct gxp_dev *gxp = vd->gxp;
 	struct gxp_mapped_resource pool;
@@ -318,55 +317,27 @@
 }
 
 static int gxp_vd_imgcfg_map(void *data, dma_addr_t daddr, phys_addr_t paddr,
-			     size_t size, unsigned int flags)
+			     size_t size, unsigned int cfg_map_flags, unsigned int cfg_op_flags)
 {
 	struct gxp_virtual_device *vd = data;
 
-	if (flags & GCIP_IMAGE_CONFIG_FLAGS_SECURE)
+	if (cfg_op_flags & GCIP_IMAGE_CONFIG_FLAGS_SECURE)
 		return 0;
 
 	return map_ns_region(vd, daddr, size);
 }
 
 static void gxp_vd_imgcfg_unmap(void *data, dma_addr_t daddr, size_t size,
-				unsigned int flags)
+				unsigned int cfg_map_flags, unsigned int cfg_op_flags)
 {
 	struct gxp_virtual_device *vd = data;
 
-	if (flags & GCIP_IMAGE_CONFIG_FLAGS_SECURE)
+	if (cfg_op_flags & GCIP_IMAGE_CONFIG_FLAGS_SECURE)
 		return;
 
 	unmap_ns_region(vd, daddr);
 }
 
-/* TODO(b/299037074): Remove core's accesses to LPM. */
-static int map_remote_lpm(struct gxp_virtual_device *vd,
-			  struct gcip_image_config *img_cfg)
-{
-	struct gxp_mapped_resource res;
-	int ret;
-
-	if (img_cfg->num_iommu_mappings != REMOTE_LPM_IDX + 1)
-		/* Core doesn't require remote lpm */
-		return 0;
-
-	res.daddr = img_cfg->iommu_mappings[REMOTE_LPM_IDX].virt_address;
-	res.paddr = (img_cfg->iommu_mappings[REMOTE_LPM_IDX].image_config_value) &
-		    GCIP_IMG_CFG_ADDR_MASK;
-	res.size = gcip_config_to_size(img_cfg->iommu_mappings[REMOTE_LPM_IDX].image_config_value);
-	ret = map_resource(vd, &res);
-	if (ret)
-		return ret;
-
-	vd->lpm = res;
-	return 0;
-}
-
-static void unmap_remote_lpm(struct gxp_virtual_device *vd)
-{
-	unmap_resource(vd, &vd->lpm);
-}
-
 static int
 map_fw_image_config(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
 		    struct gxp_firmware_loader_manager *fw_loader_mgr)
@@ -395,13 +366,6 @@
 		goto err;
 	}
 
-	ret = map_remote_lpm(vd, cfg);
-	if (ret) {
-		dev_err(gxp->dev, "Remote LPM mapping failed");
-		unmap_cfg_regions(vd);
-		goto err;
-	}
-
 	return 0;
 err:
 	gcip_image_config_clear(&vd->cfg_parser);
@@ -411,7 +375,6 @@
 static void unmap_fw_image_config(struct gxp_dev *gxp,
 				  struct gxp_virtual_device *vd)
 {
-	unmap_remote_lpm(vd);
 	unmap_cfg_regions(vd);
 	gcip_image_config_clear(&vd->cfg_parser);
 }
@@ -699,6 +662,8 @@
 	mutex_init(&vd->fence_list_lock);
 	INIT_LIST_HEAD(&vd->gxp_fence_list);
 	mutex_init(&vd->debug_dump_lock);
+	init_waitqueue_head(&vd->finished_dump_processing_waitq);
+	atomic_set(&vd->core_dump_generated_list, 0);
 
 #ifdef GXP_USE_DEFAULT_DOMAIN
 	vd->domain = gxp_iommu_get_domain_for_dev(gxp);
@@ -1050,8 +1015,46 @@
 	debug_dump_unlock(vd);
 }
 
-static inline uint select_core(struct gxp_virtual_device *vd, uint virt_core,
-			       uint phys_core)
+void gxp_vd_check_and_wait_for_debug_dump(struct gxp_virtual_device *vd)
+{
+	struct gxp_dev *gxp = vd->gxp;
+	bool vd_crashed = 0;
+	uint phys_core;
+	uint remaining_time;
+
+	if (!gxp_is_direct_mode(gxp))
+		return;
+
+	/*
+	 * Check if any of the cores for the given virtual device has generated the debug dump or
+	 * has been requested to generate the forced debug dump.
+	 */
+	for (phys_core = 0; phys_core < GXP_NUM_CORES; phys_core++) {
+		if (!(vd->core_list & BIT(phys_core)))
+			continue;
+
+		vd_crashed |= gxp_firmware_get_generate_debug_dump(gxp, vd, phys_core) |
+			      gxp_firmware_get_debug_dump_generated(gxp, vd, phys_core);
+	}
+
+	if (vd_crashed) {
+		/*
+		 * Successive prccessing of debug dumps demand a delay a second. This
+		 * delay is due to the current implementation of the SSCD module which
+		 * generates the dump files whose names are at precision of a second i.e.
+		 * coredump_<SUBSYSTEM_NAME>_<%Y-%m-%d_%H-%M-%S>.bin Thus max wait time is
+		 * kept to be GXP_NUM_CORES seconds.
+		 */
+		remaining_time = wait_event_timeout(
+			vd->finished_dump_processing_waitq,
+			atomic_read(&vd->core_dump_generated_list) == vd->core_list,
+			msecs_to_jiffies(GXP_NUM_CORES * SSCD_REPORT_WAIT_TIME));
+		if (!remaining_time)
+			dev_warn(gxp->dev, "Debug dump processing timedout.\n");
+	}
+}
+
+static inline uint select_core(struct gxp_virtual_device *vd, uint virt_core, uint phys_core)
 {
 	return virt_core;
 }
diff --git a/gxp-vd.h b/gxp-vd.h
index f49479d..5dd3fb2 100644
--- a/gxp-vd.h
+++ b/gxp-vd.h
@@ -155,6 +155,16 @@
 	bool mcu_crashed;
 	/* The manager of IOMMU reserve regions. */
 	struct gcip_iommu_reserve_manager *iommu_reserve_mgr;
+	/*
+	 * Bitmap of cores for which KD has processed the core dump. This flag is used only
+	 * in direct mode.
+	 */
+	atomic_t core_dump_generated_list;
+	/*
+	 * Wait queue to wait till the debug dump processing finishes for all the cores for
+	 * the current VD before powering them down. This flag is used only in direct mode.
+	 */
+	wait_queue_head_t finished_dump_processing_waitq;
 };
 
 /*
@@ -230,6 +240,23 @@
  */
 void gxp_vd_stop(struct gxp_virtual_device *vd);
 
+/**
+ * gxp_vd_check_and_wait_for_debug_dump() - Checks if any of the cores for the given vd has
+ *                                          generated the debug dump or has been requested to
+ *                                          generate the forced debug dump. If yes then waits
+ *                                          till all the debug dumps are processed.
+ * @vd: The virtual device to be checked for the crash.
+ *
+ * The caller should never lock the gxp->vd_semaphore before calling this function. The logic
+ * inside this function waits till the debug dumps for all the cores for the given VD are
+ * processed. Processing of debug dump logic requires acquiring of gxp->vd_semaphore which will
+ * end up in deadlock if this function is called after locking the gxp->vd_semaphore.
+ *
+ * This function is only meaningful in direct mode. On MCU mode it returns directly.
+ */
+
+void gxp_vd_check_and_wait_for_debug_dump(struct gxp_virtual_device *vd);
+
 /*
  * Returns the physical core ID for the specified virtual_core belonging to
  * this virtual device or -EINVAL if this virtual core is not running on a
@@ -493,8 +520,14 @@
 void gxp_vd_unlink_offload_vmbox(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
 				 u32 offload_client_id, u8 offload_chip_type);
 #else /* !GXP_HAS_MCU */
-#define gxp_vd_release_vmbox(...)
-#define gxp_vd_unlink_offload_vmbox(...)
+static inline void gxp_vd_release_vmbox(struct gxp_dev *gxp, struct gxp_virtual_device *vd)
+{
+}
+
+static inline void gxp_vd_unlink_offload_vmbox(struct gxp_dev *gxp, struct gxp_virtual_device *vd,
+					       u32 offload_client_id, u8 offload_chip_type)
+{
+}
 #endif /* GXP_HAS_MCU */
 
 /*
diff --git a/gxp.h b/gxp.h
index e4537b7..74010e5 100644
--- a/gxp.h
+++ b/gxp.h
@@ -13,7 +13,7 @@
 
 /* Interface Version */
 #define GXP_INTERFACE_VERSION_MAJOR 1
-#define GXP_INTERFACE_VERSION_MINOR 26
+#define GXP_INTERFACE_VERSION_MINOR 27
 #define GXP_INTERFACE_VERSION_BUILD 0
 
 /* mmap offsets for MCU logging and tracing buffers */
@@ -272,116 +272,6 @@
 #define GXP_ALLOCATE_VIRTUAL_DEVICE                                            \
 	_IOWR(GXP_IOCTL_BASE, 6, struct gxp_virtual_device_ioctl)
 
-#define ETM_TRACE_LSB_MASK 0x1
-#define ETM_TRACE_SYNC_MSG_PERIOD_MIN 8
-#define ETM_TRACE_SYNC_MSG_PERIOD_MAX 256
-#define ETM_TRACE_PC_MATCH_MASK_LEN_MAX 31
-
-/*
- * For all *_enable and pc_match_sense fields, only the least significant bit is
- * considered. All other bits are ignored.
- */
-struct gxp_etm_trace_start_ioctl {
-	__u16 virtual_core_id;
-	__u8 trace_ram_enable; /* Enables local trace memory. */
-	/* When set, trace output is sent out on the ATB interface. */
-	__u8 atb_enable;
-	/* Enables embedding timestamp information in trace messages. */
-	__u8 timestamp_enable;
-	/*
-	 * Determines the rate at which synchronization messages are
-	 * automatically emitted in the output trace.
-	 * Valid values: 0, 8, 16, 32, 64, 128, 256
-	 * Eg. A value of 16 means 1 synchronization message will be emitted
-	 * every 16 messages.
-	 * A value of 0 means no synchronization messages will be emitted.
-	 */
-	__u16 sync_msg_period;
-	__u8 pc_match_enable; /* PC match causes Stop trigger. */
-	/*
-	 * 32-bit address to compare to processor PC when pc_match_enable = 1.
-	 * A match for a given executed instruction triggers trace stop.
-	 * Note: trigger_pc is ignored when pc_match_enable = 0.
-	 */
-	__u32 trigger_pc;
-	/*
-	 * Indicates how many of the lower bits of trigger_pc to ignore.
-	 * Valid values: 0 to 31
-	 * Note: pc_match_mask_length is ignored when pc_match_enable = 0.
-	 */
-	__u8 pc_match_mask_length;
-	/* When 0, match when the processor's PC is in-range of trigger_pc and
-	 * mask. When 1, match when the processor's PC is out-of-range of
-	 * trigger_pc and mask.
-	 * Note: pc_match_sense is ignored when pc_match_enable = 0.
-	 */
-	__u8 pc_match_sense;
-};
-
-/*
- * Configure ETM trace registers and start ETM tracing.
- *
- * The client must hold a VIRTUAL_DEVICE wakelock.
- */
-#define GXP_ETM_TRACE_START_COMMAND                                            \
-	_IOW(GXP_IOCTL_BASE, 7, struct gxp_etm_trace_start_ioctl)
-
-/*
- * Halts trace generation via a software trigger. The virtual core id is passed
- * in as an input.
- *
- * The client must hold a VIRTUAL_DEVICE wakelock.
- */
-#define GXP_ETM_TRACE_SW_STOP_COMMAND _IOW(GXP_IOCTL_BASE, 8, __u16)
-
-/*
- * Users should call this IOCTL after tracing has been stopped for the last
- * trace session of the core. Otherwise, there is a risk of having up to 3 bytes
- * of trace data missing towards the end of the trace session.
- * This is a workaround for b/180728272 and b/181623511.
- * The virtual core id is passed in as an input.
- *
- * The client must hold a VIRTUAL_DEVICE wakelock.
- */
-#define GXP_ETM_TRACE_CLEANUP_COMMAND _IOW(GXP_IOCTL_BASE, 9, __u16)
-
-#define GXP_TRACE_HEADER_SIZE 256
-#define GXP_TRACE_RAM_SIZE 4096
-struct gxp_etm_get_trace_info_ioctl {
-	/*
-	 * Input:
-	 * The virtual core to fetch a response from.
-	 */
-	__u16 virtual_core_id;
-	/*
-	 * Input:
-	 * The type of data to retrieve.
-	 * 0: Trace Header only
-	 * 1: Trace Header + Trace Data in Trace RAM
-	 */
-	__u8 type;
-	/*
-	 * Input:
-	 * Trace header user space address to contain trace header information
-	 * that is used for decoding the trace.
-	 */
-	__u64 trace_header_addr;
-	/*
-	 * Input:
-	 * Trace data user space address to contain Trace RAM data.
-	 * Note: trace_data field will be empty if type == 0
-	 */
-	__u64 trace_data_addr;
-};
-
-/*
- * Retrieves trace header and/or trace data for decoding purposes.
- *
- * The client must hold a VIRTUAL_DEVICE wakelock.
- */
-#define GXP_ETM_GET_TRACE_INFO_COMMAND                                         \
-	_IOWR(GXP_IOCTL_BASE, 10, struct gxp_etm_get_trace_info_ioctl)
-
 #define GXP_TELEMETRY_TYPE_LOGGING (0)
 #define GXP_TELEMETRY_TYPE_TRACING (1)
 
diff --git a/include/trace/events/gxp.h b/include/trace/events/gxp.h
index 095bd7c..93b699e 100644
--- a/include/trace/events/gxp.h
+++ b/include/trace/events/gxp.h
@@ -2,7 +2,7 @@
 /*
  * Trace events for gxp
  *
- * Copyright (c) 2023 Google LLC
+ * Copyright (c) 2023-2024 Google LLC
  */
 
 #undef TRACE_SYSTEM
@@ -215,6 +215,54 @@
 
 	    TP_printk("vdid = %d", __entry->vdid));
 
+TRACE_EVENT(gxp_uci_cmd_start,
+
+	    TP_PROTO(u64 seq),
+
+	    TP_ARGS(seq),
+
+	    TP_STRUCT__entry(__field(u64, seq)),
+
+	    TP_fast_assign(__entry->seq = seq;),
+
+	    TP_printk("seq = %llu", __entry->seq));
+
+TRACE_EVENT(gxp_uci_cmd_end,
+
+	    TP_PROTO(u64 seq),
+
+	    TP_ARGS(seq),
+
+	    TP_STRUCT__entry(__field(u64, seq)),
+
+	    TP_fast_assign(__entry->seq = seq;),
+
+	    TP_printk("seq = %llu", __entry->seq));
+
+TRACE_EVENT(gxp_uci_rsp_start,
+
+	    TP_PROTO(int irq),
+
+	    TP_ARGS(irq),
+
+	    TP_STRUCT__entry(__field(int, irq)),
+
+	    TP_fast_assign(__entry->irq = irq;),
+
+	    TP_printk("irq = %d", __entry->irq));
+
+TRACE_EVENT(gxp_uci_rsp_end,
+
+	    TP_PROTO(u64 seq),
+
+	    TP_ARGS(seq),
+
+	    TP_STRUCT__entry(__field(u64, seq)),
+
+	    TP_fast_assign(__entry->seq = seq;),
+
+	    TP_printk("seq = %llu", __entry->seq));
+
 #endif /* _TRACE_GXP_H */
 
 /* This part must be outside protection */
diff --git a/mobile-soc-gsx01.c b/mobile-soc-gsx01.c
index 65e368a..51efd4e 100644
--- a/mobile-soc-gsx01.c
+++ b/mobile-soc-gsx01.c
@@ -2,21 +2,19 @@
 /*
  * SoC specific function definitions for GSx01.
  *
- * Copyright (C) 2023 Google LLC
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #include <linux/acpm_dvfs.h>
 #include <linux/slab.h>
 #include <soc/google/exynos_pm_qos.h>
 
-#include <gcip/gcip-kci.h>
 #include <gcip/gcip-slc.h>
 
 #include "gxp-config.h"
 #include "gxp-core-telemetry.h"
 #include "gxp-firmware.h"
 #include "gxp-gsx01-ssmt.h"
-#include "gxp-kci.h"
 #include "gxp-lpm.h"
 #include "gxp-pm.h"
 #include "mobile-soc-gsx01.h"
@@ -80,10 +78,14 @@
 {
 	exynos_pm_qos_add_request(&gxp->soc_data->int_min, PM_QOS_DEVICE_THROUGHPUT, 0);
 	exynos_pm_qos_add_request(&gxp->soc_data->mif_min, PM_QOS_BUS_THROUGHPUT, 0);
+	exynos_pm_qos_add_request(&gxp->soc_data->bci_min, PM_QOS_BCI_THROUGHPUT, 0);
+	exynos_pm_qos_add_request(&gxp->soc_data->dsu_min, PM_QOS_DSU_THROUGHPUT, 0);
 }
 
 void gxp_soc_pm_exit(struct gxp_dev *gxp)
 {
+	exynos_pm_qos_remove_request(&gxp->soc_data->dsu_min);
+	exynos_pm_qos_remove_request(&gxp->soc_data->bci_min);
 	exynos_pm_qos_remove_request(&gxp->soc_data->mif_min);
 	exynos_pm_qos_remove_request(&gxp->soc_data->int_min);
 }
@@ -113,6 +115,8 @@
 {
 	exynos_pm_qos_update_request(&gxp->soc_data->int_min, 0);
 	exynos_pm_qos_update_request(&gxp->soc_data->mif_min, 0);
+	exynos_pm_qos_update_request(&gxp->soc_data->bci_min, 0);
+	exynos_pm_qos_update_request(&gxp->soc_data->dsu_min, 0);
 }
 
 int gxp_soc_init(struct gxp_dev *gxp)
diff --git a/mobile-soc-gsx01.h b/mobile-soc-gsx01.h
index 2d12fe4..d817998 100644
--- a/mobile-soc-gsx01.h
+++ b/mobile-soc-gsx01.h
@@ -2,7 +2,7 @@
 /*
  * Header file for GSx01.
  *
- * Copyright (C) 2023 Google LLC
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #ifndef __MOBILE_SOC_GSX01_H__
@@ -18,6 +18,8 @@
 	/* INT/MIF requests for memory bandwidth. */
 	struct exynos_pm_qos_request int_min;
 	struct exynos_pm_qos_request mif_min;
+	struct exynos_pm_qos_request bci_min;
+	struct exynos_pm_qos_request dsu_min;
 	/* The Stream Security Mapping Table support. */
 	struct gxp_ssmt ssmt;
 	/* The SLC configurations. */
diff --git a/mobile-soc.h b/mobile-soc.h
index 9b378ce..42a5ec3 100644
--- a/mobile-soc.h
+++ b/mobile-soc.h
@@ -2,7 +2,7 @@
 /*
  * SoC-specific function headers.
  *
- * Copyright (C) 2023 Google LLC
+ * Copyright (C) 2023-2024 Google LLC
  */
 
 #ifndef __MOBILE_SOC_H__