mirror of
https://github.com/torvalds/linux.git
synced 2024-11-21 19:46:16 +00:00
b7ed87ffc7
Abort all jobs that belong to contexts generating MMU faults in order to avoid flooding host with MMU IRQs. Jobs are cancelled with: - SSID_RELEASE command when OS scheduling is enabled - DESTROY_CMDQ command when HW scheduling is enabled Signed-off-by: Maciej Falkowski <maciej.falkowski@intel.com> Co-developed-by: Wachowski, Karol <karol.wachowski@intel.com> Signed-off-by: Wachowski, Karol <karol.wachowski@intel.com> Reviewed-by: Jacek Lawrynowicz <jacek.lawrynowicz@linux.intel.com> Signed-off-by: Jacek Lawrynowicz <jacek.lawrynowicz@linux.intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20240611120433.1012423-3-jacek.lawrynowicz@linux.intel.com
113 lines
3.1 KiB
C
113 lines
3.1 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Copyright (C) 2020-2024 Intel Corporation
|
|
*/
|
|
|
|
#ifndef __IVPU_IPC_H__
|
|
#define __IVPU_IPC_H__
|
|
|
|
#include <linux/interrupt.h>
|
|
#include <linux/spinlock.h>
|
|
|
|
#include "vpu_jsm_api.h"
|
|
|
|
struct ivpu_bo;
|
|
|
|
/* VPU FW boot notification */
|
|
#define IVPU_IPC_CHAN_BOOT_MSG 0x3ff
|
|
#define IVPU_IPC_BOOT_MSG_DATA_ADDR 0x424f4f54
|
|
|
|
/* The alignment to be used for IPC Buffers and IPC Data. */
|
|
#define IVPU_IPC_ALIGNMENT 64
|
|
|
|
#define IVPU_IPC_HDR_FREE 0
|
|
#define IVPU_IPC_HDR_ALLOCATED 1
|
|
|
|
/**
|
|
* struct ivpu_ipc_hdr - The IPC message header structure, exchanged
|
|
* with the VPU device firmware.
|
|
* @data_addr: The VPU address of the payload (JSM message)
|
|
* @data_size: The size of the payload.
|
|
* @channel: The channel used.
|
|
* @src_node: The Node ID of the sender.
|
|
* @dst_node: The Node ID of the intended receiver.
|
|
* @status: IPC buffer usage status
|
|
*/
|
|
struct ivpu_ipc_hdr {
|
|
u32 data_addr;
|
|
u32 data_size;
|
|
u16 channel;
|
|
u8 src_node;
|
|
u8 dst_node;
|
|
u8 status;
|
|
} __packed __aligned(IVPU_IPC_ALIGNMENT);
|
|
|
|
typedef void (*ivpu_ipc_rx_callback_t)(struct ivpu_device *vdev,
|
|
struct ivpu_ipc_hdr *ipc_hdr,
|
|
struct vpu_jsm_msg *jsm_msg);
|
|
|
|
struct ivpu_ipc_rx_msg {
|
|
struct list_head link;
|
|
struct ivpu_ipc_hdr *ipc_hdr;
|
|
struct vpu_jsm_msg *jsm_msg;
|
|
ivpu_ipc_rx_callback_t callback;
|
|
};
|
|
|
|
struct ivpu_ipc_consumer {
|
|
struct list_head link;
|
|
u32 channel;
|
|
u32 tx_vpu_addr;
|
|
u32 request_id;
|
|
bool aborted;
|
|
ivpu_ipc_rx_callback_t rx_callback;
|
|
|
|
spinlock_t rx_lock; /* Protects rx_msg_list and aborted */
|
|
struct list_head rx_msg_list;
|
|
wait_queue_head_t rx_msg_wq;
|
|
};
|
|
|
|
struct ivpu_ipc_info {
|
|
struct gen_pool *mm_tx;
|
|
struct ivpu_bo *mem_tx;
|
|
struct ivpu_bo *mem_rx;
|
|
|
|
atomic_t rx_msg_count;
|
|
|
|
spinlock_t cons_lock; /* Protects cons_list and cb_msg_list */
|
|
struct list_head cons_list;
|
|
struct list_head cb_msg_list;
|
|
|
|
atomic_t request_id;
|
|
struct mutex lock; /* Lock on status */
|
|
bool on;
|
|
};
|
|
|
|
int ivpu_ipc_init(struct ivpu_device *vdev);
|
|
void ivpu_ipc_fini(struct ivpu_device *vdev);
|
|
|
|
void ivpu_ipc_enable(struct ivpu_device *vdev);
|
|
void ivpu_ipc_disable(struct ivpu_device *vdev);
|
|
void ivpu_ipc_reset(struct ivpu_device *vdev);
|
|
|
|
void ivpu_ipc_irq_handler(struct ivpu_device *vdev);
|
|
void ivpu_ipc_irq_thread_handler(struct ivpu_device *vdev);
|
|
|
|
void ivpu_ipc_consumer_add(struct ivpu_device *vdev, struct ivpu_ipc_consumer *cons,
|
|
u32 channel, ivpu_ipc_rx_callback_t callback);
|
|
void ivpu_ipc_consumer_del(struct ivpu_device *vdev, struct ivpu_ipc_consumer *cons);
|
|
|
|
int ivpu_ipc_send(struct ivpu_device *vdev, struct ivpu_ipc_consumer *cons,
|
|
struct vpu_jsm_msg *req);
|
|
int ivpu_ipc_receive(struct ivpu_device *vdev, struct ivpu_ipc_consumer *cons,
|
|
struct ivpu_ipc_hdr *ipc_buf, struct vpu_jsm_msg *jsm_msg,
|
|
unsigned long timeout_ms);
|
|
|
|
int ivpu_ipc_send_receive_active(struct ivpu_device *vdev, struct vpu_jsm_msg *req,
|
|
enum vpu_ipc_msg_type expected_resp, struct vpu_jsm_msg *resp,
|
|
u32 channel, unsigned long timeout_ms);
|
|
int ivpu_ipc_send_receive(struct ivpu_device *vdev, struct vpu_jsm_msg *req,
|
|
enum vpu_ipc_msg_type expected_resp, struct vpu_jsm_msg *resp,
|
|
u32 channel, unsigned long timeout_ms);
|
|
|
|
#endif /* __IVPU_IPC_H__ */
|