| /* Copyright (c) 2002,2007-2014, The Linux Foundation. All rights reserved. |
| * |
| * This program is free software; you can redistribute it and/or modify |
| * it under the terms of the GNU General Public License version 2 and |
| * only version 2 as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope that it will be useful, |
| * but WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the |
| * GNU General Public License for more details. |
| * |
| */ |
| #ifndef __KGSL_DEVICE_H |
| #define __KGSL_DEVICE_H |
| |
| #include <linux/slab.h> |
| #include <linux/idr.h> |
| #include <linux/pm_qos.h> |
| #include <linux/sched.h> |
| |
| #include "kgsl.h" |
| #include "kgsl_mmu.h" |
| #include "kgsl_pwrctrl.h" |
| #include "kgsl_log.h" |
| #include "kgsl_pwrscale.h" |
| #include <linux/sync.h> |
| |
| #define KGSL_TIMEOUT_NONE 0 |
| #define KGSL_TIMEOUT_DEFAULT 0xFFFFFFFF |
| #define KGSL_TIMEOUT_PART 50 /* 50 msec */ |
| #define KGSL_TIMEOUT_LONG_IB_DETECTION 2000 /* 2 sec*/ |
| |
| #define FIRST_TIMEOUT (HZ / 2) |
| |
| |
| /* KGSL device state is initialized to INIT when platform_probe * |
| * sucessfully initialized the device. Once a device has been opened * |
| * (started) it becomes active. NAP implies that only low latency * |
| * resources (for now clocks on some platforms) are off. SLEEP implies * |
| * that the KGSL module believes a device is idle (has been inactive * |
| * past its timer) and all system resources are released. SUSPEND is * |
| * requested by the kernel and will be enforced upon all open devices. */ |
| |
| #define KGSL_STATE_NONE 0x00000000 |
| #define KGSL_STATE_INIT 0x00000001 |
| #define KGSL_STATE_ACTIVE 0x00000002 |
| #define KGSL_STATE_NAP 0x00000004 |
| #define KGSL_STATE_SLEEP 0x00000008 |
| #define KGSL_STATE_SUSPEND 0x00000010 |
| #define KGSL_STATE_HUNG 0x00000020 |
| #define KGSL_STATE_SLUMBER 0x00000080 |
| |
| #define KGSL_GRAPHICS_MEMORY_LOW_WATERMARK 0x1000000 |
| |
| #define KGSL_IS_PAGE_ALIGNED(addr) (!((addr) & (~PAGE_MASK))) |
| |
| /* |
| * KGSL event types - these are passed to the event callback when the event |
| * expires or is cancelled |
| */ |
| |
| #define KGSL_EVENT_TIMESTAMP_RETIRED 0 |
| #define KGSL_EVENT_CANCELLED 1 |
| |
| #define KGSL_FLAG_WAKE_ON_TOUCH BIT(0) |
| |
| /* |
| * "list" of event types for ftrace symbolic magic |
| */ |
| |
| #define KGSL_EVENT_TYPES \ |
| { KGSL_EVENT_TIMESTAMP_RETIRED, "retired" }, \ |
| { KGSL_EVENT_CANCELLED, "cancelled" } |
| |
| struct kgsl_device; |
| struct platform_device; |
| struct kgsl_device_private; |
| struct kgsl_context; |
| struct kgsl_power_stats; |
| struct kgsl_event; |
| struct kgsl_cmdbatch; |
| |
| struct kgsl_functable { |
| /* Mandatory functions - these functions must be implemented |
| by the client device. The driver will not check for a NULL |
| pointer before calling the hook. |
| */ |
| void (*regread) (struct kgsl_device *device, |
| unsigned int offsetwords, unsigned int *value); |
| void (*regwrite) (struct kgsl_device *device, |
| unsigned int offsetwords, unsigned int value); |
| int (*idle) (struct kgsl_device *device); |
| bool (*isidle) (struct kgsl_device *device); |
| int (*suspend_context) (struct kgsl_device *device); |
| int (*init) (struct kgsl_device *device); |
| int (*start) (struct kgsl_device *device, int priority); |
| int (*stop) (struct kgsl_device *device); |
| int (*getproperty) (struct kgsl_device *device, |
| enum kgsl_property_type type, void *value, |
| unsigned int sizebytes); |
| int (*waittimestamp) (struct kgsl_device *device, |
| struct kgsl_context *context, unsigned int timestamp, |
| unsigned int msecs); |
| unsigned int (*readtimestamp) (struct kgsl_device *device, |
| struct kgsl_context *context, enum kgsl_timestamp_type type); |
| int (*issueibcmds) (struct kgsl_device_private *dev_priv, |
| struct kgsl_context *context, struct kgsl_cmdbatch *cmdbatch, |
| uint32_t *timestamps); |
| int (*setup_pt)(struct kgsl_device *device, |
| struct kgsl_pagetable *pagetable); |
| void (*cleanup_pt)(struct kgsl_device *device, |
| struct kgsl_pagetable *pagetable); |
| void (*power_stats)(struct kgsl_device *device, |
| struct kgsl_power_stats *stats); |
| void (*irqctrl)(struct kgsl_device *device, int state); |
| unsigned int (*gpuid)(struct kgsl_device *device, unsigned int *chipid); |
| void * (*snapshot)(struct kgsl_device *device, void *snapshot, |
| int *remain, int hang); |
| irqreturn_t (*irq_handler)(struct kgsl_device *device); |
| int (*drain)(struct kgsl_device *device); |
| /* Optional functions - these functions are not mandatory. The |
| driver will check that the function pointer is not NULL before |
| calling the hook */ |
| int (*setstate) (struct kgsl_device *device, unsigned int context_id, |
| uint32_t flags); |
| struct kgsl_context *(*drawctxt_create) (struct kgsl_device_private *, |
| uint32_t *flags); |
| int (*drawctxt_detach) (struct kgsl_context *context); |
| void (*drawctxt_destroy) (struct kgsl_context *context); |
| long (*ioctl) (struct kgsl_device_private *dev_priv, |
| unsigned int cmd, void *data); |
| int (*setproperty) (struct kgsl_device_private *dev_priv, |
| enum kgsl_property_type type, void *value, |
| unsigned int sizebytes); |
| int (*postmortem_dump) (struct kgsl_device *device, int manual); |
| int (*next_event)(struct kgsl_device *device, |
| struct kgsl_event *event); |
| void (*drawctxt_sched)(struct kgsl_device *device, |
| struct kgsl_context *context); |
| void (*resume)(struct kgsl_device *device); |
| }; |
| |
| /* MH register values */ |
| struct kgsl_mh { |
| unsigned int mharb; |
| unsigned int mh_intf_cfg1; |
| unsigned int mh_intf_cfg2; |
| uint32_t mpu_base; |
| int mpu_range; |
| }; |
| |
| typedef void (*kgsl_event_func)(struct kgsl_device *, void *, u32, u32, u32); |
| |
| struct kgsl_event { |
| struct kgsl_context *context; |
| uint32_t timestamp; |
| kgsl_event_func func; |
| void *priv; |
| struct list_head list; |
| void *owner; |
| unsigned int created; |
| }; |
| |
| /** |
| * struct kgsl_cmdbatch - KGSl command descriptor |
| * @device: KGSL GPU device that the command was created for |
| * @context: KGSL context that created the command |
| * @timestamp: Timestamp assigned to the command |
| * @flags: flags |
| * @priv: Internal flags |
| * @fault_policy: Internal policy describing how to handle this command in case |
| * of a fault |
| * @fault_recovery: recovery actions actually tried for this batch |
| * @ibcount: Number of IBs in the command list |
| * @ibdesc: Pointer to the list of IBs |
| * @expires: Point in time when the cmdbatch is considered to be hung |
| * @invalid: non-zero if the dispatcher determines the command and the owning |
| * context should be invalidated |
| * @refcount: kref structure to maintain the reference count |
| * @synclist: List of context/timestamp tuples to wait for before issuing |
| * |
| * This struture defines an atomic batch of command buffers issued from |
| * userspace. |
| */ |
| struct kgsl_cmdbatch { |
| struct kgsl_device *device; |
| struct kgsl_context *context; |
| spinlock_t lock; |
| uint32_t timestamp; |
| uint32_t flags; |
| unsigned long priv; |
| unsigned long fault_policy; |
| unsigned long fault_recovery; |
| uint32_t ibcount; |
| struct kgsl_ibdesc *ibdesc; |
| unsigned long expires; |
| int invalid; |
| struct kref refcount; |
| struct list_head synclist; |
| }; |
| |
| /** |
| * enum kgsl_cmdbatch_priv - Internal cmdbatch flags |
| * @CMDBATCH_FLAG_SKIP - skip the entire command batch |
| * @CMDBATCH_FLAG_FORCE_PREAMBLE - Force the preamble on for the cmdbatch |
| * @CMDBATCH_FLAG_WFI - Force wait-for-idle for the submission |
| */ |
| |
| enum kgsl_cmdbatch_priv { |
| CMDBATCH_FLAG_SKIP = 0, |
| CMDBATCH_FLAG_FORCE_PREAMBLE, |
| CMDBATCH_FLAG_WFI, |
| }; |
| |
| struct kgsl_device { |
| struct device *dev; |
| const char *name; |
| unsigned int ver_major; |
| unsigned int ver_minor; |
| uint32_t flags; |
| enum kgsl_deviceid id; |
| |
| /* Starting physical address for GPU registers */ |
| unsigned long reg_phys; |
| |
| /* Starting Kernel virtual address for GPU registers */ |
| void *reg_virt; |
| |
| /* Total memory size for all GPU registers */ |
| unsigned int reg_len; |
| |
| /* Kernel virtual address for GPU shader memory */ |
| void *shader_mem_virt; |
| |
| /* Starting physical address for GPU shader memory */ |
| unsigned long shader_mem_phys; |
| |
| /* GPU shader memory size */ |
| unsigned int shader_mem_len; |
| struct kgsl_memdesc memstore; |
| const char *iomemname; |
| const char *shadermemname; |
| |
| struct kgsl_mh mh; |
| struct kgsl_mmu mmu; |
| struct completion hwaccess_gate; |
| struct completion cmdbatch_gate; |
| const struct kgsl_functable *ftbl; |
| struct work_struct idle_check_ws; |
| struct timer_list idle_timer; |
| struct kgsl_pwrctrl pwrctrl; |
| int open_count; |
| |
| struct mutex mutex; |
| uint32_t state; |
| uint32_t requested_state; |
| |
| atomic_t active_cnt; |
| |
| wait_queue_head_t wait_queue; |
| wait_queue_head_t active_cnt_wq; |
| struct workqueue_struct *work_queue; |
| struct device *parentdev; |
| struct dentry *d_debugfs; |
| struct idr context_idr; |
| rwlock_t context_lock; |
| |
| void *snapshot; /* Pointer to the snapshot memory region */ |
| int snapshot_maxsize; /* Max size of the snapshot region */ |
| int snapshot_size; /* Current size of the snapshot region */ |
| u32 snapshot_timestamp; /* Timestamp of the last valid snapshot */ |
| u32 snapshot_faultcount; /* Total number of faults since boot */ |
| int snapshot_frozen; /* 1 if the snapshot output is frozen until |
| it gets read by the user. This avoids |
| losing the output on multiple hangs */ |
| struct kobject snapshot_kobj; |
| |
| /* |
| * List of GPU buffers that have been frozen in memory until they can be |
| * dumped |
| */ |
| struct list_head snapshot_obj_list; |
| |
| /* Logging levels */ |
| int cmd_log; |
| int ctxt_log; |
| int drv_log; |
| int mem_log; |
| int pwr_log; |
| int pm_dump_enable; |
| struct kgsl_pwrscale pwrscale; |
| struct kobject pwrscale_kobj; |
| struct work_struct ts_expired_ws; |
| struct list_head events; |
| struct list_head events_pending_list; |
| unsigned int events_last_timestamp; |
| |
| /* Postmortem Control switches */ |
| int pm_regs_enabled; |
| int pm_ib_enabled; |
| |
| int reset_counter; /* Track how many GPU core resets have occured */ |
| int cff_dump_enable; |
| }; |
| |
| void kgsl_process_events(struct work_struct *work); |
| |
| #define KGSL_DEVICE_COMMON_INIT(_dev) \ |
| .hwaccess_gate = COMPLETION_INITIALIZER((_dev).hwaccess_gate),\ |
| .cmdbatch_gate = COMPLETION_INITIALIZER((_dev).cmdbatch_gate),\ |
| .idle_check_ws = __WORK_INITIALIZER((_dev).idle_check_ws,\ |
| kgsl_idle_check),\ |
| .ts_expired_ws = __WORK_INITIALIZER((_dev).ts_expired_ws,\ |
| kgsl_process_events),\ |
| .context_idr = IDR_INIT((_dev).context_idr),\ |
| .events = LIST_HEAD_INIT((_dev).events),\ |
| .events_pending_list = LIST_HEAD_INIT((_dev).events_pending_list), \ |
| .wait_queue = __WAIT_QUEUE_HEAD_INITIALIZER((_dev).wait_queue),\ |
| .active_cnt_wq = __WAIT_QUEUE_HEAD_INITIALIZER((_dev).active_cnt_wq),\ |
| .mutex = __MUTEX_INITIALIZER((_dev).mutex),\ |
| .state = KGSL_STATE_INIT,\ |
| .ver_major = DRIVER_VERSION_MAJOR,\ |
| .ver_minor = DRIVER_VERSION_MINOR |
| |
| |
| /* bits for struct kgsl_context.priv */ |
| /* the context has been destroyed by userspace and is no longer using the gpu */ |
| #define KGSL_CONTEXT_DETACHED 0 |
| /* the context has caused a pagefault */ |
| #define KGSL_CONTEXT_PAGEFAULT 1 |
| |
| struct kgsl_process_private; |
| /** |
| * struct kgsl_context - Master structure for a KGSL context object |
| * @refcount: kref object for reference counting the context |
| * @id: integer identifier for the context |
| * @priv: in-kernel context flags, use KGSL_CONTEXT_* values |
| * @dev_priv: pointer to the owning device instance |
| * @reset_status: status indication whether a gpu reset occured and whether |
| * this context was responsible for causing it |
| * @wait_on_invalid_ts: flag indicating if this context has tried to wait on a |
| * bad timestamp |
| * @timeline: sync timeline used to create fences that can be signaled when a |
| * sync_pt timestamp expires |
| * @events: list head of pending events for this context |
| * @events_list: list node for the list of all contexts that have pending events |
| * @pid: process that owns this context. |
| * @tid: task that created this context. |
| * @pagefault_ts: global timestamp of the pagefault, if KGSL_CONTEXT_PAGEFAULT |
| * is set. |
| * @flags: flags from userspace controlling the behavior of this context |
| * @pwr_constraint: power constraint from userspace for this context |
| */ |
| struct kgsl_context { |
| struct kref refcount; |
| uint32_t id; |
| pid_t pid; |
| pid_t tid; |
| struct kgsl_device_private *dev_priv; |
| struct kgsl_process_private *proc_priv; |
| unsigned long priv; |
| struct kgsl_device *device; |
| unsigned int reset_status; |
| bool wait_on_invalid_ts; |
| struct sync_timeline *timeline; |
| struct list_head events; |
| struct list_head events_list; |
| unsigned int pagefault_ts; |
| unsigned int flags; |
| struct kgsl_pwr_constraint pwr_constraint; |
| }; |
| |
| /** |
| * struct kgsl_process_private - Private structure for a KGSL process (across |
| * all devices) |
| * @priv: Internal flags, use KGSL_PROCESS_* values |
| * @pid: ID for the task owner of the process |
| * @mem_lock: Spinlock to protect the process memory lists |
| * @refcount: kref object for reference counting the process |
| * @process_private_mutex: Mutex to synchronize access to the process struct |
| * @mem_rb: RB tree node for the memory owned by this process |
| * @idr: Iterator for assigning IDs to memory allocations |
| * @pagetable: Pointer to the pagetable owned by this process |
| * @kobj: Pointer to a kobj for the sysfs directory for this process |
| * @debug_root: Pointer to the debugfs root for this process |
| * @stats: Memory allocation statistics for this process |
| */ |
| struct kgsl_process_private { |
| unsigned long priv; |
| pid_t pid; |
| spinlock_t mem_lock; |
| |
| /* General refcount for process private struct obj */ |
| struct kref refcount; |
| /* Mutex to synchronize access to each process_private struct obj */ |
| struct mutex process_private_mutex; |
| |
| struct rb_root mem_rb; |
| struct idr mem_idr; |
| struct kgsl_pagetable *pagetable; |
| struct list_head list; |
| struct kobject kobj; |
| struct dentry *debug_root; |
| |
| struct { |
| unsigned int cur; |
| unsigned int max; |
| } stats[KGSL_MEM_ENTRY_MAX]; |
| }; |
| |
| /** |
| * enum kgsl_process_priv_flags - Private flags for kgsl_process_private |
| * @KGSL_PROCESS_INIT: Set if the process structure has been set up |
| */ |
| enum kgsl_process_priv_flags { |
| KGSL_PROCESS_INIT = 0, |
| }; |
| |
| struct kgsl_device_private { |
| struct kgsl_device *device; |
| struct kgsl_process_private *process_priv; |
| }; |
| |
| struct kgsl_device *kgsl_get_device(int dev_idx); |
| |
| int kgsl_add_event(struct kgsl_device *device, u32 id, u32 ts, |
| kgsl_event_func func, void *priv, void *owner); |
| |
| void kgsl_cancel_event(struct kgsl_device *device, struct kgsl_context *context, |
| unsigned int timestamp, kgsl_event_func func, void *priv); |
| |
| static inline void kgsl_process_add_stats(struct kgsl_process_private *priv, |
| unsigned int type, size_t size) |
| { |
| priv->stats[type].cur += size; |
| if (priv->stats[type].max < priv->stats[type].cur) |
| priv->stats[type].max = priv->stats[type].cur; |
| } |
| |
| static inline void kgsl_regread(struct kgsl_device *device, |
| unsigned int offsetwords, |
| unsigned int *value) |
| { |
| device->ftbl->regread(device, offsetwords, value); |
| } |
| |
| static inline void kgsl_regwrite(struct kgsl_device *device, |
| unsigned int offsetwords, |
| unsigned int value) |
| { |
| device->ftbl->regwrite(device, offsetwords, value); |
| } |
| |
| static inline int kgsl_idle(struct kgsl_device *device) |
| { |
| return device->ftbl->idle(device); |
| } |
| |
| static inline unsigned int kgsl_gpuid(struct kgsl_device *device, |
| unsigned int *chipid) |
| { |
| return device->ftbl->gpuid(device, chipid); |
| } |
| |
| static inline unsigned int kgsl_readtimestamp(struct kgsl_device *device, |
| struct kgsl_context *context, |
| enum kgsl_timestamp_type type) |
| { |
| return device->ftbl->readtimestamp(device, context, type); |
| } |
| |
| static inline int kgsl_create_device_sysfs_files(struct device *root, |
| const struct device_attribute **list) |
| { |
| int ret = 0, i; |
| for (i = 0; list[i] != NULL; i++) |
| ret |= device_create_file(root, list[i]); |
| return ret; |
| } |
| |
| static inline void kgsl_remove_device_sysfs_files(struct device *root, |
| const struct device_attribute **list) |
| { |
| int i; |
| for (i = 0; list[i] != NULL; i++) |
| device_remove_file(root, list[i]); |
| } |
| |
| static inline struct kgsl_mmu * |
| kgsl_get_mmu(struct kgsl_device *device) |
| { |
| return (struct kgsl_mmu *) (device ? &device->mmu : NULL); |
| } |
| |
| static inline struct kgsl_device *kgsl_device_from_dev(struct device *dev) |
| { |
| int i; |
| |
| for (i = 0; i < KGSL_DEVICE_MAX; i++) { |
| if (kgsl_driver.devp[i] && kgsl_driver.devp[i]->dev == dev) |
| return kgsl_driver.devp[i]; |
| } |
| |
| return NULL; |
| } |
| |
| static inline int kgsl_create_device_workqueue(struct kgsl_device *device) |
| { |
| device->work_queue = create_singlethread_workqueue(device->name); |
| if (!device->work_queue) { |
| KGSL_DRV_ERR(device, |
| "create_singlethread_workqueue(%s) failed\n", |
| device->name); |
| return -EINVAL; |
| } |
| return 0; |
| } |
| |
| int kgsl_check_timestamp(struct kgsl_device *device, |
| struct kgsl_context *context, unsigned int timestamp); |
| |
| int kgsl_device_platform_probe(struct kgsl_device *device); |
| |
| void kgsl_device_platform_remove(struct kgsl_device *device); |
| |
| const char *kgsl_pwrstate_to_str(unsigned int state); |
| |
| int kgsl_device_snapshot_init(struct kgsl_device *device); |
| int kgsl_device_snapshot(struct kgsl_device *device, int hang); |
| void kgsl_device_snapshot_close(struct kgsl_device *device); |
| |
| static inline struct kgsl_device_platform_data * |
| kgsl_device_get_drvdata(struct kgsl_device *dev) |
| { |
| struct platform_device *pdev = |
| container_of(dev->parentdev, struct platform_device, dev); |
| |
| return pdev->dev.platform_data; |
| } |
| |
| void kgsl_context_destroy(struct kref *kref); |
| |
| int kgsl_context_init(struct kgsl_device_private *, struct kgsl_context |
| *context); |
| int kgsl_context_detach(struct kgsl_context *context); |
| |
| /** |
| * kgsl_context_put() - Release context reference count |
| * @context: Pointer to the KGSL context to be released |
| * |
| * Reduce the reference count on a KGSL context and destroy it if it is no |
| * longer needed |
| */ |
| static inline void |
| kgsl_context_put(struct kgsl_context *context) |
| { |
| if (context) |
| kref_put(&context->refcount, kgsl_context_destroy); |
| } |
| |
| /** |
| * kgsl_context_detached() - check if a context is detached |
| * @context: the context |
| * |
| * Check if a context has been destroyed by userspace and is only waiting |
| * for reference counts to go away. This check is used to weed out |
| * contexts that shouldn't use the gpu so NULL is considered detached. |
| */ |
| static inline bool kgsl_context_detached(struct kgsl_context *context) |
| { |
| return (context == NULL || test_bit(KGSL_CONTEXT_DETACHED, |
| &context->priv)); |
| } |
| |
| |
| /** |
| * kgsl_context_get() - get a pointer to a KGSL context |
| * @device: Pointer to the KGSL device that owns the context |
| * @id: Context ID |
| * |
| * Find the context associated with the given ID number, increase the reference |
| * count on it and return it. The caller must make sure that this call is |
| * paired with a kgsl_context_put. This function is for internal use because it |
| * doesn't validate the ownership of the context with the calling process - use |
| * kgsl_context_get_owner for that |
| */ |
| static inline struct kgsl_context *kgsl_context_get(struct kgsl_device *device, |
| uint32_t id) |
| { |
| int result = 0; |
| struct kgsl_context *context = NULL; |
| |
| read_lock(&device->context_lock); |
| |
| context = idr_find(&device->context_idr, id); |
| |
| /* Don't return a context that has been detached */ |
| if (kgsl_context_detached(context)) |
| context = NULL; |
| else |
| result = kref_get_unless_zero(&context->refcount); |
| |
| read_unlock(&device->context_lock); |
| |
| if (!result) |
| return NULL; |
| return context; |
| } |
| |
| /** |
| * _kgsl_context_get() - lightweight function to just increment the ref count |
| * @context: Pointer to the KGSL context |
| * |
| * Get a reference to the specified KGSL context structure. This is a |
| * lightweight way to just increase the refcount on a known context rather than |
| * walking through kgsl_context_get and searching the iterator |
| */ |
| static inline int _kgsl_context_get(struct kgsl_context *context) |
| { |
| int ret = 0; |
| |
| if (context) { |
| ret = kref_get_unless_zero(&context->refcount); |
| /* |
| * We shouldn't realistically fail kref_get_unless_zero unless |
| * we did something really dumb so make the failure both public |
| * and painful |
| */ |
| |
| WARN_ON(!ret); |
| } |
| |
| return ret; |
| } |
| |
| /** |
| * kgsl_context_get_owner() - get a pointer to a KGSL context in a specific |
| * process |
| * @dev_priv: Pointer to the process struct |
| * @id: Context ID to return |
| * |
| * Find the context associated with the given ID number, increase the reference |
| * count on it and return it. The caller must make sure that this call is |
| * paired with a kgsl_context_put. This function validates that the context id |
| * given is owned by the dev_priv instancet that is passed in. See |
| * kgsl_context_get for the internal version that doesn't do the check |
| */ |
| static inline struct kgsl_context *kgsl_context_get_owner( |
| struct kgsl_device_private *dev_priv, uint32_t id) |
| { |
| struct kgsl_context *context; |
| |
| context = kgsl_context_get(dev_priv->device, id); |
| |
| /* Verify that the context belongs to current calling process. */ |
| if (context != NULL && context->pid != dev_priv->process_priv->pid) { |
| kgsl_context_put(context); |
| return NULL; |
| } |
| |
| return context; |
| } |
| |
| /** |
| * kgsl_context_cancel_events() - Cancel all events for a context |
| * @device: Pointer to the KGSL device structure for the GPU |
| * @context: Pointer to the KGSL context |
| * |
| * Signal all pending events on the context with KGSL_EVENT_CANCELLED |
| */ |
| static inline void kgsl_context_cancel_events(struct kgsl_device *device, |
| struct kgsl_context *context) |
| { |
| kgsl_signal_events(device, context, KGSL_EVENT_CANCELLED); |
| } |
| |
| /** |
| * kgsl_context_cancel_events_timestamp() - cancel events for a given timestamp |
| * @device: Pointer to the KGSL device that owns the context |
| * @context: Pointer to the context that owns the event or NULL for global |
| * @timestamp: Timestamp to cancel events for |
| * |
| * Cancel events pending for a specific timestamp |
| */ |
| static inline void kgsl_cancel_events_timestamp(struct kgsl_device *device, |
| struct kgsl_context *context, unsigned int timestamp) |
| { |
| kgsl_signal_event(device, context, timestamp, KGSL_EVENT_CANCELLED); |
| } |
| |
| void kgsl_cmdbatch_destroy(struct kgsl_cmdbatch *cmdbatch); |
| |
| void kgsl_cmdbatch_destroy_object(struct kref *kref); |
| |
| /** |
| * kgsl_cmdbatch_put() - Decrement the refcount for a command batch object |
| * @cmdbatch: Pointer to the command batch object |
| */ |
| static inline void kgsl_cmdbatch_put(struct kgsl_cmdbatch *cmdbatch) |
| { |
| if (cmdbatch) |
| kref_put(&cmdbatch->refcount, kgsl_cmdbatch_destroy_object); |
| } |
| |
| /** |
| * kgsl_cmdbatch_sync_pending() - return true if the cmdbatch is waiting |
| * @cmdbatch: Pointer to the command batch object to check |
| * |
| * Return non-zero if the specified command batch is still waiting for sync |
| * point dependencies to be satisfied |
| */ |
| static inline int kgsl_cmdbatch_sync_pending(struct kgsl_cmdbatch *cmdbatch) |
| { |
| int ret; |
| |
| if (cmdbatch == NULL) |
| return 0; |
| |
| spin_lock(&cmdbatch->lock); |
| ret = list_empty(&cmdbatch->synclist) ? 0 : 1; |
| spin_unlock(&cmdbatch->lock); |
| |
| return ret; |
| } |
| |
| #endif /* __KGSL_DEVICE_H */ |