android_kernel_xiaomi_sm8350/drivers/soc/qcom/llcc_perfmon.c
Avinash Philip c3aaf706c3 drver:soc:llcc_perfmon: qdss clk node control
qdss clock needs to be enabled on llcc_perfmon counter to function.
Enable/disable clock on Start/Stop llcc_perfmon commands.

Change-Id: Ic75119a08f0be1deadcdc694f27098296bfda07e
Signed-off-by: Avinash Philip <avinashp@codeaurora.org>
2020-06-14 15:10:21 +05:30

1339 lines
37 KiB
C

// SPDX-License-Identifier: GPL-2.0-only
/*
* Copyright (c) 2019-2020, The Linux Foundation. All rights reserved.
*/
#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
#include <linux/of_device.h>
#include <linux/platform_device.h>
#include <linux/mutex.h>
#include <linux/bitops.h>
#include <linux/io.h>
#include <linux/hrtimer.h>
#include <linux/regmap.h>
#include <linux/soc/qcom/llcc-qcom.h>
#include <linux/module.h>
#include <linux/clk.h>
#include "llcc_events.h"
#include "llcc_perfmon.h"
#define LLCC_PERFMON_NAME "qcom_llcc_perfmon"
#define MAX_CNTR 16
#define MAX_NUMBER_OF_PORTS 8
#define NUM_CHANNELS 16
#define DELIM_CHAR " "
/**
* struct llcc_perfmon_counter_map - llcc perfmon counter map info
* @port_sel: Port selected for configured counter
* @event_sel: Event selected for configured counter
* @counter_dump: Cumulative counter dump
*/
struct llcc_perfmon_counter_map {
unsigned int port_sel;
unsigned int event_sel;
unsigned long long counter_dump[NUM_CHANNELS];
};
struct llcc_perfmon_private;
/**
* struct event_port_ops - event port operation
* @event_config: Counter config support for port & event
* @event_enable: Counter enable support for port
* @event_filter_config: Port filter config support
*/
struct event_port_ops {
void (*event_config)(struct llcc_perfmon_private *priv,
unsigned int type, unsigned int *num, bool enable);
void (*event_enable)(struct llcc_perfmon_private *priv, bool enable);
void (*event_filter_config)(struct llcc_perfmon_private *priv,
enum filter_type filter, unsigned long match,
unsigned long mask, bool enable);
};
/**
* struct llcc_perfmon_private - llcc perfmon private
* @llcc_map: llcc register address space map
* @llcc_bcast_map: llcc broadcast register address space map
* @bank_off: Offset of llcc banks
* @num_banks: Number of banks supported
* @port_ops: struct event_port_ops
* @configured: Mapping of configured event counters
* @configured_cntrs: Count of configured counters.
* @enables_port: Port enabled for perfmon configuration
* @filtered_ports: Port filter enabled
* @port_configd: Number of perfmon port configuration supported
* @mutex: mutex to protect this structure
* @hrtimer: hrtimer instance for timer functionality
* @expires: timer expire time in nano seconds
* @num_mc: number of MCS
* @version: Version information of llcc block
* @clk: clock node to enable qdss
*/
struct llcc_perfmon_private {
struct regmap *llcc_map;
struct regmap *llcc_bcast_map;
unsigned int bank_off[NUM_CHANNELS];
unsigned int num_banks;
struct event_port_ops *port_ops[MAX_NUMBER_OF_PORTS];
struct llcc_perfmon_counter_map configured[MAX_CNTR];
unsigned int configured_cntrs;
unsigned int enables_port;
unsigned int filtered_ports;
unsigned int port_configd;
struct mutex mutex;
struct hrtimer hrtimer;
ktime_t expires;
unsigned int num_mc;
unsigned int version;
struct clk *clock;
};
static inline void llcc_bcast_write(struct llcc_perfmon_private *llcc_priv,
unsigned int offset, uint32_t val)
{
regmap_write(llcc_priv->llcc_bcast_map, offset, val);
}
static inline void llcc_bcast_read(struct llcc_perfmon_private *llcc_priv,
unsigned int offset, uint32_t *val)
{
regmap_read(llcc_priv->llcc_bcast_map, offset, val);
}
static void llcc_bcast_modify(struct llcc_perfmon_private *llcc_priv,
unsigned int offset, uint32_t val, uint32_t mask)
{
uint32_t readval;
llcc_bcast_read(llcc_priv, offset, &readval);
readval &= ~mask;
readval |= val & mask;
llcc_bcast_write(llcc_priv, offset, readval);
}
static void perfmon_counter_dump(struct llcc_perfmon_private *llcc_priv)
{
struct llcc_perfmon_counter_map *counter_map;
uint32_t val;
unsigned int i, j;
if (!llcc_priv->configured_cntrs)
return;
llcc_bcast_write(llcc_priv, PERFMON_DUMP, MONITOR_DUMP);
for (i = 0; i < llcc_priv->configured_cntrs; i++) {
counter_map = &llcc_priv->configured[i];
for (j = 0; j < llcc_priv->num_banks; j++) {
regmap_read(llcc_priv->llcc_map, llcc_priv->bank_off[j]
+ LLCC_COUNTER_n_VALUE(i), &val);
counter_map->counter_dump[j] += val;
}
}
}
static ssize_t perfmon_counter_dump_show(struct device *dev,
struct device_attribute *attr, char *buf)
{
struct llcc_perfmon_private *llcc_priv = dev_get_drvdata(dev);
struct llcc_perfmon_counter_map *counter_map;
unsigned int i, j;
unsigned long long total;
ssize_t cnt = 0;
if (llcc_priv->configured_cntrs == 0) {
pr_err("counters not configured\n");
return cnt;
}
perfmon_counter_dump(llcc_priv);
for (i = 0; i < llcc_priv->configured_cntrs - 1; i++) {
total = 0;
counter_map = &llcc_priv->configured[i];
cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt,
"Port %02d,Event %02d,",
counter_map->port_sel, counter_map->event_sel);
if ((counter_map->port_sel == EVENT_PORT_BEAC) &&
(llcc_priv->num_mc > 1)) {
/* DBX uses 2 counters for BEAC 0 & 1 */
i++;
for (j = 0; j < llcc_priv->num_banks; j++) {
total += counter_map->counter_dump[j];
counter_map->counter_dump[j] = 0;
total += counter_map[1].counter_dump[j];
counter_map[1].counter_dump[j] = 0;
}
} else {
for (j = 0; j < llcc_priv->num_banks; j++) {
total += counter_map->counter_dump[j];
counter_map->counter_dump[j] = 0;
}
}
cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt, "0x%016llx\n",
total);
}
cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt, "CYCLE COUNT, ,");
total = 0;
counter_map = &llcc_priv->configured[i];
for (j = 0; j < llcc_priv->num_banks; j++) {
total += counter_map->counter_dump[j];
counter_map->counter_dump[j] = 0;
}
cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt, "0x%016llx\n", total);
if (llcc_priv->expires)
hrtimer_forward_now(&llcc_priv->hrtimer, llcc_priv->expires);
return cnt;
}
static ssize_t perfmon_configure_store(struct device *dev,
struct device_attribute *attr, const char *buf, size_t count)
{
struct llcc_perfmon_private *llcc_priv = dev_get_drvdata(dev);
struct event_port_ops *port_ops;
struct llcc_perfmon_counter_map *counter_map;
unsigned int j = 0, k, end_cntrs;
unsigned long port_sel, event_sel;
uint32_t val;
char *token, *delim = DELIM_CHAR;
mutex_lock(&llcc_priv->mutex);
if (llcc_priv->configured_cntrs) {
pr_err("Counters configured already, remove & try again\n");
mutex_unlock(&llcc_priv->mutex);
return -EINVAL;
}
token = strsep((char **)&buf, delim);
while (token != NULL) {
if (kstrtoul(token, 0, &port_sel))
break;
if (port_sel >= llcc_priv->port_configd)
break;
token = strsep((char **)&buf, delim);
if (token == NULL)
break;
if (kstrtoul(token, 0, &event_sel))
break;
token = strsep((char **)&buf, delim);
if (event_sel >= EVENT_NUM_MAX) {
pr_err("unsupported event num %ld\n", event_sel);
continue;
}
/* Last perfmon counter for cycle counter */
end_cntrs = 1;
if (port_sel == EVENT_PORT_BEAC)
end_cntrs = llcc_priv->num_mc;
if (j == (MAX_CNTR - end_cntrs))
break;
counter_map = &llcc_priv->configured[j];
counter_map->port_sel = port_sel;
counter_map->event_sel = event_sel;
for (k = 0; k < llcc_priv->num_banks; k++)
counter_map->counter_dump[k] = 0;
port_ops = llcc_priv->port_ops[port_sel];
port_ops->event_config(llcc_priv, event_sel, &j, true);
pr_info("counter %2d configured for event %2ld from port %ld\n",
j++, event_sel, port_sel);
if (((llcc_priv->enables_port & (1 << port_sel)) == 0) &&
(port_ops->event_enable))
port_ops->event_enable(llcc_priv, true);
llcc_priv->enables_port |= (1 << port_sel);
}
/* configure clock event */
val = COUNT_CLOCK_EVENT | CLEAR_ON_ENABLE | CLEAR_ON_DUMP;
llcc_bcast_write(llcc_priv, PERFMON_COUNTER_n_CONFIG(j++), val);
llcc_priv->configured_cntrs = j;
mutex_unlock(&llcc_priv->mutex);
return count;
}
static ssize_t perfmon_remove_store(struct device *dev,
struct device_attribute *attr, const char *buf, size_t count)
{
struct llcc_perfmon_private *llcc_priv = dev_get_drvdata(dev);
struct event_port_ops *port_ops;
struct llcc_perfmon_counter_map *counter_map;
unsigned int j = 0, end_cntrs;
unsigned long port_sel, event_sel;
char *token, *delim = DELIM_CHAR;
mutex_lock(&llcc_priv->mutex);
if (!llcc_priv->configured_cntrs) {
pr_err("Counters not configured\n");
mutex_unlock(&llcc_priv->mutex);
return -EINVAL;
}
token = strsep((char **)&buf, delim);
while (token != NULL) {
if (kstrtoul(token, 0, &port_sel))
break;
if (port_sel >= llcc_priv->port_configd)
break;
token = strsep((char **)&buf, delim);
if (token == NULL)
break;
if (kstrtoul(token, 0, &event_sel))
break;
token = strsep((char **)&buf, delim);
if (event_sel >= EVENT_NUM_MAX) {
pr_err("unsupported event num %ld\n", event_sel);
continue;
}
/* Last perfmon counter for cycle counter */
end_cntrs = 1;
if (port_sel == EVENT_PORT_BEAC)
end_cntrs = llcc_priv->num_mc;
if (j == (llcc_priv->configured_cntrs - end_cntrs))
break;
/* put dummy values */
counter_map = &llcc_priv->configured[j];
counter_map->port_sel = MAX_NUMBER_OF_PORTS;
counter_map->event_sel = 0;
port_ops = llcc_priv->port_ops[port_sel];
port_ops->event_config(llcc_priv, event_sel, &j, false);
pr_info("removed counter %2d for event %2ld from port %2ld\n",
j++, event_sel, port_sel);
if ((llcc_priv->enables_port & (1 << port_sel)) &&
(port_ops->event_enable))
port_ops->event_enable(llcc_priv, false);
llcc_priv->enables_port &= ~(1 << port_sel);
}
/* remove clock event */
llcc_bcast_write(llcc_priv, PERFMON_COUNTER_n_CONFIG(j), 0);
llcc_priv->configured_cntrs = 0;
mutex_unlock(&llcc_priv->mutex);
return count;
}
static enum filter_type find_filter_type(char *filter)
{
enum filter_type ret = UNKNOWN;
if (!strcmp(filter, "SCID"))
ret = SCID;
else if (!strcmp(filter, "MID"))
ret = MID;
else if (!strcmp(filter, "PROFILING_TAG"))
ret = PROFILING_TAG;
else if (!strcmp(filter, "WAY_ID"))
ret = WAY_ID;
else if (!strcmp(filter, "OPCODE"))
ret = OPCODE;
else if (!strcmp(filter, "CACHEALLOC"))
ret = CACHEALLOC;
return ret;
}
static ssize_t perfmon_filter_config_store(struct device *dev,
struct device_attribute *attr, const char *buf,
size_t count)
{
struct llcc_perfmon_private *llcc_priv = dev_get_drvdata(dev);
unsigned long port, mask, match;
struct event_port_ops *port_ops;
char *token, *delim = DELIM_CHAR;
enum filter_type filter = UNKNOWN;
if (llcc_priv->configured_cntrs) {
pr_err("remove configured events and try\n");
return count;
}
mutex_lock(&llcc_priv->mutex);
token = strsep((char **)&buf, delim);
if (token != NULL)
filter = find_filter_type(token);
if (filter == UNKNOWN) {
pr_err("filter configuration failed, Unsupported filter\n");
goto filter_config_free;
}
token = strsep((char **)&buf, delim);
if (token == NULL) {
pr_err("filter configuration failed, Wrong input\n");
goto filter_config_free;
}
if (kstrtoul(token, 0, &match)) {
pr_err("filter configuration failed, Wrong format\n");
goto filter_config_free;
}
if ((filter == SCID) && (match >= SCID_MAX)) {
pr_err("filter configuration failed, SCID above MAX value\n");
goto filter_config_free;
}
token = strsep((char **)&buf, delim);
if (token == NULL) {
pr_err("filter configuration failed, Wrong input\n");
goto filter_config_free;
}
if (kstrtoul(token, 0, &mask)) {
pr_err("filter configuration failed, Wrong format\n");
goto filter_config_free;
}
while (token != NULL) {
token = strsep((char **)&buf, delim);
if (token == NULL)
break;
if (kstrtoul(token, 0, &port))
break;
llcc_priv->filtered_ports |= 1 << port;
port_ops = llcc_priv->port_ops[port];
if (port_ops->event_filter_config)
port_ops->event_filter_config(llcc_priv, filter, match,
mask, true);
}
mutex_unlock(&llcc_priv->mutex);
return count;
filter_config_free:
mutex_unlock(&llcc_priv->mutex);
return -EINVAL;
}
static ssize_t perfmon_filter_remove_store(struct device *dev,
struct device_attribute *attr, const char *buf, size_t count)
{
struct llcc_perfmon_private *llcc_priv = dev_get_drvdata(dev);
struct event_port_ops *port_ops;
unsigned long port, mask, match;
char *token, *delim = DELIM_CHAR;
enum filter_type filter = UNKNOWN;
mutex_lock(&llcc_priv->mutex);
token = strsep((char **)&buf, delim);
if (token != NULL)
filter = find_filter_type(token);
if (filter == UNKNOWN) {
pr_err("filter configuration failed, Unsupported filter\n");
goto filter_remove_free;
}
token = strsep((char **)&buf, delim);
if (token == NULL) {
pr_err("filter configuration failed, Wrong input\n");
goto filter_remove_free;
}
if (kstrtoul(token, 0, &match)) {
pr_err("filter configuration failed, Wrong format\n");
goto filter_remove_free;
}
if ((filter == SCID) && (match >= SCID_MAX)) {
pr_err("filter configuration failed, SCID above MAX value\n");
goto filter_remove_free;
}
token = strsep((char **)&buf, delim);
if (token == NULL) {
pr_err("filter configuration failed, Wrong input\n");
goto filter_remove_free;
}
if (kstrtoul(token, 0, &mask)) {
pr_err("filter configuration failed, Wrong format\n");
goto filter_remove_free;
}
while (token != NULL) {
token = strsep((char **)&buf, delim);
if (token == NULL)
break;
if (kstrtoul(token, 0, &port))
break;
llcc_priv->filtered_ports &= ~(1 << port);
port_ops = llcc_priv->port_ops[port];
if (port_ops->event_filter_config)
port_ops->event_filter_config(llcc_priv, filter, match,
mask, false);
}
filter_remove_free:
mutex_unlock(&llcc_priv->mutex);
return count;
}
static ssize_t perfmon_start_store(struct device *dev,
struct device_attribute *attr, const char *buf,
size_t count)
{
struct llcc_perfmon_private *llcc_priv = dev_get_drvdata(dev);
uint32_t val = 0, mask_val;
unsigned long start;
int ret;
if (kstrtoul(buf, 0, &start))
return -EINVAL;
mutex_lock(&llcc_priv->mutex);
if (start) {
if (!llcc_priv->configured_cntrs) {
pr_err("start failed. perfmon not configured\n");
mutex_unlock(&llcc_priv->mutex);
return -EINVAL;
}
ret = clk_prepare_enable(llcc_priv->clock);
if (ret) {
mutex_unlock(&llcc_priv->mutex);
return -EINVAL;
}
val = MANUAL_MODE | MONITOR_EN;
if (llcc_priv->expires) {
if (hrtimer_is_queued(&llcc_priv->hrtimer))
hrtimer_forward_now(&llcc_priv->hrtimer,
llcc_priv->expires);
else
hrtimer_start(&llcc_priv->hrtimer,
llcc_priv->expires,
HRTIMER_MODE_REL_PINNED);
}
} else {
if (llcc_priv->expires)
hrtimer_cancel(&llcc_priv->hrtimer);
if (!llcc_priv->configured_cntrs)
pr_err("stop failed. perfmon not configured\n");
}
mask_val = PERFMON_MODE_MONITOR_MODE_MASK |
PERFMON_MODE_MONITOR_EN_MASK;
llcc_bcast_modify(llcc_priv, PERFMON_MODE, val, mask_val);
if (!start)
clk_disable_unprepare(llcc_priv->clock);
mutex_unlock(&llcc_priv->mutex);
return count;
}
static ssize_t perfmon_ns_periodic_dump_store(struct device *dev,
struct device_attribute *attr, const char *buf,
size_t count)
{
struct llcc_perfmon_private *llcc_priv = dev_get_drvdata(dev);
if (kstrtos64(buf, 0, &llcc_priv->expires))
return -EINVAL;
mutex_lock(&llcc_priv->mutex);
if (!llcc_priv->expires) {
hrtimer_cancel(&llcc_priv->hrtimer);
mutex_unlock(&llcc_priv->mutex);
return count;
}
if (hrtimer_is_queued(&llcc_priv->hrtimer))
hrtimer_forward_now(&llcc_priv->hrtimer, llcc_priv->expires);
else
hrtimer_start(&llcc_priv->hrtimer, llcc_priv->expires,
HRTIMER_MODE_REL_PINNED);
mutex_unlock(&llcc_priv->mutex);
return count;
}
static ssize_t perfmon_scid_status_show(struct device *dev,
struct device_attribute *attr, char *buf)
{
struct llcc_perfmon_private *llcc_priv = dev_get_drvdata(dev);
uint32_t val;
unsigned int i, j, offset;
ssize_t cnt = 0;
unsigned long total;
for (i = 0; i < SCID_MAX; i++) {
total = 0;
offset = TRP_SCID_n_STATUS(i);
for (j = 0; j < llcc_priv->num_banks; j++) {
regmap_read(llcc_priv->llcc_map,
llcc_priv->bank_off[j] + offset, &val);
val = (val & TRP_SCID_STATUS_CURRENT_CAP_MASK) >>
TRP_SCID_STATUS_CURRENT_CAP_SHIFT;
total += val;
}
llcc_bcast_read(llcc_priv, offset, &val);
if (val & TRP_SCID_STATUS_ACTIVE_MASK)
cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt,
"SCID %02d %10s", i, "ACTIVE");
else
cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt,
"SCID %02d %10s", i, "DEACTIVE");
cnt += scnprintf(buf + cnt, PAGE_SIZE - cnt, ",0x%08lx\n",
total);
}
return cnt;
}
static DEVICE_ATTR_RO(perfmon_counter_dump);
static DEVICE_ATTR_WO(perfmon_configure);
static DEVICE_ATTR_WO(perfmon_remove);
static DEVICE_ATTR_WO(perfmon_filter_config);
static DEVICE_ATTR_WO(perfmon_filter_remove);
static DEVICE_ATTR_WO(perfmon_start);
static DEVICE_ATTR_RO(perfmon_scid_status);
static DEVICE_ATTR_WO(perfmon_ns_periodic_dump);
static struct attribute *llcc_perfmon_attrs[] = {
&dev_attr_perfmon_counter_dump.attr,
&dev_attr_perfmon_configure.attr,
&dev_attr_perfmon_remove.attr,
&dev_attr_perfmon_filter_config.attr,
&dev_attr_perfmon_filter_remove.attr,
&dev_attr_perfmon_start.attr,
&dev_attr_perfmon_scid_status.attr,
&dev_attr_perfmon_ns_periodic_dump.attr,
NULL,
};
static struct attribute_group llcc_perfmon_group = {
.attrs = llcc_perfmon_attrs,
};
static void perfmon_cntr_config(struct llcc_perfmon_private *llcc_priv,
unsigned int port, unsigned int counter_num, bool enable)
{
uint32_t val = 0;
if (counter_num >= MAX_CNTR)
return;
if (enable)
val = (port & PERFMON_PORT_SELECT_MASK) |
((counter_num << EVENT_SELECT_SHIFT) &
PERFMON_EVENT_SELECT_MASK) | CLEAR_ON_ENABLE |
CLEAR_ON_DUMP;
llcc_bcast_write(llcc_priv, PERFMON_COUNTER_n_CONFIG(counter_num), val);
}
static void feac_event_config(struct llcc_perfmon_private *llcc_priv,
unsigned int event_type, unsigned int *counter_num,
bool enable)
{
uint32_t val = 0, mask_val;
mask_val = EVENT_SEL_MASK;
if (llcc_priv->version == REV_2)
mask_val = EVENT_SEL_MASK7;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_FEAC))
mask_val |= FILTER_SEL_MASK | FILTER_EN_MASK;
if (enable) {
val = (event_type << EVENT_SEL_SHIFT) & EVENT_SEL_MASK;
if (llcc_priv->version == REV_2)
val = (event_type << EVENT_SEL_SHIFT) &
EVENT_SEL_MASK7;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_FEAC))
val |= (FILTER_0 << FILTER_SEL_SHIFT) | FILTER_EN;
}
llcc_bcast_modify(llcc_priv, FEAC_PROF_EVENT_n_CFG(*counter_num),
val, mask_val);
perfmon_cntr_config(llcc_priv, EVENT_PORT_FEAC, *counter_num, enable);
}
static void feac_event_enable(struct llcc_perfmon_private *llcc_priv,
bool enable)
{
uint32_t val = 0, mask_val;
if (enable) {
val = (BYTE_SCALING << BYTE_SCALING_SHIFT) |
(BEAT_SCALING << BEAT_SCALING_SHIFT) | PROF_EN;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_FEAC)) {
if (llcc_priv->version == REV_0)
val |= (FILTER_0 <<
FEAC_SCALING_FILTER_SEL_SHIFT) |
FEAC_SCALING_FILTER_EN;
else
val |= (FILTER_0 <<
FEAC_WR_BEAT_FILTER_SEL_SHIFT) |
FEAC_WR_BEAT_FILTER_EN |
(FILTER_0 <<
FEAC_WR_BYTE_FILTER_SEL_SHIFT) |
FEAC_WR_BYTE_FILTER_EN |
(FILTER_0 <<
FEAC_RD_BEAT_FILTER_SEL_SHIFT) |
FEAC_RD_BEAT_FILTER_EN |
(FILTER_0 <<
FEAC_RD_BYTE_FILTER_SEL_SHIFT) |
FEAC_RD_BYTE_FILTER_EN;
}
}
mask_val = PROF_CFG_BEAT_SCALING_MASK | PROF_CFG_BYTE_SCALING_MASK
| PROF_CFG_EN_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_FEAC)) {
if (llcc_priv->version == REV_0)
mask_val |= FEAC_SCALING_FILTER_SEL_MASK |
FEAC_SCALING_FILTER_EN_MASK;
else
mask_val |= FEAC_WR_BEAT_FILTER_SEL_MASK |
FEAC_WR_BEAT_FILTER_EN_MASK |
FEAC_WR_BYTE_FILTER_SEL_MASK |
FEAC_WR_BYTE_FILTER_EN_MASK |
FEAC_RD_BEAT_FILTER_SEL_MASK |
FEAC_RD_BEAT_FILTER_EN_MASK |
FEAC_RD_BYTE_FILTER_SEL_MASK |
FEAC_RD_BYTE_FILTER_EN_MASK;
}
llcc_bcast_modify(llcc_priv, FEAC_PROF_CFG, val, mask_val);
}
static void feac_event_filter_config(struct llcc_perfmon_private *llcc_priv,
enum filter_type filter, unsigned long match,
unsigned long mask, bool enable)
{
uint32_t val = 0, mask_val;
if (filter == SCID) {
if (llcc_priv->version == REV_0) {
if (enable)
val = (match << SCID_MATCH_SHIFT) |
(mask << SCID_MASK_SHIFT);
mask_val = SCID_MATCH_MASK | SCID_MASK_MASK;
} else {
if (enable)
val = (1 << match);
mask_val = SCID_MULTI_MATCH_MASK;
}
llcc_bcast_modify(llcc_priv, FEAC_PROF_FILTER_0_CFG6, val,
mask_val);
} else if (filter == MID) {
if (enable)
val = (match << MID_MATCH_SHIFT) |
(mask << MID_MASK_SHIFT);
mask_val = MID_MATCH_MASK | MID_MASK_MASK;
llcc_bcast_modify(llcc_priv, FEAC_PROF_FILTER_0_CFG5, val,
mask_val);
} else if (filter == OPCODE) {
if (enable)
val = (match << OPCODE_MATCH_SHIFT) |
(mask << OPCODE_MASK_SHIFT);
mask_val = OPCODE_MATCH_MASK | OPCODE_MASK_MASK;
llcc_bcast_modify(llcc_priv, FEAC_PROF_FILTER_0_CFG3, val,
mask_val);
} else if (filter == CACHEALLOC) {
if (enable)
val = (match << CACHEALLOC_MATCH_SHIFT) |
(mask << CACHEALLOC_MASK_SHIFT);
mask_val = CACHEALLOC_MATCH_MASK | CACHEALLOC_MASK_MASK;
llcc_bcast_modify(llcc_priv, FEAC_PROF_FILTER_0_CFG3, val,
mask_val);
} else {
pr_err("unknown filter/not supported\n");
}
}
static struct event_port_ops feac_port_ops = {
.event_config = feac_event_config,
.event_enable = feac_event_enable,
.event_filter_config = feac_event_filter_config,
};
static void ferc_event_config(struct llcc_perfmon_private *llcc_priv,
unsigned int event_type, unsigned int *counter_num,
bool enable)
{
uint32_t val = 0, mask_val;
mask_val = EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_FERC))
mask_val |= FILTER_SEL_MASK | FILTER_EN_MASK;
if (enable) {
val = event_type << EVENT_SEL_SHIFT;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_FERC))
val |= (FILTER_0 << FILTER_SEL_SHIFT) | FILTER_EN;
}
llcc_bcast_modify(llcc_priv, FERC_PROF_EVENT_n_CFG(*counter_num),
val, mask_val);
perfmon_cntr_config(llcc_priv, EVENT_PORT_FERC, *counter_num, enable);
}
static void ferc_event_enable(struct llcc_perfmon_private *llcc_priv,
bool enable)
{
uint32_t val = 0, mask_val;
if (enable)
val = (BYTE_SCALING << BYTE_SCALING_SHIFT) |
(BEAT_SCALING << BEAT_SCALING_SHIFT) | PROF_EN;
mask_val = PROF_CFG_BEAT_SCALING_MASK | PROF_CFG_BYTE_SCALING_MASK |
PROF_CFG_EN_MASK;
llcc_bcast_modify(llcc_priv, FERC_PROF_CFG, val, mask_val);
}
static void ferc_event_filter_config(struct llcc_perfmon_private *llcc_priv,
enum filter_type filter, unsigned long match,
unsigned long mask, bool enable)
{
uint32_t val = 0, mask_val;
if (filter != PROFILING_TAG) {
pr_err("unknown filter/not supported\n");
return;
}
if (enable)
val = (match << PROFTAG_MATCH_SHIFT) |
(mask << PROFTAG_MASK_SHIFT);
mask_val = PROFTAG_MATCH_MASK | PROFTAG_MASK_MASK;
llcc_bcast_modify(llcc_priv, FERC_PROF_FILTER_0_CFG0, val, mask_val);
}
static struct event_port_ops ferc_port_ops = {
.event_config = ferc_event_config,
.event_enable = ferc_event_enable,
.event_filter_config = ferc_event_filter_config,
};
static void fewc_event_config(struct llcc_perfmon_private *llcc_priv,
unsigned int event_type, unsigned int *counter_num,
bool enable)
{
uint32_t val = 0, mask_val;
mask_val = EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_FEWC))
mask_val |= FILTER_SEL_MASK | FILTER_EN_MASK;
if (enable) {
val = (event_type << EVENT_SEL_SHIFT) & EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_FEWC))
val |= (FILTER_0 << FILTER_SEL_SHIFT) | FILTER_EN;
}
llcc_bcast_modify(llcc_priv, FEWC_PROF_EVENT_n_CFG(*counter_num),
val, mask_val);
perfmon_cntr_config(llcc_priv, EVENT_PORT_FEWC, *counter_num, enable);
}
static void fewc_event_filter_config(struct llcc_perfmon_private *llcc_priv,
enum filter_type filter, unsigned long match,
unsigned long mask, bool enable)
{
uint32_t val = 0, mask_val;
if (filter != PROFILING_TAG) {
pr_err("unknown filter/not supported\n");
return;
}
if (enable)
val = (match << PROFTAG_MATCH_SHIFT) |
(mask << PROFTAG_MASK_SHIFT);
mask_val = PROFTAG_MATCH_MASK | PROFTAG_MASK_MASK;
llcc_bcast_modify(llcc_priv, FEWC_PROF_FILTER_0_CFG0, val, mask_val);
}
static struct event_port_ops fewc_port_ops = {
.event_config = fewc_event_config,
.event_filter_config = fewc_event_filter_config,
};
static void beac_event_config(struct llcc_perfmon_private *llcc_priv,
unsigned int event_type, unsigned int *counter_num,
bool enable)
{
uint32_t val = 0, mask_val;
uint32_t valcfg = 0, mask_valcfg = 0;
unsigned int mc_cnt, offset;
struct llcc_perfmon_counter_map *counter_map;
mask_val = EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_BEAC)) {
mask_val |= FILTER_SEL_MASK | FILTER_EN_MASK;
if (llcc_priv->version == REV_2)
mask_valcfg = BEAC_WR_BEAT_FILTER_SEL_MASK |
BEAC_WR_BEAT_FILTER_EN_MASK |
BEAC_RD_BEAT_FILTER_SEL_MASK |
BEAC_RD_BEAT_FILTER_EN_MASK;
}
if (enable) {
val = (event_type << EVENT_SEL_SHIFT) & EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_BEAC)) {
val |= (FILTER_0 << FILTER_SEL_SHIFT) | FILTER_EN;
if (llcc_priv->version == REV_2)
valcfg = (FILTER_0 <<
BEAC_WR_BEAT_FILTER_SEL_SHIFT) |
BEAC_WR_BEAT_FILTER_EN |
(FILTER_0 <<
BEAC_RD_BEAT_FILTER_SEL_SHIFT) |
BEAC_RD_BEAT_FILTER_EN;
}
}
for (mc_cnt = 0; mc_cnt < llcc_priv->num_mc; mc_cnt++) {
offset = BEAC_PROF_EVENT_n_CFG(*counter_num + mc_cnt) +
mc_cnt * BEAC_INST_OFF;
llcc_bcast_modify(llcc_priv, offset, val, mask_val);
offset = BEAC_PROF_CFG + mc_cnt * BEAC_INST_OFF;
llcc_bcast_modify(llcc_priv, offset, valcfg, mask_valcfg);
perfmon_cntr_config(llcc_priv, EVENT_PORT_BEAC, *counter_num,
enable);
/* DBX uses 2 counters for BEAC 0 & 1 */
if (mc_cnt == 1)
perfmon_cntr_config(llcc_priv, EVENT_PORT_BEAC1,
*counter_num + mc_cnt, enable);
}
/* DBX uses 2 counters for BEAC 0 & 1 */
if (llcc_priv->num_mc > 1) {
counter_map = &llcc_priv->configured[(*counter_num)++];
if (enable) {
counter_map->port_sel = EVENT_PORT_BEAC;
counter_map->event_sel = event_type;
} else {
counter_map->port_sel = MAX_NUMBER_OF_PORTS;
counter_map->event_sel = 100;
}
}
}
static void beac_event_enable(struct llcc_perfmon_private *llcc_priv,
bool enable)
{
uint32_t val = 0, mask_val;
unsigned int mc_cnt, offset;
if (enable)
val = (BYTE_SCALING << BYTE_SCALING_SHIFT) |
(BEAT_SCALING << BEAT_SCALING_SHIFT) | PROF_EN;
mask_val = PROF_CFG_BEAT_SCALING_MASK | PROF_CFG_BYTE_SCALING_MASK
| PROF_CFG_EN_MASK;
for (mc_cnt = 0; mc_cnt < llcc_priv->num_mc; mc_cnt++) {
offset = BEAC_PROF_CFG + mc_cnt * BEAC_INST_OFF;
llcc_bcast_modify(llcc_priv, offset, val, mask_val);
}
}
static void beac_event_filter_config(struct llcc_perfmon_private *llcc_priv,
enum filter_type filter, unsigned long match,
unsigned long mask, bool enable)
{
uint32_t val = 0, mask_val;
unsigned int mc_cnt, offset;
if (filter != PROFILING_TAG) {
pr_err("unknown filter/not supported\n");
return;
}
if (enable)
val = (match << BEAC_PROFTAG_MATCH_SHIFT) |
(mask << BEAC_PROFTAG_MASK_SHIFT);
mask_val = BEAC_PROFTAG_MASK_MASK | BEAC_PROFTAG_MATCH_MASK;
for (mc_cnt = 0; mc_cnt < llcc_priv->num_mc; mc_cnt++) {
offset = BEAC_PROF_FILTER_0_CFG5 + mc_cnt * BEAC_INST_OFF;
llcc_bcast_modify(llcc_priv, offset, val, mask_val);
}
if (enable)
val = match << BEAC_MC_PROFTAG_SHIFT;
mask_val = BEAC_MC_PROFTAG_MASK;
for (mc_cnt = 0; mc_cnt < llcc_priv->num_mc; mc_cnt++) {
offset = BEAC_PROF_CFG + mc_cnt * BEAC_INST_OFF;
llcc_bcast_modify(llcc_priv, offset, val, mask_val);
}
}
static struct event_port_ops beac_port_ops = {
.event_config = beac_event_config,
.event_enable = beac_event_enable,
.event_filter_config = beac_event_filter_config,
};
static void berc_event_config(struct llcc_perfmon_private *llcc_priv,
unsigned int event_type, unsigned int *counter_num,
bool enable)
{
uint32_t val = 0, mask_val;
mask_val = EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_BERC))
mask_val |= FILTER_SEL_MASK | FILTER_EN_MASK;
if (enable) {
val = (event_type << EVENT_SEL_SHIFT) & EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_BERC))
val |= (FILTER_0 << FILTER_SEL_SHIFT) | FILTER_EN;
}
llcc_bcast_modify(llcc_priv, BERC_PROF_EVENT_n_CFG(*counter_num),
val, mask_val);
perfmon_cntr_config(llcc_priv, EVENT_PORT_BERC, *counter_num, enable);
}
static void berc_event_enable(struct llcc_perfmon_private *llcc_priv,
bool enable)
{
uint32_t val = 0, mask_val;
if (enable)
val = (BYTE_SCALING << BYTE_SCALING_SHIFT) |
(BEAT_SCALING << BEAT_SCALING_SHIFT) | PROF_EN;
mask_val = PROF_CFG_BEAT_SCALING_MASK | PROF_CFG_BYTE_SCALING_MASK
| PROF_CFG_EN_MASK;
llcc_bcast_modify(llcc_priv, BERC_PROF_CFG, val, mask_val);
}
static void berc_event_filter_config(struct llcc_perfmon_private *llcc_priv,
enum filter_type filter, unsigned long match,
unsigned long mask, bool enable)
{
uint32_t val = 0, mask_val;
if (filter != PROFILING_TAG) {
pr_err("unknown filter/not supported\n");
return;
}
if (enable)
val = (match << PROFTAG_MATCH_SHIFT) |
(mask << PROFTAG_MASK_SHIFT);
mask_val = PROFTAG_MATCH_MASK | PROFTAG_MASK_MASK;
llcc_bcast_modify(llcc_priv, BERC_PROF_FILTER_0_CFG0, val, mask_val);
}
static struct event_port_ops berc_port_ops = {
.event_config = berc_event_config,
.event_enable = berc_event_enable,
.event_filter_config = berc_event_filter_config,
};
static void trp_event_config(struct llcc_perfmon_private *llcc_priv,
unsigned int event_type, unsigned int *counter_num,
bool enable)
{
uint32_t val = 0, mask_val;
mask_val = EVENT_SEL_MASK;
if (llcc_priv->version == REV_2)
mask_val = EVENT_SEL_MASK7;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_TRP))
mask_val |= FILTER_SEL_MASK | FILTER_EN_MASK;
if (enable) {
val = (event_type << EVENT_SEL_SHIFT) & EVENT_SEL_MASK;
if (llcc_priv->version == REV_2)
val = (event_type << EVENT_SEL_SHIFT) &
EVENT_SEL_MASK7;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_TRP))
val |= (FILTER_0 << FILTER_SEL_SHIFT) | FILTER_EN;
}
llcc_bcast_modify(llcc_priv, TRP_PROF_EVENT_n_CFG(*counter_num),
val, mask_val);
perfmon_cntr_config(llcc_priv, EVENT_PORT_TRP, *counter_num, enable);
}
static void trp_event_filter_config(struct llcc_perfmon_private *llcc_priv,
enum filter_type filter, unsigned long match,
unsigned long mask, bool enable)
{
uint32_t val = 0, mask_val;
if (filter == SCID) {
if (llcc_priv->version == REV_2) {
if (enable)
val = (1 << match);
mask_val = SCID_MULTI_MATCH_MASK;
} else {
if (enable)
val = (match << TRP_SCID_MATCH_SHIFT) |
(mask << TRP_SCID_MASK_SHIFT);
mask_val = TRP_SCID_MATCH_MASK | TRP_SCID_MASK_MASK;
}
} else if (filter == WAY_ID) {
if (enable)
val = (match << TRP_WAY_ID_MATCH_SHIFT) |
(mask << TRP_WAY_ID_MASK_SHIFT);
mask_val = TRP_WAY_ID_MATCH_MASK | TRP_WAY_ID_MASK_MASK;
} else if (filter == PROFILING_TAG) {
if (enable)
val = (match << TRP_PROFTAG_MATCH_SHIFT) |
(mask << TRP_PROFTAG_MASK_SHIFT);
mask_val = TRP_PROFTAG_MATCH_MASK | TRP_PROFTAG_MASK_MASK;
} else {
pr_err("unknown filter/not supported\n");
return;
}
if ((llcc_priv->version == REV_2) && (filter == SCID))
llcc_bcast_modify(llcc_priv, TRP_PROF_FILTER_0_CFG2, val,
mask_val);
else
llcc_bcast_modify(llcc_priv, TRP_PROF_FILTER_0_CFG1, val,
mask_val);
}
static struct event_port_ops trp_port_ops = {
.event_config = trp_event_config,
.event_filter_config = trp_event_filter_config,
};
static void drp_event_config(struct llcc_perfmon_private *llcc_priv,
unsigned int event_type, unsigned int *counter_num,
bool enable)
{
uint32_t val = 0, mask_val;
mask_val = EVENT_SEL_MASK;
if (llcc_priv->version == REV_2)
mask_val = EVENT_SEL_MASK7;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_DRP))
mask_val |= FILTER_SEL_MASK | FILTER_EN_MASK;
if (enable) {
val = (event_type << EVENT_SEL_SHIFT) & EVENT_SEL_MASK;
if (llcc_priv->version == REV_2)
val = (event_type << EVENT_SEL_SHIFT) &
EVENT_SEL_MASK7;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_DRP))
val |= (FILTER_0 << FILTER_SEL_SHIFT) | FILTER_EN;
}
llcc_bcast_modify(llcc_priv, DRP_PROF_EVENT_n_CFG(*counter_num),
val, mask_val);
perfmon_cntr_config(llcc_priv, EVENT_PORT_DRP, *counter_num, enable);
}
static void drp_event_enable(struct llcc_perfmon_private *llcc_priv,
bool enable)
{
uint32_t val = 0, mask_val;
if (enable)
val = (BEAT_SCALING << BEAT_SCALING_SHIFT) | PROF_EN;
mask_val = PROF_CFG_BEAT_SCALING_MASK | PROF_CFG_EN_MASK;
llcc_bcast_modify(llcc_priv, DRP_PROF_CFG, val, mask_val);
}
static struct event_port_ops drp_port_ops = {
.event_config = drp_event_config,
.event_enable = drp_event_enable,
};
static void pmgr_event_config(struct llcc_perfmon_private *llcc_priv,
unsigned int event_type, unsigned int *counter_num,
bool enable)
{
uint32_t val = 0, mask_val;
mask_val = EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_PMGR))
mask_val |= FILTER_SEL_MASK | FILTER_EN_MASK;
if (enable) {
val = (event_type << EVENT_SEL_SHIFT) & EVENT_SEL_MASK;
if (llcc_priv->filtered_ports & (1 << EVENT_PORT_PMGR))
val |= (FILTER_0 << FILTER_SEL_SHIFT) | FILTER_EN;
}
llcc_bcast_modify(llcc_priv, PMGR_PROF_EVENT_n_CFG(*counter_num),
val, mask_val);
perfmon_cntr_config(llcc_priv, EVENT_PORT_PMGR, *counter_num, enable);
}
static struct event_port_ops pmgr_port_ops = {
.event_config = pmgr_event_config,
};
static void llcc_register_event_port(struct llcc_perfmon_private *llcc_priv,
struct event_port_ops *ops, unsigned int event_port_num)
{
if (llcc_priv->port_configd >= MAX_NUMBER_OF_PORTS) {
pr_err("Register port Failure!\n");
return;
}
llcc_priv->port_configd = llcc_priv->port_configd + 1;
llcc_priv->port_ops[event_port_num] = ops;
}
static enum hrtimer_restart llcc_perfmon_timer_handler(struct hrtimer *hrtimer)
{
struct llcc_perfmon_private *llcc_priv = container_of(hrtimer,
struct llcc_perfmon_private, hrtimer);
perfmon_counter_dump(llcc_priv);
hrtimer_forward_now(&llcc_priv->hrtimer, llcc_priv->expires);
return HRTIMER_RESTART;
}
static int llcc_perfmon_probe(struct platform_device *pdev)
{
int result = 0;
struct llcc_perfmon_private *llcc_priv;
struct llcc_drv_data *llcc_driv_data = pdev->dev.platform_data;
uint32_t val;
llcc_priv = devm_kzalloc(&pdev->dev, sizeof(*llcc_priv), GFP_KERNEL);
if (llcc_priv == NULL)
return -ENOMEM;
if (!llcc_driv_data)
return -ENOMEM;
if ((llcc_driv_data->regmap == NULL) ||
(llcc_driv_data->bcast_regmap == NULL))
return -ENODEV;
llcc_priv->llcc_map = llcc_driv_data->regmap;
llcc_priv->llcc_bcast_map = llcc_driv_data->bcast_regmap;
llcc_bcast_read(llcc_priv, LLCC_COMMON_STATUS0, &val);
llcc_priv->num_mc = (val & NUM_MC_MASK) >> NUM_MC_SHIFT;
/* Setting to 1, as some platforms it read as 0 */
if (llcc_priv->num_mc == 0)
llcc_priv->num_mc = 1;
llcc_priv->num_banks = (val & LB_CNT_MASK) >> LB_CNT_SHIFT;
for (val = 0; val < llcc_priv->num_banks; val++)
llcc_priv->bank_off[val] = BANK_OFFSET * val;
llcc_priv->version = REV_0;
llcc_bcast_read(llcc_priv, LLCC_COMMON_HW_INFO, &val);
if (val == LLCC_VERSION_1)
llcc_priv->version = REV_1;
else if (val == LLCC_VERSION_2)
llcc_priv->version = REV_2;
llcc_priv->clock = devm_clk_get(pdev->dev.parent, "qdss_clk");
if (IS_ERR_OR_NULL(llcc_priv->clock)) {
pr_err("failed to get clock node\n");
return PTR_ERR(llcc_priv->clock);
}
result = sysfs_create_group(&pdev->dev.kobj, &llcc_perfmon_group);
if (result) {
pr_err("Unable to create sysfs group\n");
return result;
}
mutex_init(&llcc_priv->mutex);
platform_set_drvdata(pdev, llcc_priv);
llcc_register_event_port(llcc_priv, &feac_port_ops, EVENT_PORT_FEAC);
llcc_register_event_port(llcc_priv, &ferc_port_ops, EVENT_PORT_FERC);
llcc_register_event_port(llcc_priv, &fewc_port_ops, EVENT_PORT_FEWC);
llcc_register_event_port(llcc_priv, &beac_port_ops, EVENT_PORT_BEAC);
llcc_register_event_port(llcc_priv, &berc_port_ops, EVENT_PORT_BERC);
llcc_register_event_port(llcc_priv, &trp_port_ops, EVENT_PORT_TRP);
llcc_register_event_port(llcc_priv, &drp_port_ops, EVENT_PORT_DRP);
llcc_register_event_port(llcc_priv, &pmgr_port_ops, EVENT_PORT_PMGR);
hrtimer_init(&llcc_priv->hrtimer, CLOCK_MONOTONIC, HRTIMER_MODE_REL);
llcc_priv->hrtimer.function = llcc_perfmon_timer_handler;
llcc_priv->expires = 0;
pr_info("Revision %d, has %d memory controllers connected with LLCC\n",
llcc_priv->version, llcc_priv->num_mc);
return 0;
}
static int llcc_perfmon_remove(struct platform_device *pdev)
{
struct llcc_perfmon_private *llcc_priv = platform_get_drvdata(pdev);
while (hrtimer_active(&llcc_priv->hrtimer))
hrtimer_cancel(&llcc_priv->hrtimer);
mutex_destroy(&llcc_priv->mutex);
sysfs_remove_group(&pdev->dev.kobj, &llcc_perfmon_group);
platform_set_drvdata(pdev, NULL);
return 0;
}
static struct platform_driver llcc_perfmon_driver = {
.probe = llcc_perfmon_probe,
.remove = llcc_perfmon_remove,
.driver = {
.name = LLCC_PERFMON_NAME,
}
};
module_platform_driver(llcc_perfmon_driver);
MODULE_DESCRIPTION("QCOM LLCC PMU MONITOR");
MODULE_LICENSE("GPL v2");