mirror of https://gitee.com/openkylin/linux.git
- Multiple panfrost fixes for regulator support and page fault handling
- Some cleanups and fixes in the self-refresh helpers - Some cleanups and fixes in the atomic helpers -----BEGIN PGP SIGNATURE----- iHUEABYIAB0WIQRcEzekXsqa64kGDp7j7w1vZxhRxQUCXYjuIAAKCRDj7w1vZxhR xfNKAP9dlMLleLnjvAnCpI/Q2YZAktGCC9mIvqSzXJz52mwXNQEA5avYLerwa2+J P5xmO7mr2/9gzeeeWE1xaufQwZnYrwY= =Z99W -----END PGP SIGNATURE----- Merge tag 'drm-misc-next-fixes-2019-09-23' of git://anongit.freedesktop.org/drm/drm-misc into drm-next - Multiple panfrost fixes for regulator support and page fault handling - Some cleanups and fixes in the self-refresh helpers - Some cleanups and fixes in the atomic helpers Signed-off-by: Dave Airlie <airlied@redhat.com> From: Maxime Ripard <mripard@kernel.org> Link: https://patchwork.freedesktop.org/patch/msgid/20190923160946.nvaqiw5j7fpcdhc7@gilmour
This commit is contained in:
commit
da3fce4af7
|
@ -5770,8 +5770,7 @@ static void amdgpu_dm_commit_planes(struct drm_atomic_state *state,
|
||||||
* change FB pitch, DCC state, rotation or mirroing.
|
* change FB pitch, DCC state, rotation or mirroing.
|
||||||
*/
|
*/
|
||||||
bundle->flip_addrs[planes_count].flip_immediate =
|
bundle->flip_addrs[planes_count].flip_immediate =
|
||||||
(crtc->state->pageflip_flags &
|
crtc->state->async_flip &&
|
||||||
DRM_MODE_PAGE_FLIP_ASYNC) != 0 &&
|
|
||||||
acrtc_state->update_type == UPDATE_TYPE_FAST;
|
acrtc_state->update_type == UPDATE_TYPE_FAST;
|
||||||
|
|
||||||
timestamp_ns = ktime_get_ns();
|
timestamp_ns = ktime_get_ns();
|
||||||
|
@ -6348,7 +6347,7 @@ static void amdgpu_dm_atomic_commit_tail(struct drm_atomic_state *state)
|
||||||
amdgpu_dm_enable_crtc_interrupts(dev, state, true);
|
amdgpu_dm_enable_crtc_interrupts(dev, state, true);
|
||||||
|
|
||||||
for_each_new_crtc_in_state(state, crtc, new_crtc_state, j)
|
for_each_new_crtc_in_state(state, crtc, new_crtc_state, j)
|
||||||
if (new_crtc_state->pageflip_flags & DRM_MODE_PAGE_FLIP_ASYNC)
|
if (new_crtc_state->async_flip)
|
||||||
wait_for_vblank = false;
|
wait_for_vblank = false;
|
||||||
|
|
||||||
/* update planes when needed per crtc*/
|
/* update planes when needed per crtc*/
|
||||||
|
|
|
@ -874,6 +874,9 @@ static int adv7511_bridge_attach(struct drm_bridge *bridge)
|
||||||
&adv7511_connector_helper_funcs);
|
&adv7511_connector_helper_funcs);
|
||||||
drm_connector_attach_encoder(&adv->connector, bridge->encoder);
|
drm_connector_attach_encoder(&adv->connector, bridge->encoder);
|
||||||
|
|
||||||
|
if (adv->type == ADV7533)
|
||||||
|
ret = adv7533_attach_dsi(adv);
|
||||||
|
|
||||||
if (adv->i2c_main->irq)
|
if (adv->i2c_main->irq)
|
||||||
regmap_write(adv->regmap, ADV7511_REG_INT_ENABLE(0),
|
regmap_write(adv->regmap, ADV7511_REG_INT_ENABLE(0),
|
||||||
ADV7511_INT0_HPD);
|
ADV7511_INT0_HPD);
|
||||||
|
@ -1219,17 +1222,8 @@ static int adv7511_probe(struct i2c_client *i2c, const struct i2c_device_id *id)
|
||||||
drm_bridge_add(&adv7511->bridge);
|
drm_bridge_add(&adv7511->bridge);
|
||||||
|
|
||||||
adv7511_audio_init(dev, adv7511);
|
adv7511_audio_init(dev, adv7511);
|
||||||
|
|
||||||
if (adv7511->type == ADV7533) {
|
|
||||||
ret = adv7533_attach_dsi(adv7511);
|
|
||||||
if (ret)
|
|
||||||
goto err_remove_bridge;
|
|
||||||
}
|
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
err_remove_bridge:
|
|
||||||
drm_bridge_remove(&adv7511->bridge);
|
|
||||||
err_unregister_cec:
|
err_unregister_cec:
|
||||||
i2c_unregister_device(adv7511->i2c_cec);
|
i2c_unregister_device(adv7511->i2c_cec);
|
||||||
if (adv7511->cec_clk)
|
if (adv7511->cec_clk)
|
||||||
|
|
|
@ -26,6 +26,7 @@
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#include <linux/dma-fence.h>
|
#include <linux/dma-fence.h>
|
||||||
|
#include <linux/ktime.h>
|
||||||
|
|
||||||
#include <drm/drm_atomic.h>
|
#include <drm/drm_atomic.h>
|
||||||
#include <drm/drm_atomic_helper.h>
|
#include <drm/drm_atomic_helper.h>
|
||||||
|
@ -1580,9 +1581,23 @@ static void commit_tail(struct drm_atomic_state *old_state)
|
||||||
{
|
{
|
||||||
struct drm_device *dev = old_state->dev;
|
struct drm_device *dev = old_state->dev;
|
||||||
const struct drm_mode_config_helper_funcs *funcs;
|
const struct drm_mode_config_helper_funcs *funcs;
|
||||||
|
ktime_t start;
|
||||||
|
s64 commit_time_ms;
|
||||||
|
|
||||||
funcs = dev->mode_config.helper_private;
|
funcs = dev->mode_config.helper_private;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We're measuring the _entire_ commit, so the time will vary depending
|
||||||
|
* on how many fences and objects are involved. For the purposes of self
|
||||||
|
* refresh, this is desirable since it'll give us an idea of how
|
||||||
|
* congested things are. This will inform our decision on how often we
|
||||||
|
* should enter self refresh after idle.
|
||||||
|
*
|
||||||
|
* These times will be averaged out in the self refresh helpers to avoid
|
||||||
|
* overreacting over one outlier frame
|
||||||
|
*/
|
||||||
|
start = ktime_get();
|
||||||
|
|
||||||
drm_atomic_helper_wait_for_fences(dev, old_state, false);
|
drm_atomic_helper_wait_for_fences(dev, old_state, false);
|
||||||
|
|
||||||
drm_atomic_helper_wait_for_dependencies(old_state);
|
drm_atomic_helper_wait_for_dependencies(old_state);
|
||||||
|
@ -1592,6 +1607,11 @@ static void commit_tail(struct drm_atomic_state *old_state)
|
||||||
else
|
else
|
||||||
drm_atomic_helper_commit_tail(old_state);
|
drm_atomic_helper_commit_tail(old_state);
|
||||||
|
|
||||||
|
commit_time_ms = ktime_ms_delta(ktime_get(), start);
|
||||||
|
if (commit_time_ms > 0)
|
||||||
|
drm_self_refresh_helper_update_avg_times(old_state,
|
||||||
|
(unsigned long)commit_time_ms);
|
||||||
|
|
||||||
drm_atomic_helper_commit_cleanup_done(old_state);
|
drm_atomic_helper_commit_cleanup_done(old_state);
|
||||||
|
|
||||||
drm_atomic_state_put(old_state);
|
drm_atomic_state_put(old_state);
|
||||||
|
@ -3275,7 +3295,7 @@ static int page_flip_common(struct drm_atomic_state *state,
|
||||||
return PTR_ERR(crtc_state);
|
return PTR_ERR(crtc_state);
|
||||||
|
|
||||||
crtc_state->event = event;
|
crtc_state->event = event;
|
||||||
crtc_state->pageflip_flags = flags;
|
crtc_state->async_flip = flags & DRM_MODE_PAGE_FLIP_ASYNC;
|
||||||
|
|
||||||
plane_state = drm_atomic_get_plane_state(state, plane);
|
plane_state = drm_atomic_get_plane_state(state, plane);
|
||||||
if (IS_ERR(plane_state))
|
if (IS_ERR(plane_state))
|
||||||
|
|
|
@ -128,7 +128,7 @@ void __drm_atomic_helper_crtc_duplicate_state(struct drm_crtc *crtc,
|
||||||
state->zpos_changed = false;
|
state->zpos_changed = false;
|
||||||
state->commit = NULL;
|
state->commit = NULL;
|
||||||
state->event = NULL;
|
state->event = NULL;
|
||||||
state->pageflip_flags = 0;
|
state->async_flip = false;
|
||||||
|
|
||||||
/* Self refresh should be canceled when a new update is available */
|
/* Self refresh should be canceled when a new update is available */
|
||||||
state->active = drm_atomic_crtc_effectively_active(state);
|
state->active = drm_atomic_crtc_effectively_active(state);
|
||||||
|
|
|
@ -1305,8 +1305,7 @@ int drm_mode_atomic_ioctl(struct drm_device *dev,
|
||||||
if (arg->reserved)
|
if (arg->reserved)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
if ((arg->flags & DRM_MODE_PAGE_FLIP_ASYNC) &&
|
if (arg->flags & DRM_MODE_PAGE_FLIP_ASYNC)
|
||||||
!dev->mode_config.async_page_flip)
|
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
/* can't test and expect an event at the same time. */
|
/* can't test and expect an event at the same time. */
|
||||||
|
|
|
@ -976,14 +976,14 @@ int drm_dev_register(struct drm_device *dev, unsigned long flags)
|
||||||
if (ret)
|
if (ret)
|
||||||
goto err_minors;
|
goto err_minors;
|
||||||
|
|
||||||
|
dev->registered = true;
|
||||||
|
|
||||||
if (dev->driver->load) {
|
if (dev->driver->load) {
|
||||||
ret = dev->driver->load(dev, flags);
|
ret = dev->driver->load(dev, flags);
|
||||||
if (ret)
|
if (ret)
|
||||||
goto err_minors;
|
goto err_minors;
|
||||||
}
|
}
|
||||||
|
|
||||||
dev->registered = true;
|
|
||||||
|
|
||||||
if (drm_core_check_feature(dev, DRIVER_MODESET))
|
if (drm_core_check_feature(dev, DRIVER_MODESET))
|
||||||
drm_modeset_register_all(dev);
|
drm_modeset_register_all(dev);
|
||||||
|
|
||||||
|
|
|
@ -336,7 +336,12 @@ drm_setclientcap(struct drm_device *dev, void *data, struct drm_file *file_priv)
|
||||||
case DRM_CLIENT_CAP_ATOMIC:
|
case DRM_CLIENT_CAP_ATOMIC:
|
||||||
if (!drm_core_check_feature(dev, DRIVER_ATOMIC))
|
if (!drm_core_check_feature(dev, DRIVER_ATOMIC))
|
||||||
return -EOPNOTSUPP;
|
return -EOPNOTSUPP;
|
||||||
if (req->value > 1)
|
/* The modesetting DDX has a totally broken idea of atomic. */
|
||||||
|
if (current->comm[0] == 'X' && req->value == 1) {
|
||||||
|
pr_info("broken atomic modeset userspace detected, disabling atomic\n");
|
||||||
|
return -EOPNOTSUPP;
|
||||||
|
}
|
||||||
|
if (req->value > 2)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
file_priv->atomic = req->value;
|
file_priv->atomic = req->value;
|
||||||
file_priv->universal_planes = req->value;
|
file_priv->universal_planes = req->value;
|
||||||
|
|
|
@ -42,7 +42,7 @@ int __drm_mode_object_add(struct drm_device *dev, struct drm_mode_object *obj,
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
WARN_ON(dev->registered && !obj_free_cb);
|
WARN_ON(!dev->driver->load && dev->registered && !obj_free_cb);
|
||||||
|
|
||||||
mutex_lock(&dev->mode_config.idr_mutex);
|
mutex_lock(&dev->mode_config.idr_mutex);
|
||||||
ret = idr_alloc(&dev->mode_config.object_idr, register_obj ? obj : NULL,
|
ret = idr_alloc(&dev->mode_config.object_idr, register_obj ? obj : NULL,
|
||||||
|
@ -104,7 +104,7 @@ void drm_mode_object_register(struct drm_device *dev,
|
||||||
void drm_mode_object_unregister(struct drm_device *dev,
|
void drm_mode_object_unregister(struct drm_device *dev,
|
||||||
struct drm_mode_object *object)
|
struct drm_mode_object *object)
|
||||||
{
|
{
|
||||||
WARN_ON(dev->registered && !object->free_cb);
|
WARN_ON(!dev->driver->load && dev->registered && !object->free_cb);
|
||||||
|
|
||||||
mutex_lock(&dev->mode_config.idr_mutex);
|
mutex_lock(&dev->mode_config.idr_mutex);
|
||||||
if (object->id) {
|
if (object->id) {
|
||||||
|
|
|
@ -5,6 +5,7 @@
|
||||||
* Authors:
|
* Authors:
|
||||||
* Sean Paul <seanpaul@chromium.org>
|
* Sean Paul <seanpaul@chromium.org>
|
||||||
*/
|
*/
|
||||||
|
#include <linux/average.h>
|
||||||
#include <linux/bitops.h>
|
#include <linux/bitops.h>
|
||||||
#include <linux/slab.h>
|
#include <linux/slab.h>
|
||||||
#include <linux/workqueue.h>
|
#include <linux/workqueue.h>
|
||||||
|
@ -50,11 +51,17 @@
|
||||||
* atomic_check when &drm_crtc_state.self_refresh_active is true.
|
* atomic_check when &drm_crtc_state.self_refresh_active is true.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
#define SELF_REFRESH_AVG_SEED_MS 200
|
||||||
|
|
||||||
|
DECLARE_EWMA(psr_time, 4, 4)
|
||||||
|
|
||||||
struct drm_self_refresh_data {
|
struct drm_self_refresh_data {
|
||||||
struct drm_crtc *crtc;
|
struct drm_crtc *crtc;
|
||||||
struct delayed_work entry_work;
|
struct delayed_work entry_work;
|
||||||
struct drm_atomic_state *save_state;
|
|
||||||
unsigned int entry_delay_ms;
|
struct mutex avg_mutex;
|
||||||
|
struct ewma_psr_time entry_avg_ms;
|
||||||
|
struct ewma_psr_time exit_avg_ms;
|
||||||
};
|
};
|
||||||
|
|
||||||
static void drm_self_refresh_helper_entry_work(struct work_struct *work)
|
static void drm_self_refresh_helper_entry_work(struct work_struct *work)
|
||||||
|
@ -122,6 +129,44 @@ static void drm_self_refresh_helper_entry_work(struct work_struct *work)
|
||||||
drm_modeset_acquire_fini(&ctx);
|
drm_modeset_acquire_fini(&ctx);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* drm_self_refresh_helper_update_avg_times - Updates a crtc's SR time averages
|
||||||
|
* @state: the state which has just been applied to hardware
|
||||||
|
* @commit_time_ms: the amount of time in ms that this commit took to complete
|
||||||
|
*
|
||||||
|
* Called after &drm_mode_config_funcs.atomic_commit_tail, this function will
|
||||||
|
* update the average entry/exit self refresh times on self refresh transitions.
|
||||||
|
* These averages will be used when calculating how long to delay before
|
||||||
|
* entering self refresh mode after activity.
|
||||||
|
*/
|
||||||
|
void drm_self_refresh_helper_update_avg_times(struct drm_atomic_state *state,
|
||||||
|
unsigned int commit_time_ms)
|
||||||
|
{
|
||||||
|
struct drm_crtc *crtc;
|
||||||
|
struct drm_crtc_state *old_crtc_state, *new_crtc_state;
|
||||||
|
int i;
|
||||||
|
|
||||||
|
for_each_oldnew_crtc_in_state(state, crtc, old_crtc_state,
|
||||||
|
new_crtc_state, i) {
|
||||||
|
struct drm_self_refresh_data *sr_data = crtc->self_refresh_data;
|
||||||
|
struct ewma_psr_time *time;
|
||||||
|
|
||||||
|
if (old_crtc_state->self_refresh_active ==
|
||||||
|
new_crtc_state->self_refresh_active)
|
||||||
|
continue;
|
||||||
|
|
||||||
|
if (new_crtc_state->self_refresh_active)
|
||||||
|
time = &sr_data->entry_avg_ms;
|
||||||
|
else
|
||||||
|
time = &sr_data->exit_avg_ms;
|
||||||
|
|
||||||
|
mutex_lock(&sr_data->avg_mutex);
|
||||||
|
ewma_psr_time_add(time, commit_time_ms);
|
||||||
|
mutex_unlock(&sr_data->avg_mutex);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(drm_self_refresh_helper_update_avg_times);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* drm_self_refresh_helper_alter_state - Alters the atomic state for SR exit
|
* drm_self_refresh_helper_alter_state - Alters the atomic state for SR exit
|
||||||
* @state: the state currently being checked
|
* @state: the state currently being checked
|
||||||
|
@ -153,6 +198,7 @@ void drm_self_refresh_helper_alter_state(struct drm_atomic_state *state)
|
||||||
|
|
||||||
for_each_new_crtc_in_state(state, crtc, crtc_state, i) {
|
for_each_new_crtc_in_state(state, crtc, crtc_state, i) {
|
||||||
struct drm_self_refresh_data *sr_data;
|
struct drm_self_refresh_data *sr_data;
|
||||||
|
unsigned int delay;
|
||||||
|
|
||||||
/* Don't trigger the entry timer when we're already in SR */
|
/* Don't trigger the entry timer when we're already in SR */
|
||||||
if (crtc_state->self_refresh_active)
|
if (crtc_state->self_refresh_active)
|
||||||
|
@ -162,8 +208,13 @@ void drm_self_refresh_helper_alter_state(struct drm_atomic_state *state)
|
||||||
if (!sr_data)
|
if (!sr_data)
|
||||||
continue;
|
continue;
|
||||||
|
|
||||||
|
mutex_lock(&sr_data->avg_mutex);
|
||||||
|
delay = (ewma_psr_time_read(&sr_data->entry_avg_ms) +
|
||||||
|
ewma_psr_time_read(&sr_data->exit_avg_ms)) * 2;
|
||||||
|
mutex_unlock(&sr_data->avg_mutex);
|
||||||
|
|
||||||
mod_delayed_work(system_wq, &sr_data->entry_work,
|
mod_delayed_work(system_wq, &sr_data->entry_work,
|
||||||
msecs_to_jiffies(sr_data->entry_delay_ms));
|
msecs_to_jiffies(delay));
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(drm_self_refresh_helper_alter_state);
|
EXPORT_SYMBOL(drm_self_refresh_helper_alter_state);
|
||||||
|
@ -171,12 +222,10 @@ EXPORT_SYMBOL(drm_self_refresh_helper_alter_state);
|
||||||
/**
|
/**
|
||||||
* drm_self_refresh_helper_init - Initializes self refresh helpers for a crtc
|
* drm_self_refresh_helper_init - Initializes self refresh helpers for a crtc
|
||||||
* @crtc: the crtc which supports self refresh supported displays
|
* @crtc: the crtc which supports self refresh supported displays
|
||||||
* @entry_delay_ms: amount of inactivity to wait before entering self refresh
|
|
||||||
*
|
*
|
||||||
* Returns zero if successful or -errno on failure
|
* Returns zero if successful or -errno on failure
|
||||||
*/
|
*/
|
||||||
int drm_self_refresh_helper_init(struct drm_crtc *crtc,
|
int drm_self_refresh_helper_init(struct drm_crtc *crtc)
|
||||||
unsigned int entry_delay_ms)
|
|
||||||
{
|
{
|
||||||
struct drm_self_refresh_data *sr_data = crtc->self_refresh_data;
|
struct drm_self_refresh_data *sr_data = crtc->self_refresh_data;
|
||||||
|
|
||||||
|
@ -190,8 +239,18 @@ int drm_self_refresh_helper_init(struct drm_crtc *crtc,
|
||||||
|
|
||||||
INIT_DELAYED_WORK(&sr_data->entry_work,
|
INIT_DELAYED_WORK(&sr_data->entry_work,
|
||||||
drm_self_refresh_helper_entry_work);
|
drm_self_refresh_helper_entry_work);
|
||||||
sr_data->entry_delay_ms = entry_delay_ms;
|
|
||||||
sr_data->crtc = crtc;
|
sr_data->crtc = crtc;
|
||||||
|
mutex_init(&sr_data->avg_mutex);
|
||||||
|
ewma_psr_time_init(&sr_data->entry_avg_ms);
|
||||||
|
ewma_psr_time_init(&sr_data->exit_avg_ms);
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Seed the averages so they're non-zero (and sufficiently large
|
||||||
|
* for even poorly performing panels). As time goes on, this will be
|
||||||
|
* averaged out and the values will trend to their true value.
|
||||||
|
*/
|
||||||
|
ewma_psr_time_add(&sr_data->entry_avg_ms, SELF_REFRESH_AVG_SEED_MS);
|
||||||
|
ewma_psr_time_add(&sr_data->exit_avg_ms, SELF_REFRESH_AVG_SEED_MS);
|
||||||
|
|
||||||
crtc->self_refresh_data = sr_data;
|
crtc->self_refresh_data = sr_data;
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -267,7 +267,7 @@ nv50_wndw_atomic_check_acquire(struct nv50_wndw *wndw, bool modeset,
|
||||||
asyw->image.pitch[0] = fb->base.pitches[0];
|
asyw->image.pitch[0] = fb->base.pitches[0];
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!(asyh->state.pageflip_flags & DRM_MODE_PAGE_FLIP_ASYNC))
|
if (!asyh->state.async_flip)
|
||||||
asyw->image.interval = 1;
|
asyw->image.interval = 1;
|
||||||
else
|
else
|
||||||
asyw->image.interval = 0;
|
asyw->image.interval = 0;
|
||||||
|
@ -383,7 +383,7 @@ nv50_wndw_atomic_check_lut(struct nv50_wndw *wndw,
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Can't do an immediate flip while changing the LUT. */
|
/* Can't do an immediate flip while changing the LUT. */
|
||||||
asyh->state.pageflip_flags &= ~DRM_MODE_PAGE_FLIP_ASYNC;
|
asyh->state.async_flip = false;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int
|
static int
|
||||||
|
|
|
@ -39,7 +39,7 @@ static int panfrost_devfreq_target(struct device *dev, unsigned long *freq,
|
||||||
* If frequency scaling from low to high, adjust voltage first.
|
* If frequency scaling from low to high, adjust voltage first.
|
||||||
* If frequency scaling from high to low, adjust frequency first.
|
* If frequency scaling from high to low, adjust frequency first.
|
||||||
*/
|
*/
|
||||||
if (old_clk_rate < target_rate && pfdev->regulator) {
|
if (old_clk_rate < target_rate) {
|
||||||
err = regulator_set_voltage(pfdev->regulator, target_volt,
|
err = regulator_set_voltage(pfdev->regulator, target_volt,
|
||||||
target_volt);
|
target_volt);
|
||||||
if (err) {
|
if (err) {
|
||||||
|
@ -53,14 +53,12 @@ static int panfrost_devfreq_target(struct device *dev, unsigned long *freq,
|
||||||
if (err) {
|
if (err) {
|
||||||
dev_err(dev, "Cannot set frequency %lu (%d)\n", target_rate,
|
dev_err(dev, "Cannot set frequency %lu (%d)\n", target_rate,
|
||||||
err);
|
err);
|
||||||
if (pfdev->regulator)
|
regulator_set_voltage(pfdev->regulator, pfdev->devfreq.cur_volt,
|
||||||
regulator_set_voltage(pfdev->regulator,
|
pfdev->devfreq.cur_volt);
|
||||||
pfdev->devfreq.cur_volt,
|
|
||||||
pfdev->devfreq.cur_volt);
|
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (old_clk_rate > target_rate && pfdev->regulator) {
|
if (old_clk_rate > target_rate) {
|
||||||
err = regulator_set_voltage(pfdev->regulator, target_volt,
|
err = regulator_set_voltage(pfdev->regulator, target_volt,
|
||||||
target_volt);
|
target_volt);
|
||||||
if (err)
|
if (err)
|
||||||
|
|
|
@ -89,12 +89,9 @@ static int panfrost_regulator_init(struct panfrost_device *pfdev)
|
||||||
{
|
{
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
pfdev->regulator = devm_regulator_get_optional(pfdev->dev, "mali");
|
pfdev->regulator = devm_regulator_get(pfdev->dev, "mali");
|
||||||
if (IS_ERR(pfdev->regulator)) {
|
if (IS_ERR(pfdev->regulator)) {
|
||||||
ret = PTR_ERR(pfdev->regulator);
|
ret = PTR_ERR(pfdev->regulator);
|
||||||
pfdev->regulator = NULL;
|
|
||||||
if (ret == -ENODEV)
|
|
||||||
return 0;
|
|
||||||
dev_err(pfdev->dev, "failed to get regulator: %d\n", ret);
|
dev_err(pfdev->dev, "failed to get regulator: %d\n", ret);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
@ -110,8 +107,7 @@ static int panfrost_regulator_init(struct panfrost_device *pfdev)
|
||||||
|
|
||||||
static void panfrost_regulator_fini(struct panfrost_device *pfdev)
|
static void panfrost_regulator_fini(struct panfrost_device *pfdev)
|
||||||
{
|
{
|
||||||
if (pfdev->regulator)
|
regulator_disable(pfdev->regulator);
|
||||||
regulator_disable(pfdev->regulator);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
int panfrost_device_init(struct panfrost_device *pfdev)
|
int panfrost_device_init(struct panfrost_device *pfdev)
|
||||||
|
|
|
@ -386,28 +386,40 @@ void panfrost_mmu_pgtable_free(struct panfrost_file_priv *priv)
|
||||||
free_io_pgtable_ops(mmu->pgtbl_ops);
|
free_io_pgtable_ops(mmu->pgtbl_ops);
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct drm_mm_node *addr_to_drm_mm_node(struct panfrost_device *pfdev, int as, u64 addr)
|
static struct panfrost_gem_object *
|
||||||
|
addr_to_drm_mm_node(struct panfrost_device *pfdev, int as, u64 addr)
|
||||||
{
|
{
|
||||||
struct drm_mm_node *node = NULL;
|
struct panfrost_gem_object *bo = NULL;
|
||||||
|
struct panfrost_file_priv *priv;
|
||||||
|
struct drm_mm_node *node;
|
||||||
u64 offset = addr >> PAGE_SHIFT;
|
u64 offset = addr >> PAGE_SHIFT;
|
||||||
struct panfrost_mmu *mmu;
|
struct panfrost_mmu *mmu;
|
||||||
|
|
||||||
spin_lock(&pfdev->as_lock);
|
spin_lock(&pfdev->as_lock);
|
||||||
list_for_each_entry(mmu, &pfdev->as_lru_list, list) {
|
list_for_each_entry(mmu, &pfdev->as_lru_list, list) {
|
||||||
struct panfrost_file_priv *priv;
|
if (as == mmu->as)
|
||||||
if (as != mmu->as)
|
break;
|
||||||
continue;
|
}
|
||||||
|
if (as != mmu->as)
|
||||||
|
goto out;
|
||||||
|
|
||||||
priv = container_of(mmu, struct panfrost_file_priv, mmu);
|
priv = container_of(mmu, struct panfrost_file_priv, mmu);
|
||||||
drm_mm_for_each_node(node, &priv->mm) {
|
|
||||||
if (offset >= node->start && offset < (node->start + node->size))
|
spin_lock(&priv->mm_lock);
|
||||||
goto out;
|
|
||||||
|
drm_mm_for_each_node(node, &priv->mm) {
|
||||||
|
if (offset >= node->start &&
|
||||||
|
offset < (node->start + node->size)) {
|
||||||
|
bo = drm_mm_node_to_panfrost_bo(node);
|
||||||
|
drm_gem_object_get(&bo->base.base);
|
||||||
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
spin_unlock(&priv->mm_lock);
|
||||||
out:
|
out:
|
||||||
spin_unlock(&pfdev->as_lock);
|
spin_unlock(&pfdev->as_lock);
|
||||||
return node;
|
return bo;
|
||||||
}
|
}
|
||||||
|
|
||||||
#define NUM_FAULT_PAGES (SZ_2M / PAGE_SIZE)
|
#define NUM_FAULT_PAGES (SZ_2M / PAGE_SIZE)
|
||||||
|
@ -415,29 +427,28 @@ static struct drm_mm_node *addr_to_drm_mm_node(struct panfrost_device *pfdev, in
|
||||||
int panfrost_mmu_map_fault_addr(struct panfrost_device *pfdev, int as, u64 addr)
|
int panfrost_mmu_map_fault_addr(struct panfrost_device *pfdev, int as, u64 addr)
|
||||||
{
|
{
|
||||||
int ret, i;
|
int ret, i;
|
||||||
struct drm_mm_node *node;
|
|
||||||
struct panfrost_gem_object *bo;
|
struct panfrost_gem_object *bo;
|
||||||
struct address_space *mapping;
|
struct address_space *mapping;
|
||||||
pgoff_t page_offset;
|
pgoff_t page_offset;
|
||||||
struct sg_table *sgt;
|
struct sg_table *sgt;
|
||||||
struct page **pages;
|
struct page **pages;
|
||||||
|
|
||||||
node = addr_to_drm_mm_node(pfdev, as, addr);
|
bo = addr_to_drm_mm_node(pfdev, as, addr);
|
||||||
if (!node)
|
if (!bo)
|
||||||
return -ENOENT;
|
return -ENOENT;
|
||||||
|
|
||||||
bo = drm_mm_node_to_panfrost_bo(node);
|
|
||||||
if (!bo->is_heap) {
|
if (!bo->is_heap) {
|
||||||
dev_WARN(pfdev->dev, "matching BO is not heap type (GPU VA = %llx)",
|
dev_WARN(pfdev->dev, "matching BO is not heap type (GPU VA = %llx)",
|
||||||
node->start << PAGE_SHIFT);
|
bo->node.start << PAGE_SHIFT);
|
||||||
return -EINVAL;
|
ret = -EINVAL;
|
||||||
|
goto err_bo;
|
||||||
}
|
}
|
||||||
WARN_ON(bo->mmu->as != as);
|
WARN_ON(bo->mmu->as != as);
|
||||||
|
|
||||||
/* Assume 2MB alignment and size multiple */
|
/* Assume 2MB alignment and size multiple */
|
||||||
addr &= ~((u64)SZ_2M - 1);
|
addr &= ~((u64)SZ_2M - 1);
|
||||||
page_offset = addr >> PAGE_SHIFT;
|
page_offset = addr >> PAGE_SHIFT;
|
||||||
page_offset -= node->start;
|
page_offset -= bo->node.start;
|
||||||
|
|
||||||
mutex_lock(&bo->base.pages_lock);
|
mutex_lock(&bo->base.pages_lock);
|
||||||
|
|
||||||
|
@ -446,7 +457,8 @@ int panfrost_mmu_map_fault_addr(struct panfrost_device *pfdev, int as, u64 addr)
|
||||||
sizeof(struct sg_table), GFP_KERNEL | __GFP_ZERO);
|
sizeof(struct sg_table), GFP_KERNEL | __GFP_ZERO);
|
||||||
if (!bo->sgts) {
|
if (!bo->sgts) {
|
||||||
mutex_unlock(&bo->base.pages_lock);
|
mutex_unlock(&bo->base.pages_lock);
|
||||||
return -ENOMEM;
|
ret = -ENOMEM;
|
||||||
|
goto err_bo;
|
||||||
}
|
}
|
||||||
|
|
||||||
pages = kvmalloc_array(bo->base.base.size >> PAGE_SHIFT,
|
pages = kvmalloc_array(bo->base.base.size >> PAGE_SHIFT,
|
||||||
|
@ -455,7 +467,8 @@ int panfrost_mmu_map_fault_addr(struct panfrost_device *pfdev, int as, u64 addr)
|
||||||
kfree(bo->sgts);
|
kfree(bo->sgts);
|
||||||
bo->sgts = NULL;
|
bo->sgts = NULL;
|
||||||
mutex_unlock(&bo->base.pages_lock);
|
mutex_unlock(&bo->base.pages_lock);
|
||||||
return -ENOMEM;
|
ret = -ENOMEM;
|
||||||
|
goto err_bo;
|
||||||
}
|
}
|
||||||
bo->base.pages = pages;
|
bo->base.pages = pages;
|
||||||
bo->base.pages_use_count = 1;
|
bo->base.pages_use_count = 1;
|
||||||
|
@ -493,12 +506,16 @@ int panfrost_mmu_map_fault_addr(struct panfrost_device *pfdev, int as, u64 addr)
|
||||||
|
|
||||||
dev_dbg(pfdev->dev, "mapped page fault @ AS%d %llx", as, addr);
|
dev_dbg(pfdev->dev, "mapped page fault @ AS%d %llx", as, addr);
|
||||||
|
|
||||||
|
drm_gem_object_put_unlocked(&bo->base.base);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
err_map:
|
err_map:
|
||||||
sg_free_table(sgt);
|
sg_free_table(sgt);
|
||||||
err_pages:
|
err_pages:
|
||||||
drm_gem_shmem_put_pages(&bo->base);
|
drm_gem_shmem_put_pages(&bo->base);
|
||||||
|
err_bo:
|
||||||
|
drm_gem_object_put_unlocked(&bo->base.base);
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -39,8 +39,6 @@
|
||||||
#include "rockchip_drm_vop.h"
|
#include "rockchip_drm_vop.h"
|
||||||
#include "rockchip_rgb.h"
|
#include "rockchip_rgb.h"
|
||||||
|
|
||||||
#define VOP_SELF_REFRESH_ENTRY_DELAY_MS 100
|
|
||||||
|
|
||||||
#define VOP_WIN_SET(vop, win, name, v) \
|
#define VOP_WIN_SET(vop, win, name, v) \
|
||||||
vop_reg_set(vop, &win->phy->name, win->base, ~0, v, #name)
|
vop_reg_set(vop, &win->phy->name, win->base, ~0, v, #name)
|
||||||
#define VOP_SCL_SET(vop, win, name, v) \
|
#define VOP_SCL_SET(vop, win, name, v) \
|
||||||
|
@ -1563,8 +1561,7 @@ static int vop_create_crtc(struct vop *vop)
|
||||||
init_completion(&vop->line_flag_completion);
|
init_completion(&vop->line_flag_completion);
|
||||||
crtc->port = port;
|
crtc->port = port;
|
||||||
|
|
||||||
ret = drm_self_refresh_helper_init(crtc,
|
ret = drm_self_refresh_helper_init(crtc);
|
||||||
VOP_SELF_REFRESH_ENTRY_DELAY_MS);
|
|
||||||
if (ret)
|
if (ret)
|
||||||
DRM_DEV_DEBUG_KMS(vop->dev,
|
DRM_DEV_DEBUG_KMS(vop->dev,
|
||||||
"Failed to init %s with SR helpers %d, ignoring\n",
|
"Failed to init %s with SR helpers %d, ignoring\n",
|
||||||
|
|
|
@ -285,12 +285,12 @@ struct drm_crtc_state {
|
||||||
u32 target_vblank;
|
u32 target_vblank;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @pageflip_flags:
|
* @async_flip:
|
||||||
*
|
*
|
||||||
* DRM_MODE_PAGE_FLIP_* flags, as passed to the page flip ioctl.
|
* This is set when DRM_MODE_PAGE_FLIP_ASYNC is set in the legacy
|
||||||
* Zero in any other case.
|
* PAGE_FLIP IOCTL. It's not wired up for the atomic IOCTL itself yet.
|
||||||
*/
|
*/
|
||||||
u32 pageflip_flags;
|
bool async_flip;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @vrr_enabled:
|
* @vrr_enabled:
|
||||||
|
@ -1108,7 +1108,7 @@ struct drm_crtc {
|
||||||
/**
|
/**
|
||||||
* @self_refresh_data: Holds the state for the self refresh helpers
|
* @self_refresh_data: Holds the state for the self refresh helpers
|
||||||
*
|
*
|
||||||
* Initialized via drm_self_refresh_helper_register().
|
* Initialized via drm_self_refresh_helper_init().
|
||||||
*/
|
*/
|
||||||
struct drm_self_refresh_data *self_refresh_data;
|
struct drm_self_refresh_data *self_refresh_data;
|
||||||
};
|
};
|
||||||
|
|
|
@ -12,9 +12,9 @@ struct drm_atomic_state;
|
||||||
struct drm_crtc;
|
struct drm_crtc;
|
||||||
|
|
||||||
void drm_self_refresh_helper_alter_state(struct drm_atomic_state *state);
|
void drm_self_refresh_helper_alter_state(struct drm_atomic_state *state);
|
||||||
|
void drm_self_refresh_helper_update_avg_times(struct drm_atomic_state *state,
|
||||||
|
unsigned int commit_time_ms);
|
||||||
|
|
||||||
int drm_self_refresh_helper_init(struct drm_crtc *crtc,
|
int drm_self_refresh_helper_init(struct drm_crtc *crtc);
|
||||||
unsigned int entry_delay_ms);
|
|
||||||
|
|
||||||
void drm_self_refresh_helper_cleanup(struct drm_crtc *crtc);
|
void drm_self_refresh_helper_cleanup(struct drm_crtc *crtc);
|
||||||
#endif
|
#endif
|
||||||
|
|
Loading…
Reference in New Issue