diff options
author | Jonathan Gray <jsg@jsg.id.au> | 2013-03-11 23:46:47 +1100 |
---|---|---|
committer | Jonathan Gray <jsg@jsg.id.au> | 2013-03-11 23:46:47 +1100 |
commit | cc9656e49a6b595557cd3c0ff1f94c024605138f (patch) | |
tree | a2e661d4bbc9196b6d6469acdbdd6d60c3bf1a04 | |
parent | 2e28413d6f14130f89149d17ec6f451635b49ad3 (diff) |
unstub the rest of intel_pm.c
-rw-r--r-- | sys/dev/pci/drm/drmP.h | 1 | ||||
-rw-r--r-- | sys/dev/pci/drm/i915_drv.h | 5 | ||||
-rw-r--r-- | sys/dev/pci/drm/intel_drv.h | 3 | ||||
-rw-r--r-- | sys/dev/pci/drm/intel_pm.c | 104 |
4 files changed, 57 insertions, 56 deletions
diff --git a/sys/dev/pci/drm/drmP.h b/sys/dev/pci/drm/drmP.h index cb4796ec8eb..0b04858cd12 100644 --- a/sys/dev/pci/drm/drmP.h +++ b/sys/dev/pci/drm/drmP.h @@ -119,6 +119,7 @@ #define PAGE_ALIGN(addr) (((addr) + PAGE_MASK) & ~PAGE_MASK) #define roundup2(x, y) (((x)+((y)-1))&(~((y)-1))) /* if y is powers of two */ +#define jiffies_to_msecs(x) (((int64_t)(x)) * 1000 / hz) #define msecs_to_jiffies(x) (((int64_t)(x)) * hz / 1000) #define time_after_eq(a,b) ((long)(b) - (long)(a) <= 0) #define drm_msleep(x, msg) delay(x * 1000) diff --git a/sys/dev/pci/drm/i915_drv.h b/sys/dev/pci/drm/i915_drv.h index 2ce618b8354..4e766505866 100644 --- a/sys/dev/pci/drm/i915_drv.h +++ b/sys/dev/pci/drm/i915_drv.h @@ -439,9 +439,8 @@ struct intel_gen6_power_mgmt { u8 min_delay; u8 max_delay; -#if 0 - struct delayed_work delayed_resume_work; -#endif + struct workq_task delayed_resume_task; + struct timeout delayed_resume_to; /* * Protects RPS/RC6 register access and PCU communication. diff --git a/sys/dev/pci/drm/intel_drv.h b/sys/dev/pci/drm/intel_drv.h index fb86620c125..5eaadbafbe9 100644 --- a/sys/dev/pci/drm/intel_drv.h +++ b/sys/dev/pci/drm/intel_drv.h @@ -376,7 +376,8 @@ struct intel_unpin_work { }; struct intel_fbc_work { -// struct delayed_work work; + struct workq_task task; + struct timeout to; struct drm_crtc *crtc; struct drm_framebuffer *fb; int interval; diff --git a/sys/dev/pci/drm/intel_pm.c b/sys/dev/pci/drm/intel_pm.c index 49041dabcf3..8aa880fd994 100644 --- a/sys/dev/pci/drm/intel_pm.c +++ b/sys/dev/pci/drm/intel_pm.c @@ -147,6 +147,12 @@ void vlv_force_wake_get(struct inteldrm_softc *); void vlv_force_wake_put(struct inteldrm_softc *); void intel_gt_reset(struct drm_device *); void intel_gt_init(struct drm_device *); +void intel_fbc_work_tick(void *); +void intel_fbc_work_fn(void *, void *); +void intel_gen6_powersave_tick(void *); +void intel_gen6_powersave_work(void *, void *); + +extern int ticks; bool intel_crtc_active(struct drm_crtc *crtc) @@ -385,13 +391,19 @@ intel_fbc_enabled(struct drm_device *dev) return dev_priv->display.fbc_enabled(dev); } -#ifdef notyet void -intel_fbc_work_fn(struct work_struct *__work) +intel_fbc_work_tick(void *arg) +{ + struct intel_fbc_work *work = arg; + + workq_queue_task(NULL, &work->task, 0, + intel_fbc_work_fn, work, NULL); +} + +void +intel_fbc_work_fn(void *arg1, void *arg2) { - struct intel_fbc_work *work = - container_of(to_delayed_work(__work), - struct intel_fbc_work, work); + struct intel_fbc_work *work = arg1; struct drm_device *dev = work->crtc->dev; struct inteldrm_softc *dev_priv = dev->dev_private; @@ -415,7 +427,6 @@ intel_fbc_work_fn(struct work_struct *__work) free(work, M_DRM); } -#endif void intel_cancel_fbc_work(struct inteldrm_softc *dev_priv) @@ -429,11 +440,9 @@ intel_cancel_fbc_work(struct inteldrm_softc *dev_priv) * dev_priv->fbc_work, so we can perform the cancellation * entirely asynchronously. */ -#ifdef notyet - if (cancel_delayed_work(&dev_priv->fbc_work->work)) + if (timeout_del(&dev_priv->fbc_work->to)) /* tasklet was killed before being run, clean up */ free(dev_priv->fbc_work, M_DRM); -#endif /* Mark the work as no longer wanted so that if it does * wake-up (because the work was already running and waiting @@ -464,7 +473,7 @@ intel_enable_fbc(struct drm_crtc *crtc, unsigned long interval) work->crtc = crtc; work->fb = crtc->fb; work->interval = interval; -// INIT_DELAYED_WORK(&work->work, intel_fbc_work_fn); + timeout_set(&work->to, intel_fbc_work_tick, work); dev_priv->fbc_work = work; @@ -481,7 +490,7 @@ intel_enable_fbc(struct drm_crtc *crtc, unsigned long interval) * and indeed performing the enable as a co-routine and not * waiting synchronously upon the vblank. */ -// schedule_delayed_work(&work->work, msecs_to_jiffies(50)); + timeout_add_msec(&work->to, 50); } void @@ -2420,9 +2429,6 @@ intel_update_sprite_watermarks(struct drm_device *dev, int pipe, struct drm_i915_gem_object * intel_alloc_context_page(struct drm_device *dev) { - printf("%s stub\n", __func__); - return NULL; -#ifdef notyet struct drm_i915_gem_object *ctx; int ret; @@ -2436,7 +2442,7 @@ intel_alloc_context_page(struct drm_device *dev) return NULL; } - ret = i915_gem_object_pin(ctx, 4096, true, false); + ret = i915_gem_object_pin(ctx, 4096, true /*, false */); if (ret) { DRM_ERROR("failed to pin power context: %d\n", ret); goto err_unref; @@ -2454,9 +2460,8 @@ err_unpin: i915_gem_object_unpin(ctx); err_unref: drm_gem_object_unreference(&ctx->base); - mutex_unlock(&dev->struct_mutex); + DRM_UNLOCK(); return NULL; -#endif } struct mutex mchdev_lock; @@ -2555,7 +2560,7 @@ ironlake_enable_drps(struct drm_device *dev) dev_priv->ips.last_count1 = I915_READ(0x112e4) + I915_READ(0x112e8) + I915_READ(0x112e0); -// dev_priv->ips.last_time1 = jiffies_to_msecs(jiffies); + dev_priv->ips.last_time1 = jiffies_to_msecs(ticks); dev_priv->ips.last_count2 = I915_READ(0x112f4); nanotime(&dev_priv->ips.last_time2); @@ -2845,8 +2850,6 @@ gen6_enable_rps(struct drm_device *dev) void gen6_update_ring_freq(struct drm_device *dev) { - printf("%s stub\n", __func__); -#ifdef notyet struct inteldrm_softc *dev_priv = dev->dev_private; int min_freq = 15; int gpu_freq; @@ -2855,6 +2858,7 @@ gen6_update_ring_freq(struct drm_device *dev) rw_assert_wrlock(&dev_priv->rps.hw_lock); +#ifdef notyet max_ia_freq = cpufreq_quick_get_max(0); /* * Default to measured freq if none found, PCU will ensure we don't go @@ -2865,6 +2869,10 @@ gen6_update_ring_freq(struct drm_device *dev) /* Convert from kHz to MHz */ max_ia_freq /= 1000; +#else + /* we ideally want the max so avoid cpuspeed */ + max_ia_freq = (curcpu()->ci_tsc_freq + 4999) / 1000000; +#endif /* * For each potential GPU frequency, load a ring frequency we'd like @@ -2874,6 +2882,7 @@ gen6_update_ring_freq(struct drm_device *dev) for (gpu_freq = dev_priv->rps.max_delay; gpu_freq >= dev_priv->rps.min_delay; gpu_freq--) { int diff = dev_priv->rps.max_delay - gpu_freq; + int d; /* * For GPU frequencies less than 750MHz, just use the lowest @@ -2883,14 +2892,14 @@ gen6_update_ring_freq(struct drm_device *dev) ia_freq = 800; else ia_freq = max_ia_freq - ((diff * scaling_factor) / 2); - ia_freq = DIV_ROUND_CLOSEST(ia_freq, 100); + d = 100; + ia_freq = (ia_freq + d / 2) / d; ia_freq <<= GEN6_PCODE_FREQ_IA_RATIO_SHIFT; sandybridge_pcode_write(dev_priv, GEN6_PCODE_WRITE_MIN_FREQ_TABLE, ia_freq | gpu_freq); } -#endif } void @@ -3052,12 +3061,9 @@ static const struct cparams { unsigned long __i915_chipset_val(struct inteldrm_softc *dev_priv) { - printf("%s stub\n", __func__); - return 0; -#ifdef notyet u64 total_count, diff, ret; u32 count1, count2, count3, m = 0, c = 0; - unsigned long now = jiffies_to_msecs(jiffies), diff1; + unsigned long now = jiffies_to_msecs(ticks), diff1; int i; // assert_spin_locked(&mchdev_lock); @@ -3095,9 +3101,9 @@ __i915_chipset_val(struct inteldrm_softc *dev_priv) } } - diff = div_u64(diff, diff1); + diff = diff / diff1; ret = ((m * diff) + c); - ret = div_u64(ret, 10); + ret = ret / 10; dev_priv->ips.last_count1 = total_count; dev_priv->ips.last_time1 = now; @@ -3105,7 +3111,6 @@ __i915_chipset_val(struct inteldrm_softc *dev_priv) dev_priv->ips.chipset_power = ret; return ret; -#endif } unsigned long @@ -3286,8 +3291,6 @@ pvid_to_extvid(struct inteldrm_softc *dev_priv, u8 pxvid) void __i915_update_gfx_val(struct inteldrm_softc *dev_priv) { - printf("%s stub\n", __func__); -#ifdef notyet struct timespec now, diff1; u64 diff; unsigned long diffms; @@ -3296,7 +3299,7 @@ __i915_update_gfx_val(struct inteldrm_softc *dev_priv) // assert_spin_locked(&mchdev_lock); nanotime(&now); - diff1 = timespec_sub(now, dev_priv->ips.last_time2); + timespecsub(&now, &dev_priv->ips.last_time2, &diff1); /* Don't divide by 0 */ diffms = diff1.tv_sec * 1000 + diff1.tv_nsec / 1000000; @@ -3317,9 +3320,8 @@ __i915_update_gfx_val(struct inteldrm_softc *dev_priv) /* More magic constants... */ diff = diff * 1181; - diff = div_u64(diff, diffms * 10); + diff = diff / (diffms * 10); dev_priv->ips.gfx_power = diff; -#endif } void @@ -3654,22 +3656,26 @@ intel_disable_gt_powersave(struct drm_device *dev) ironlake_disable_drps(dev); ironlake_disable_rc6(dev); } else if (INTEL_INFO(dev)->gen >= 6 && !IS_VALLEYVIEW(dev)) { -#ifdef notyet - cancel_delayed_work_sync(&dev_priv->rps.delayed_resume_work); -#endif + timeout_del(&dev_priv->rps.delayed_resume_to); rw_enter_write(&dev_priv->rps.hw_lock); gen6_disable_rps(dev); rw_exit_write(&dev_priv->rps.hw_lock); } } -#ifdef notyet void -intel_gen6_powersave_work(struct work_struct *work) +intel_gen6_powersave_tick(void *arg) +{ + drm_i915_private_t *dev_priv = arg; + + workq_queue_task(NULL, &dev_priv->rps.delayed_resume_task, 0, + intel_gen6_powersave_work, dev_priv, NULL); +} + +void +intel_gen6_powersave_work(void *arg1, void *arg2) { - struct inteldrm_softc *dev_priv = - container_of(work, struct inteldrm_softc, - rps.delayed_resume_work.work); + drm_i915_private_t *dev_priv = arg1; struct drm_device *dev = (struct drm_device *)dev_priv->drmdev; rw_enter_write(&dev_priv->rps.hw_lock); @@ -3677,12 +3683,11 @@ intel_gen6_powersave_work(struct work_struct *work) gen6_update_ring_freq(dev); rw_exit_write(&dev_priv->rps.hw_lock); } -#endif void intel_enable_gt_powersave(struct drm_device *dev) { -// struct inteldrm_softc *dev_priv = dev->dev_private; + struct inteldrm_softc *dev_priv = dev->dev_private; if (IS_IRONLAKE_M(dev)) { ironlake_enable_drps(dev); @@ -3694,10 +3699,7 @@ intel_enable_gt_powersave(struct drm_device *dev) * done at any specific time, so do this out of our fast path * to make resume and init faster. */ -#ifdef notyet - schedule_delayed_work(&dev_priv->rps.delayed_resume_work, - round_jiffies_up_relative(HZ)); -#endif + timeout_add_sec(&dev_priv->rps.delayed_resume_to, 1); } } @@ -4685,10 +4687,8 @@ intel_gt_init(struct drm_device *dev) dev_priv->gt.force_wake_get = __gen6_gt_force_wake_get; dev_priv->gt.force_wake_put = __gen6_gt_force_wake_put; } -#ifdef notyet - INIT_DELAYED_WORK(&dev_priv->rps.delayed_resume_work, - intel_gen6_powersave_work); -#endif + timeout_set(&dev_priv->rps.delayed_resume_to, intel_gen6_powersave_tick, + dev_priv); } int |