diff options
Diffstat (limited to 'man/intel.man')
-rw-r--r-- | man/intel.man | 168 |
1 files changed, 23 insertions, 145 deletions
diff --git a/man/intel.man b/man/intel.man index 1673702b..9344d28c 100644 --- a/man/intel.man +++ b/man/intel.man @@ -61,6 +61,16 @@ This sets the default pixel value for the YUV video overlay key. .IP Default: undefined. .TP +.BI "Option \*qDRI\*q \*q" boolean \*q +Disable or enable DRI support. +.IP +Default: DRI is enabled for configurations where it is supported. + +.PP +The following driver +.B Options +are supported for the i810 and i815 chipsets: +.TP .BI "Option \*qCacheLines\*q \*q" integer \*q This allows the user to change the amount of graphics memory used for 2D acceleration and video when XAA acceleration is enabled. Decreasing this @@ -72,41 +82,6 @@ driver attempts to allocate space for at 3 screenfuls of pixmaps plus an HD-sized XV video. The default used for a specific configuration can be found by examining the __xservername__ log file. .TP -.BI "Option \*qFramebufferCompression\*q \*q" boolean \*q -This option controls whether the framebuffer compression feature is enabled. -If possible, the front buffer will be allocated in a tiled format and compressed -periodically to save memory bandwidth and power. -This option is only available on mobile chipsets. -.IP -Default: enabled on supported configurations. -.TP -.BI "Option \*qTiling\*q \*q" boolean \*q -This option controls whether memory buffers are allocated in tiled mode. In -most cases (especially for complex rendering), tiling dramatically improves -performance. -.IP -Default: enabled. -.TP -.BI "Option \*qSwapbuffersWait\*q \*q" boolean \*q -This option controls the behavior of glXSwapBuffers and glXCopySubBufferMESA -calls by GL applications. If enabled, the calls will avoid tearing by making -sure the display scanline is outside of the area to be copied before the copy -occurs. If disabled, no scanline synchronization is performed, meaning tearing -will likely occur. Note that when enabled, this option can adversely affect -the framerate of applications that render frames at less than refresh rate. -.IP -Default: enabled. -.TP -.BI "Option \*qDRI\*q \*q" boolean \*q -Disable or enable DRI support. -.IP -Default: DRI is enabled for configurations where it is supported. - -.PP -The following driver -.B Options -are supported for the i810 and i815 chipsets: -.TP .BI "Option \*qDDC\*q \*q" boolean \*q Disable or enable DDC support. .IP @@ -162,41 +137,22 @@ server log. .IP Default: Disabled .TP -.BI "Option \*qForceEnablePipeA\*q \*q" boolean \*q -Force the driver to leave pipe A enabled. May be necessary in configurations -where the BIOS accesses pipe registers during display hotswitch or lid close, -causing a crash. If you find that your platform needs this option, please file -a bug (see REPORTING BUGS below) including the output of 'lspci -v' and 'lspci -vn'. -.TP -.BI "Option \*qLVDS24Bit\*q \*q" boolean \*q -Specify 24 bit pixel format (i.e. 8 bits per color) to be used for the -LVDS output. Some newer LCD panels expect pixels to be formatted and -sent as 8 bits per color channel instead of the more common 6 bits per -color channel. Set this option to true to enable the newer format. -Note that this concept is entirely different and independent from the -frame buffer color depth - which is still controlled in the usual way -within the X server. This option instead selects the physical format -/ sequencing of the digital bits sent to the display. Setting the -frame buffer color depth is really a matter of preference by the user, -while setting the pixel format here is a requirement of the connected -hardware. +.BI "Option \*qSwapbuffersWait\*q \*q" boolean \*q +This option controls the behavior of glXSwapBuffers and glXCopySubBufferMESA +calls by GL applications. If enabled, the calls will avoid tearing by making +sure the display scanline is outside of the area to be copied before the copy +occurs. If disabled, no scanline synchronization is performed, meaning tearing +will likely occur. Note that when enabled, this option can adversely affect +the framerate of applications that render frames at less than refresh rate. .IP -Leaving this unset implies the default value of false, -which is almost always going to be right choice. If your -LVDS-connected display on the other hand is extremely washed out -(e.g. white on a lighter white), trying this option might clear the -problem. +Default: enabled. .TP -.BI "Option \*qLVDSFixedMode\*q \*q" boolean \*q -Use a fixed set of timings for the LVDS output, independent of normal -xorg specified timings. +.BI "Option \*qTiling\*q \*q" boolean \*q +This option controls whether memory buffers are allocated in tiled mode. In +most cases (especially for complex rendering), tiling dramatically improves +performance. .IP -The default value if left unspecified is -true, which is what you want for a normal LVDS-connected LCD type of -panel. If you are not sure about this, leave it at its default, which -allows the driver to automatically figure out the correct fixed panel -timings. See further in the section about LVDS fixed timing for more -information. +Default: enabled. .TP .BI "Option \*qXvMC\*q \*q" boolean \*q Enable XvMC driver. Current support MPEG2 MC on 915/945 and G33 series. @@ -332,84 +288,6 @@ sections with these outputs for configuration. Associating Monitor sections with each output can be helpful if you need to ignore a specific output, for example, or statically configure an extended desktop monitor layout. -.SH HARDWARE LVDS FIXED TIMINGS AND SCALING - -Following here is a discussion that should shed some light on the -nature and reasoning behind the LVDSFixedMode option. - -Unlike a CRT display, an LCD has a "native" resolution corresponding -to the actual pixel geometry. A graphics controller under all normal -circumstances should always output that resolution (and timings) to -the display. Anything else and the image might not fill the display, -it might not be centered, or it might have information missing - any -manner of strange effects can happen if an LCD panel is not fed with -the expected resolution and timings. - -However there are cases where one might want to run an LCD panel at an -effective resolution other than the native one. And for this reason, -GPUs which drive LCD panels typically include a hardware scaler to -match the user-configured frame buffer size to the actual size of the -panel. Thus when one "sets" his/her 1280x1024 panel to only 1024x768, -the GPU happily configures a 1024x768 frame buffer, but it scans the -buffer out in such a way that the image is scaled to 1280x1024 and in -fact sends 1280x1024 to the panel. This is normally invisible to the -user; when a "fuzzy" LCD image is seen, scaling like this is why this -happens. - -In order to make this magic work, this driver logically has to be -configured with two sets of monitor timings - the set specified (or -otherwise determined) as the normal xorg "mode", and the "fixed" -timings that are actually sent to the monitor. But with xorg, it's -only possible to specify the first user-driven set, and not the second -fixed set. So how does the driver figure out the correct fixed panel -timings? Normally it will attempt to detect the fixed timings, and it -uses a number of strategies to figure this out. First it attempts to -read EDID data from whatever is connected to the LVDS port. Failing -that, it will check if the LVDS output is already configured (perhaps -previously by the video BIOS) and will adopt those settings if found. -Failing that, it will scan the video BIOS ROM, looking for an embedded -mode table from which it can infer the proper timings. If even that -fails, then the driver gives up, prints the message "Couldn't detect -panel mode. Disabling panel" to the X server log, and shuts down the -LVDS output. - -Under most circumstances, the detection scheme works. However there -are cases when it can go awry. For example, if you have a panel -without EDID support and it isn't integral to the motherboard -(i.e. not a laptop), then odds are the driver is either not going to -find something suitable to use or it is going to find something -flat-out wrong, leaving a messed up display. Remember that this is -about the fixed timings being discussed here and not the -user-specified timings which can always be set in xorg.conf in the -worst case. So when this process goes awry there seems to be little -recourse. This sort of scenario can happen in some embedded -applications. - -The LVDSFixedMode option is present to deal with this. This option -normally enables the above-described detection strategy. And since it -defaults to true, this is in fact what normally happens. However if -the detection fails to do the right thing, the LVDSFixedMode option -can instead be set to false, which disables all the magic. With -LVDSFixedMode set to false, the detection steps are skipped and the -driver proceeds without a specified fixed mode timing. This then -causes the hardware scaler to be disabled, and the actual timings then -used fall back to those normally configured via the usual xorg -mechanisms. - -Having LVDSFixedMode set to false means that whatever is used for the -monitor's mode (e.g. a modeline setting) is precisely what is sent to -the device connected to the LVDS port. This also means that the user -now has to determine the correct mode to use - but it's really no -different than the work for correctly configuring an old-school CRT -anyway, and the alternative if detection fails will be a useless -display. - -In short, leave LVDSFixedMode alone (thus set to true) and normal -fixed mode detection will take place, which in most cases is exactly -what is needed. Set LVDSFixedMode to false and then the user has full -control over the resolution and timings sent to the LVDS-connected -device, through the usual means in xorg. - .SH MULTIHEAD CONFIGURATIONS The number of independent outputs is dictated by the number of CRTCs |