summaryrefslogtreecommitdiff
path: root/man/intel.man
diff options
context:
space:
mode:
Diffstat (limited to 'man/intel.man')
-rw-r--r--man/intel.man168
1 files changed, 23 insertions, 145 deletions
diff --git a/man/intel.man b/man/intel.man
index 1673702b..9344d28c 100644
--- a/man/intel.man
+++ b/man/intel.man
@@ -61,6 +61,16 @@ This sets the default pixel value for the YUV video overlay key.
.IP
Default: undefined.
.TP
+.BI "Option \*qDRI\*q \*q" boolean \*q
+Disable or enable DRI support.
+.IP
+Default: DRI is enabled for configurations where it is supported.
+
+.PP
+The following driver
+.B Options
+are supported for the i810 and i815 chipsets:
+.TP
.BI "Option \*qCacheLines\*q \*q" integer \*q
This allows the user to change the amount of graphics memory used for
2D acceleration and video when XAA acceleration is enabled. Decreasing this
@@ -72,41 +82,6 @@ driver attempts to allocate space for at 3 screenfuls of pixmaps plus an
HD-sized XV video. The default used for a specific configuration can be found
by examining the __xservername__ log file.
.TP
-.BI "Option \*qFramebufferCompression\*q \*q" boolean \*q
-This option controls whether the framebuffer compression feature is enabled.
-If possible, the front buffer will be allocated in a tiled format and compressed
-periodically to save memory bandwidth and power.
-This option is only available on mobile chipsets.
-.IP
-Default: enabled on supported configurations.
-.TP
-.BI "Option \*qTiling\*q \*q" boolean \*q
-This option controls whether memory buffers are allocated in tiled mode. In
-most cases (especially for complex rendering), tiling dramatically improves
-performance.
-.IP
-Default: enabled.
-.TP
-.BI "Option \*qSwapbuffersWait\*q \*q" boolean \*q
-This option controls the behavior of glXSwapBuffers and glXCopySubBufferMESA
-calls by GL applications. If enabled, the calls will avoid tearing by making
-sure the display scanline is outside of the area to be copied before the copy
-occurs. If disabled, no scanline synchronization is performed, meaning tearing
-will likely occur. Note that when enabled, this option can adversely affect
-the framerate of applications that render frames at less than refresh rate.
-.IP
-Default: enabled.
-.TP
-.BI "Option \*qDRI\*q \*q" boolean \*q
-Disable or enable DRI support.
-.IP
-Default: DRI is enabled for configurations where it is supported.
-
-.PP
-The following driver
-.B Options
-are supported for the i810 and i815 chipsets:
-.TP
.BI "Option \*qDDC\*q \*q" boolean \*q
Disable or enable DDC support.
.IP
@@ -162,41 +137,22 @@ server log.
.IP
Default: Disabled
.TP
-.BI "Option \*qForceEnablePipeA\*q \*q" boolean \*q
-Force the driver to leave pipe A enabled. May be necessary in configurations
-where the BIOS accesses pipe registers during display hotswitch or lid close,
-causing a crash. If you find that your platform needs this option, please file
-a bug (see REPORTING BUGS below) including the output of 'lspci -v' and 'lspci -vn'.
-.TP
-.BI "Option \*qLVDS24Bit\*q \*q" boolean \*q
-Specify 24 bit pixel format (i.e. 8 bits per color) to be used for the
-LVDS output. Some newer LCD panels expect pixels to be formatted and
-sent as 8 bits per color channel instead of the more common 6 bits per
-color channel. Set this option to true to enable the newer format.
-Note that this concept is entirely different and independent from the
-frame buffer color depth - which is still controlled in the usual way
-within the X server. This option instead selects the physical format
-/ sequencing of the digital bits sent to the display. Setting the
-frame buffer color depth is really a matter of preference by the user,
-while setting the pixel format here is a requirement of the connected
-hardware.
+.BI "Option \*qSwapbuffersWait\*q \*q" boolean \*q
+This option controls the behavior of glXSwapBuffers and glXCopySubBufferMESA
+calls by GL applications. If enabled, the calls will avoid tearing by making
+sure the display scanline is outside of the area to be copied before the copy
+occurs. If disabled, no scanline synchronization is performed, meaning tearing
+will likely occur. Note that when enabled, this option can adversely affect
+the framerate of applications that render frames at less than refresh rate.
.IP
-Leaving this unset implies the default value of false,
-which is almost always going to be right choice. If your
-LVDS-connected display on the other hand is extremely washed out
-(e.g. white on a lighter white), trying this option might clear the
-problem.
+Default: enabled.
.TP
-.BI "Option \*qLVDSFixedMode\*q \*q" boolean \*q
-Use a fixed set of timings for the LVDS output, independent of normal
-xorg specified timings.
+.BI "Option \*qTiling\*q \*q" boolean \*q
+This option controls whether memory buffers are allocated in tiled mode. In
+most cases (especially for complex rendering), tiling dramatically improves
+performance.
.IP
-The default value if left unspecified is
-true, which is what you want for a normal LVDS-connected LCD type of
-panel. If you are not sure about this, leave it at its default, which
-allows the driver to automatically figure out the correct fixed panel
-timings. See further in the section about LVDS fixed timing for more
-information.
+Default: enabled.
.TP
.BI "Option \*qXvMC\*q \*q" boolean \*q
Enable XvMC driver. Current support MPEG2 MC on 915/945 and G33 series.
@@ -332,84 +288,6 @@ sections with these outputs for configuration. Associating Monitor sections
with each output can be helpful if you need to ignore a specific output, for
example, or statically configure an extended desktop monitor layout.
-.SH HARDWARE LVDS FIXED TIMINGS AND SCALING
-
-Following here is a discussion that should shed some light on the
-nature and reasoning behind the LVDSFixedMode option.
-
-Unlike a CRT display, an LCD has a "native" resolution corresponding
-to the actual pixel geometry. A graphics controller under all normal
-circumstances should always output that resolution (and timings) to
-the display. Anything else and the image might not fill the display,
-it might not be centered, or it might have information missing - any
-manner of strange effects can happen if an LCD panel is not fed with
-the expected resolution and timings.
-
-However there are cases where one might want to run an LCD panel at an
-effective resolution other than the native one. And for this reason,
-GPUs which drive LCD panels typically include a hardware scaler to
-match the user-configured frame buffer size to the actual size of the
-panel. Thus when one "sets" his/her 1280x1024 panel to only 1024x768,
-the GPU happily configures a 1024x768 frame buffer, but it scans the
-buffer out in such a way that the image is scaled to 1280x1024 and in
-fact sends 1280x1024 to the panel. This is normally invisible to the
-user; when a "fuzzy" LCD image is seen, scaling like this is why this
-happens.
-
-In order to make this magic work, this driver logically has to be
-configured with two sets of monitor timings - the set specified (or
-otherwise determined) as the normal xorg "mode", and the "fixed"
-timings that are actually sent to the monitor. But with xorg, it's
-only possible to specify the first user-driven set, and not the second
-fixed set. So how does the driver figure out the correct fixed panel
-timings? Normally it will attempt to detect the fixed timings, and it
-uses a number of strategies to figure this out. First it attempts to
-read EDID data from whatever is connected to the LVDS port. Failing
-that, it will check if the LVDS output is already configured (perhaps
-previously by the video BIOS) and will adopt those settings if found.
-Failing that, it will scan the video BIOS ROM, looking for an embedded
-mode table from which it can infer the proper timings. If even that
-fails, then the driver gives up, prints the message "Couldn't detect
-panel mode. Disabling panel" to the X server log, and shuts down the
-LVDS output.
-
-Under most circumstances, the detection scheme works. However there
-are cases when it can go awry. For example, if you have a panel
-without EDID support and it isn't integral to the motherboard
-(i.e. not a laptop), then odds are the driver is either not going to
-find something suitable to use or it is going to find something
-flat-out wrong, leaving a messed up display. Remember that this is
-about the fixed timings being discussed here and not the
-user-specified timings which can always be set in xorg.conf in the
-worst case. So when this process goes awry there seems to be little
-recourse. This sort of scenario can happen in some embedded
-applications.
-
-The LVDSFixedMode option is present to deal with this. This option
-normally enables the above-described detection strategy. And since it
-defaults to true, this is in fact what normally happens. However if
-the detection fails to do the right thing, the LVDSFixedMode option
-can instead be set to false, which disables all the magic. With
-LVDSFixedMode set to false, the detection steps are skipped and the
-driver proceeds without a specified fixed mode timing. This then
-causes the hardware scaler to be disabled, and the actual timings then
-used fall back to those normally configured via the usual xorg
-mechanisms.
-
-Having LVDSFixedMode set to false means that whatever is used for the
-monitor's mode (e.g. a modeline setting) is precisely what is sent to
-the device connected to the LVDS port. This also means that the user
-now has to determine the correct mode to use - but it's really no
-different than the work for correctly configuring an old-school CRT
-anyway, and the alternative if detection fails will be a useless
-display.
-
-In short, leave LVDSFixedMode alone (thus set to true) and normal
-fixed mode detection will take place, which in most cases is exactly
-what is needed. Set LVDSFixedMode to false and then the user has full
-control over the resolution and timings sent to the LVDS-connected
-device, through the usual means in xorg.
-
.SH MULTIHEAD CONFIGURATIONS
The number of independent outputs is dictated by the number of CRTCs