xf86-video-intel: man/intel.man

Jesse Barnes jbarnes at kemper.freedesktop.org
Wed Dec 2 14:45:07 PST 2009


 man/intel.man |  168 +++++++---------------------------------------------------
 1 file changed, 23 insertions(+), 145 deletions(-)

New commits:
commit c439207ec0cc16d7d9f523598fcdebf7cec893e2
Author: Jesse Barnes <jbarnes at virtuousgeek.org>
Date:   Wed Dec 2 14:43:17 2009 -0800

    Update man page to reflect currently available options
    
    Many have been removed or are obsolete now that UMS is gone.  And some
    are only available on i810/i815 or i830+, so move them to the
    appropriate section.
    
    Signed-off-by: Jesse Barnes <jbarnes at virtuousgeek.org>

diff --git a/man/intel.man b/man/intel.man
index 1673702..9344d28 100644
--- a/man/intel.man
+++ b/man/intel.man
@@ -61,6 +61,16 @@ This sets the default pixel value for the YUV video overlay key.
 .IP
 Default: undefined.
 .TP
+.BI "Option \*qDRI\*q \*q" boolean \*q
+Disable or enable DRI support.
+.IP
+Default: DRI is enabled for configurations where it is supported.
+
+.PP
+The following driver
+.B Options
+are supported for the i810 and i815 chipsets:
+.TP
 .BI "Option \*qCacheLines\*q \*q" integer \*q
 This allows the user to change the amount of graphics memory used for
 2D acceleration and video when XAA acceleration is enabled.  Decreasing this
@@ -72,41 +82,6 @@ driver attempts to allocate space for at 3 screenfuls of pixmaps plus an
 HD-sized XV video.  The default used for a specific configuration can be found
 by examining the __xservername__ log file.
 .TP
-.BI "Option \*qFramebufferCompression\*q \*q" boolean \*q
-This option controls whether the framebuffer compression feature is enabled.
-If possible, the front buffer will be allocated in a tiled format and compressed
-periodically to save memory bandwidth and power.
-This option is only available on mobile chipsets.
-.IP
-Default: enabled on supported configurations.
-.TP
-.BI "Option \*qTiling\*q \*q" boolean \*q
-This option controls whether memory buffers are allocated in tiled mode.  In
-most cases (especially for complex rendering), tiling dramatically improves
-performance.
-.IP
-Default: enabled.
-.TP
-.BI "Option \*qSwapbuffersWait\*q \*q" boolean \*q
-This option controls the behavior of glXSwapBuffers and glXCopySubBufferMESA
-calls by GL applications.  If enabled, the calls will avoid tearing by making
-sure the display scanline is outside of the area to be copied before the copy
-occurs.  If disabled, no scanline synchronization is performed, meaning tearing
-will likely occur.  Note that when enabled, this option can adversely affect
-the framerate of applications that render frames at less than refresh rate.
-.IP
-Default: enabled.
-.TP
-.BI "Option \*qDRI\*q \*q" boolean \*q
-Disable or enable DRI support.
-.IP
-Default: DRI is enabled for configurations where it is supported.
-
-.PP
-The following driver
-.B Options
-are supported for the i810 and i815 chipsets:
-.TP
 .BI "Option \*qDDC\*q \*q" boolean \*q
 Disable or enable DDC support.
 .IP
@@ -162,41 +137,22 @@ server log.
 .IP
 Default: Disabled
 .TP
-.BI "Option \*qForceEnablePipeA\*q \*q" boolean \*q
-Force the driver to leave pipe A enabled.  May be necessary in configurations
-where the BIOS accesses pipe registers during display hotswitch or lid close,
-causing a crash.  If you find that your platform needs this option, please file
-a bug (see REPORTING BUGS below) including the output of 'lspci -v' and 'lspci -vn'.
-.TP
-.BI "Option \*qLVDS24Bit\*q \*q" boolean \*q
-Specify 24 bit pixel format (i.e. 8 bits per color) to be used for the
-LVDS output.  Some newer LCD panels expect pixels to be formatted and
-sent as 8 bits per color channel instead of the more common 6 bits per
-color channel.  Set this option to true to enable the newer format.
-Note that this concept is entirely different and independent from the
-frame buffer color depth - which is still controlled in the usual way
-within the X server.  This option instead selects the physical format
-/ sequencing of the digital bits sent to the display.  Setting the
-frame buffer color depth is really a matter of preference by the user,
-while setting the pixel format here is a requirement of the connected
-hardware.
+.BI "Option \*qSwapbuffersWait\*q \*q" boolean \*q
+This option controls the behavior of glXSwapBuffers and glXCopySubBufferMESA
+calls by GL applications.  If enabled, the calls will avoid tearing by making
+sure the display scanline is outside of the area to be copied before the copy
+occurs.  If disabled, no scanline synchronization is performed, meaning tearing
+will likely occur.  Note that when enabled, this option can adversely affect
+the framerate of applications that render frames at less than refresh rate.
 .IP
-Leaving this unset implies the default value of false,
-which is almost always going to be right choice.  If your
-LVDS-connected display on the other hand is extremely washed out
-(e.g. white on a lighter white), trying this option might clear the
-problem.
+Default: enabled.
 .TP
-.BI "Option \*qLVDSFixedMode\*q \*q" boolean \*q
-Use a fixed set of timings for the LVDS output, independent of normal
-xorg specified timings.
+.BI "Option \*qTiling\*q \*q" boolean \*q
+This option controls whether memory buffers are allocated in tiled mode.  In
+most cases (especially for complex rendering), tiling dramatically improves
+performance.
 .IP
-The default value if left unspecified is
-true, which is what you want for a normal LVDS-connected LCD type of
-panel.  If you are not sure about this, leave it at its default, which
-allows the driver to automatically figure out the correct fixed panel
-timings.  See further in the section about LVDS fixed timing for more
-information.
+Default: enabled.
 .TP
 .BI "Option \*qXvMC\*q \*q" boolean \*q
 Enable XvMC driver. Current support MPEG2 MC on 915/945 and G33 series.
@@ -332,84 +288,6 @@ sections with these outputs for configuration.  Associating Monitor sections
 with each output can be helpful if you need to ignore a specific output, for
 example, or statically configure an extended desktop monitor layout.
 
-.SH HARDWARE LVDS FIXED TIMINGS AND SCALING
-
-Following here is a discussion that should shed some light on the
-nature and reasoning behind the LVDSFixedMode option.
-
-Unlike a CRT display, an LCD has a "native" resolution corresponding
-to the actual pixel geometry.  A graphics controller under all normal
-circumstances should always output that resolution (and timings) to
-the display.  Anything else and the image might not fill the display,
-it might not be centered, or it might have information missing - any
-manner of strange effects can happen if an LCD panel is not fed with
-the expected resolution and timings.
-
-However there are cases where one might want to run an LCD panel at an
-effective resolution other than the native one.  And for this reason,
-GPUs which drive LCD panels typically include a hardware scaler to
-match the user-configured frame buffer size to the actual size of the
-panel.  Thus when one "sets" his/her 1280x1024 panel to only 1024x768,
-the GPU happily configures a 1024x768 frame buffer, but it scans the
-buffer out in such a way that the image is scaled to 1280x1024 and in
-fact sends 1280x1024 to the panel.  This is normally invisible to the
-user; when a "fuzzy" LCD image is seen, scaling like this is why this
-happens.
-
-In order to make this magic work, this driver logically has to be
-configured with two sets of monitor timings - the set specified (or
-otherwise determined) as the normal xorg "mode", and the "fixed"
-timings that are actually sent to the monitor.  But with xorg, it's
-only possible to specify the first user-driven set, and not the second
-fixed set.  So how does the driver figure out the correct fixed panel
-timings?  Normally it will attempt to detect the fixed timings, and it
-uses a number of strategies to figure this out.  First it attempts to
-read EDID data from whatever is connected to the LVDS port.  Failing
-that, it will check if the LVDS output is already configured (perhaps
-previously by the video BIOS) and will adopt those settings if found.
-Failing that, it will scan the video BIOS ROM, looking for an embedded
-mode table from which it can infer the proper timings.  If even that
-fails, then the driver gives up, prints the message "Couldn't detect
-panel mode.  Disabling panel" to the X server log, and shuts down the
-LVDS output.
-
-Under most circumstances, the detection scheme works.  However there
-are cases when it can go awry.  For example, if you have a panel
-without EDID support and it isn't integral to the motherboard
-(i.e. not a laptop), then odds are the driver is either not going to
-find something suitable to use or it is going to find something
-flat-out wrong, leaving a messed up display.  Remember that this is
-about the fixed timings being discussed here and not the
-user-specified timings which can always be set in xorg.conf in the
-worst case.  So when this process goes awry there seems to be little
-recourse.  This sort of scenario can happen in some embedded
-applications.
-
-The LVDSFixedMode option is present to deal with this.  This option
-normally enables the above-described detection strategy.  And since it
-defaults to true, this is in fact what normally happens.  However if
-the detection fails to do the right thing, the LVDSFixedMode option
-can instead be set to false, which disables all the magic.  With
-LVDSFixedMode set to false, the detection steps are skipped and the
-driver proceeds without a specified fixed mode timing.  This then
-causes the hardware scaler to be disabled, and the actual timings then
-used fall back to those normally configured via the usual xorg
-mechanisms.
-
-Having LVDSFixedMode set to false means that whatever is used for the
-monitor's mode (e.g. a modeline setting) is precisely what is sent to
-the device connected to the LVDS port.  This also means that the user
-now has to determine the correct mode to use - but it's really no
-different than the work for correctly configuring an old-school CRT
-anyway, and the alternative if detection fails will be a useless
-display.
-
-In short, leave LVDSFixedMode alone (thus set to true) and normal
-fixed mode detection will take place, which in most cases is exactly
-what is needed.  Set LVDSFixedMode to false and then the user has full
-control over the resolution and timings sent to the LVDS-connected
-device, through the usual means in xorg.
-
 .SH MULTIHEAD CONFIGURATIONS
 
 The number of independent outputs is dictated by the number of CRTCs


More information about the xorg-commit mailing list