Lost sliding windows w/ DVI Syncmaster monitor
mrmazda at earthlink.net
Thu Apr 2 01:58:34 UTC 2020
Paul Rogers composed on 2020-04-01 13:02 (UTC-0700):
>> I don't think it has anything to do with the Samsung or the Dell, but
>> rather with
>> however you are trying to configure. How exactly are you configuring?
> See attachment. Or are you asking for the build options for the server?
xorgconfd.tgz had what I was looking for. :-)
> >> Please show us your GPU specification, thus:
> >> # inxi -Gxx
> Device-1: NVIDIA GK208 [GeForce GT 635] driver: N/A bus ID: 01:00.0
> chip ID: 10de:1280
> Display: server: X.Org 1.18.4 driver: vesa resolution: 1024x768~N/A
> OpenGL: renderer: Gallium 0.4 on llvmpipe (LLVM 3.8 256 bits)
> v: 3.3 Mesa 12.0.1 compat-v: 3.0 direct render: Yes
Rather than a configuration of the type you're familiar with, I'll bet the
newer way will work, if not with your custom installed system, at least with
a common recent live media distro as a means of proving. Modern Xorg automagic
should get the whole job done, up until the point of panning configuration, if
instead of VESA, using the newer technology default DDX driver provided by the
XServer rather than in a separate package.
Start by eliminating 20-vesa.conf, 40-samsung.conf and 90-screen.conf from
xorg.conf.d/. Let X startup in the Samsung's native 1280x1024 mode using the
default "modesetting" DDX. Then try as follows:
# xrandr | head
Screen 0: minimum 320 x 200, current 1280 x 1024, maximum 8192 x 8192
DVI-I-1 connected 1280x1024+0+0 (normal left inverted right x axis y axis) 338mm x 270mm
1280x1024 60.02*+ 75.02
1024x768 60.04 75.03 70.07 60.00
# xrandr --fb 1280x1280 --output DVI-I-1 --mode 1024x768 --panning 1280x1280
# xrandr | head
Screen 0: minimum 320 x 200, current 1280 x 1280, maximum 8192 x 8192
DVI-I-1 connected 1280x1280+0+0 (normal left inverted right x axis y axis) 338mm x 270mm panning 1280x1280+0+0
1280x1024 60.02 + 75.02
1024x768 60.04 75.03* 70.07 60.00
# inxi -SGxx
System: Host: g5eas Kernel: 4.4.180-102-default x86_64 bits: 64 compiler: gcc v: 4.8.5 Desktop: Trinity R14.0.6
tk: Qt 3.5.0 wm: Twin dm: startx Distro: openSUSE Leap 42.3
Graphics: Device-1: XGI Z7/Z9 vendor: Gigabyte driver: xgifb v: kernel bus ID: 0a:03.0 chip ID: 18ca:0020
Device-2: NVIDIA G98 [GeForce 8400 GS Rev. 2] driver: nouveau v: kernel bus ID: 0b:00.0 chip ID: 10de:06e4
Display: server: X.Org 1.18.3 driver: modesetting tty: N/A
OpenGL: renderer: Gallium 0.4 on llvmpipe (LLVM 3.8 128 bits) v: 3.3 Mesa 17.0.5 compat-v: 3.0 direct render: Yes
I used nearly the same kernel and XServer as yours, as well as an NVidia GPU
not terribly newer, and a Dell LCD. I do have a 1280x1024 Samsung SyncMaster
914V, but it's upstairs, and has no DVI port, so more bother to test with.
Ignore what you see about the XGI GPU. This Gigabyte's BIOS doesn't understand
how to do it's job correctly when a PCI GeForce is installed.
If it works as I expect, you can put xrandr in whatever script you start
X with, or another. If it fails, I'd like to see the resulting Xorg.0.logs
from both your custom build and any recent live distro that depends fully
on automagic X configuration.
Be sure you do not have the Xorg nouveau DDX installed, as it will usurp the
modesetting DDX, unless you perform additional manual configuration.
Evolution as taught in public schools is religion, not science.
Team OS/2 ** Reg. Linux User #211409 ** a11y rocks!
Felix Miata *** http://fm.no-ip.com/
More information about the xorg