2 LCD VGA Monitors connected to one DVI video port on an adapter, resolution changes on restart

I added a 2nd monitor to my PC so I could have a 2nd workstation;  this is not a dual monitor setup per se.  I want both monitors to display the same thing.  The video card I have is a ATI HD 3600 with only one DVI port.  I'm using a DVI to VGA adapter connector and then a Y splittle cable, with a VGA cable going to each monitor.  One monitor is a Dell 22" Widescreen 2208WFP Ultrasharp, the other monitor is new and it's a ASUS 22" VW224U.  Both are native resolution 1680 x 1050, using 32 bit color.  I'm running Vista Premium 32 bit.

The monitors both worked great together at the same time at 1680 x 1050, but then when I restarted the PC when Vista finished booting up both monitors were now a much lower resolution and consequently the text, icons, etc were larger.  I went into the adapter settings and it was maxed out (I think it was 1024 x 768) and I could not move the slider anymore to the right to increase the resolution.  The monitor driver being used is the generic pnp LCD driver, not the specific driver for either monitor.

My fix was to disconnect one of the monitors, reboot, then go into the adapter setting and now I could set it to 1680 x 1050.  Then reconnect the other monitor and now both monitors displayed at 1680 x 1050.  Works fine until I restart again with both monitors connected.

Why is this and how I can I permanently fix it.  I'm sure it has something to do with autodetecting monitors?  

Thanks
Joe
JoeBoydAsked:
Who is Participating?

[Product update] Infrastructure Analysis Tool is now available with Business Accounts.Learn More

x
I wear a lot of hats...

"The solutions and answers provided on Experts Exchange have been extremely helpful to me over the last few years. I wear a lot of hats - Developer, Database Administrator, Help Desk, etc., so I know a lot of things but not a lot about one thing. Experts Exchange gives me answers from people who do know a lot about one thing, in a easy to use platform." -Todd S.

CallandorCommented:
It does have to do with autodetecting monitors - when you split a signal, the voltage drops in half, and the monitor and cards don't detect things properly.
0
CallandorCommented:
In order to fix it, you need to get a video distribution amp, which will split the signal, but amplify it back to the original level, such as a Kramer VP-200.
0

Experts Exchange Solution brought to you by

Your issues matter to us.

Facing a tech roadblock? Get the help and guidance you need from experienced professionals who care. Ask your question anytime, anywhere, with no hassle.

Start your 7-day free trial
JoeBoydAuthor Commented:
Thanks.  Not only did you give me a reason but also a solution.  That video amp looks pretty good and it's not super expensive.  Appreciate it!
0
It's more than this solution.Get answers and train to solve all your tech problems - anytime, anywhere.Try it for free Edge Out The Competitionfor your dream job with proven skills and certifications.Get started today Stand Outas the employee with proven skills.Start learning today for free Move Your Career Forwardwith certification training in the latest technologies.Start your trial today
Displays / Monitors

From novice to tech pro — start learning today.