Solved

dg4odbc vs dg4msql

Posted on 2012-12-24
4
1,297 Views
Last Modified: 2013-01-08
I was using a direct database link to an mssql using dg4odbc..

because of performance issues i started looking for alternatives and installed transparent gateway with dg4msql

performance is unchanged. there is no difference in querying time

eg

select * from rmordp@"dblink" where ORDITMNUM = 89129 and ordordqty = 48 and ordseqnum = 24 and ordprcgrp = 8242;

takes 9.351 seconds over the database link.

if i run that same query from the mssql server, it takes less than half a second..


can someone tell me what I am doing wrong or how i should go about getting remote data via oracle in a timely fashion?
0
Comment
Question by:futureDBA
[X]
Welcome to Experts Exchange

Add your voice to the tech community where 5M+ people just like you are talking about what matters.

  • Help others & share knowledge
  • Earn cash & points
  • Learn & ask questions
  • 2
4 Comments
 
LVL 12

Expert Comment

by:Praveen Kumar Chandrashekatr
ID: 38718910
when you are querying remotely it also depends on the network bandwidth which you have to communicate between the servers.

please check what is your bandwidth.
0
 

Author Comment

by:futureDBA
ID: 38718928
its not a bandwidth issue, when i use TOAD for SQL Server to query the sql server directly (without having oracle as a middle man)

my execution times are pretty fast, i am doing this from the same machine that I have sql developer on
0
 
LVL 77

Accepted Solution

by:
slightwv (䄆 Netminder) earned 500 total points
ID: 38721954
I'm not a Toad user but I've seen posts that Toad buffers data.  How are you timing the Toad runs?  If it is from the first data you see, that could be the issue.  You are seeing the first few rows brought back not all of them.
0
 

Author Closing Comment

by:futureDBA
ID: 38755330
figured it out, thanks
0

Featured Post

What Is Transaction Monitoring and who needs it?

Synthetic Transaction Monitoring that you need for the day to day, which ensures your business website keeps running optimally, and that there is no downtime to impact your customer experience.

Question has a verified solution.

If you are experiencing a similar issue, please ask a related question

When table data gets too large to manage or queries take too long to execute the solution is often to buy bigger hardware or assign more CPUs and memory resources to the machine to solve the problem. However, the best, cheapest and most effective so…
In this blog post, we’ll look at how ClickHouse performs in a general analytical workload using the star schema benchmark test.
Video by: Steve
Using examples as well as descriptions, step through each of the common simple join types, explaining differences in syntax, differences in expected outputs and showing how the queries run along with the actual outputs based upon a simple set of dem…
This video shows syntax for various backup options while discussing how the different basic backup types work.  It explains how to take full backups, incremental level 0 backups, incremental level 1 backups in both differential and cumulative mode a…

687 members asked questions and received personalized solutions in the past 7 days.

Join the community of 500,000 technology professionals and ask your questions.

Join & Ask a Question