INDEXING with Cast / Convert function in SQL Server (Enhancing Query Performance)

Posted on 2013-12-12
Medium Priority
Last Modified: 2014-02-10
I have several SQL Server tables which are dropped and reloaded through text files on a regular basis. The result is dozens of tables which all have the data type NVARCHAR(255). In the past this has worked fine since I can cast the data type to whatever I want later without having to worry about type conversion issue on the daily job. This practice has now created an issue since I would like to create an index on a datetime derived field that is stored as a multiple text fields. Is it possible to create an index on a value that is derived from the table? In my case I would like something similar to:

ON [dbo].[Position_Dta]
([Position Number - Position Dta],
DATEADD(Minute, cast([Effective Date Sequence # - Position Dta] as int), [Effective Date - Position Dta]),
cast(tbl_Secondary.[Action Date - Position Dta] as date))

Obviously the above syntax does not work but hopefully that helps explain what I am looking for. Essentially the query I have in the example is one of dozens of subqueries with the same issue and the result is queries that take 4 hours to run. I suppose I would create a temp table with the correct data types and index the temp table but I was really hoping for a solution that would not cause me to have to update a lot of functioning code (even though it is super slow it works). Additionally I am not the only person who accesses these tables so I am really trying to avoid modifying the source tables.

Attached is a copy of a typical example where I am trying to create a more effective index.  I am open to any ideas that will increase performance and the less modification to the existing code the better! I appreciate the help experts!
Question by:HRISTeam
Welcome to Experts Exchange

Add your voice to the tech community where 5M+ people just like you are talking about what matters.

  • Help others & share knowledge
  • Earn cash & points
  • Learn & ask questions
  • 2
  • 2
  • 2
  • +1
LVL 20

Expert Comment

ID: 39714917
You can add computed columns to the tables and then make the index on those.

Here is a topic about computed columns: http://technet.microsoft.com/en-us/library/ms191250(v=sql.105).aspx
And here about indexes on them: http://technet.microsoft.com/en-us/library/ms189292.aspx
LVL 42

Expert Comment

ID: 39715018
To create an index on TWO tables as you showed in the question is not good solution probably...

You should try to create index on  [Position Number - Position Dta]  column on both tables and see what happens. Then you may add indexes on [Effective Date Sequence # - Position Dta] and [Effective Date - Position Dta] which will optimize your subquery.

The "less than comparison part" of the WHERE condition isn't easily optimizable and the index on computed column could be a good start BUT I am skeptic...

Author Comment

ID: 39715181
Sorry, the screen shot is not 100% accurate. All the data is coming from the same table. So the table alias tbl_Primary and tbl_Secondary are actually the exact same table.
Get MongoDB database support online, now!

At Percona’s web store you can order your MongoDB database support needs in minutes. No hassles, no fuss, just pick and click. Pay online with a credit card. Handle your MongoDB database support now!

LVL 42

Expert Comment

ID: 39715322
In such case I would rather create a temporary table which will be used in place of the first subselect. This subselect must slow down the whole query because it calculates minimum from unoptimized data.

So create a table where each  [Position Number - Position Dta]  value has one corresponding value calculated as MIN(DateAdd(MINUTE, ...  WHERE ... etc.
and join this table by LEFT JOIN to your Position_dta table. This should optimize the subselect part.

If the result will be still slow then you may optimize the correlated subquery in the WHERE part.
LVL 20

Expert Comment

ID: 39715333
What about the computed columns I wrote about?

Author Comment

ID: 39715549
The computed columns looks promising but it would seemingly require me to rewrite much of my code. I am going to try a small test with the concept and test the performance gains.
LVL 66

Accepted Solution

Jim Horn earned 1500 total points
ID: 39716774
>several SQL Server tables which are dropped and reloaded through text files on a regular basis. The result is dozens of tables which all have the data type NVARCHAR(255).

That's an excellent start, but a vastly better idea would be to treat all of these nvarchar(255) tables as 'staging' tables, whose purpose in life is to ensure that all rows in your text file(s) make it into SQL Server.

Then, do some validations to make sure dates are dates, numbers are numbers, etc. and then INSERT rows from those staging tables into destination tables where the date columns have a date data type, numeric columns have numeric data types, etc.   If you have time you can also build a way to gracefully handle rows where a value that should have been a date was not.

Then index/relate the destination tables, and use them as the source of data, and not the nvarchar(255) tables, in everything that needs the data.

The benefits of this approach will include you don't have to CAST() all the time, which means indexes will be used.

Featured Post

Get real performance insights from real users

Key features:
- Total Pages Views and Load times
- Top Pages Viewed and Load Times
- Real Time Site Page Build Performance
- Users’ Browser and Platform Performance
- Geographic User Breakdown
- And more

Question has a verified solution.

If you are experiencing a similar issue, please ask a related question

In the first part of this tutorial we will cover the prerequisites for installing SQL Server vNext on Linux.
In this blog post, we’ll look at how using thread_statistics can cause high memory usage.
Viewers will learn how to use the SELECT statement in SQL and will be exposed to the many uses the SELECT statement has.
This is a high-level webinar that covers the history of enterprise open source database use. It addresses both the advantages companies see in using open source database technologies, as well as the fears and reservations they might have. In this…
Suggested Courses

777 members asked questions and received personalized solutions in the past 7 days.

Join the community of 500,000 technology professionals and ask your questions.

Join & Ask a Question