• Status: Solved
  • Priority: Medium
  • Security: Public
  • Views: 232
  • Last Modified:

complete cache of DB

hey guys i've got a big report i need to run and the database is on the network. that means that the calls always takes very long and the multiple trips are a penalty. how do i create a local cache of the relevant tables when i'm running this mammoth report at the end of every week?

i'm thinking 3 ways

1) Access's synchronisation
2) DIY synchronisation
3) just copy the whole database to local machine, change the links of the table to the local database, run the report, link the tables back to the network database, then delete the local database

what do yall think guys? looking forward to your guidance and sharing guys = )
0
developingprogrammer
Asked:
developingprogrammer
  • 4
  • 4
  • 4
  • +1
2 Solutions
 
Rey Obrero (Capricorn1)Commented:
if the records are not going to change or the db is not in used at the end of the week,
the best way is option 3.
0
 
DatabaseMX (Joe Anderson - Microsoft Access MVP)Database ArchitectCommented:
I vote for # 3 also.
0
 
developingprogrammerAuthor Commented:
Cool, thanks guys, but if I didn't suggest number 3, how would y'all have done it? I think your methods should definitely be better than mine with all the experience y'all have! = )
0
Get 10% Off Your First Squarespace Website

Ready to showcase your work, publish content or promote your business online? With Squarespace’s award-winning templates and 24/7 customer service, getting started is simple. Head to Squarespace.com and use offer code ‘EXPERTS’ to get 10% off your first purchase.

 
DatabaseMX (Joe Anderson - Microsoft Access MVP)Database ArchitectCommented:
Yes  I would have suggested it, because that is exactly what I do (in an automated fashion) on one of our reporting KPI databases.

:-)
0
 
Dale FyeCommented:
The 4th way would be to use temporary tables, creating only the necessary tables locally (see my article on temp tables).
0
 
developingprogrammerAuthor Commented:
Hrmm nice guys. I super like the ideas!! = ))

Hrmm may be a silly question, but if a db has one table with 100,000 records, is it faster in terms of "localising" it by

1) copying the whole database
2) reading the linked table and copying to temp table?

One is wholesale one is pinpoint (so to speak if there were more tables we want to ignore)
0
 
Dale FyeCommented:
I generally just copy exactly what I need into the temp table, not scientific, just the way I do things.
0
 
developingprogrammerAuthor Commented:
Yup I agree. Looks more elegant that way as well. What do you think MX? And also from a performance angle MX? = ) thanks guys!!
0
 
DatabaseMX (Joe Anderson - Microsoft Access MVP)Database ArchitectCommented:
What I do in the case I posted is ... what fyed does. I load a local temp table with records from approx 12 different dbs on the server, then run reports from that. The local table - toward the end of the year, can  have well over a million records. It takes less than a minute to load the temp table over our network.
0
 
Dale FyeCommented:
Joe,

That's because you have a blazingly fast network!

;-)
0
 
DatabaseMX (Joe Anderson - Microsoft Access MVP)Database ArchitectCommented:
Yep ... But it's **not** the only one in the World, contrary to popular belief :-)
0
 
Dale FyeCommented:
I'm just jealous.
0
 
developingprogrammerAuthor Commented:
Superb!!! Thanks guys!! = ))
0
Question has a verified solution.

Are you are experiencing a similar issue? Get a personalized answer when you ask a related question.

Have a better answer? Share it in a comment.

Join & Write a Comment

Featured Post

Easily Design & Build Your Next Website

Squarespace’s all-in-one platform gives you everything you need to express yourself creatively online, whether it is with a domain, website, or online store. Get started with your free trial today, and when ready, take 10% off your first purchase with offer code 'EXPERTS'.

  • 4
  • 4
  • 4
  • +1
Tackle projects and never again get stuck behind a technical roadblock.
Join Now