Skip to main content

CONDG Entity Framework Presentation Follow-ups

Thanks to everyone who came out last night (73...a great showing considering the start of Memorial Day weekend and the Indiana Jones IV opening) to CONDG for my Entity Framework presentation. We had some great questions I didn't know the answers to which I'll investigate and respond to soon.
  • Is there a caching mechanism for ObjectContext? Thinking of this in a web (ASP.NET), disconnected between request/response environment.
  • Is the conceptual, mapping, and schema XML loaded up all at once or is the XML representing entities loaded up individually upon instantiation / use?
  • The EF wizard interrogates the physical data store and produces the 1:1, Type per Table initial EDM. Is it possible to model the EDM first leveraging it to generate the physical schema?
  • Referring to this performance comparison between the traditional Sql Client, Entity SQL, and LINQ to Entities, what's the break-down of time consumed?
    • The ADO.NET team posted a follow-up with the break-down here. Granted, it's beta3 bits but there is some fascinating insight into the innards of EF on this post. Here is the final installment of posts on performance.
  • (I responded to this one but want to delve in further...) Are transactions built into EF and the ObjectContext? Is the ObjectContext transaction-aware?
  • Is the EF and ObjectContext interface based? Meaning, can I leverage a mock testing framework to unit test my EDM?
  • Clarification on lazy loading in EF. Thanks to Kevin Sprague for correcting me on how NHibernate handles both lazy and eager loading.
    • I can't state it any better than the EF FAQ. Lazy loading is the default but eager loading is supported on a query-by-query basis. One needs to explicitly call Load (or Include) to pull related entities into the object graph. Related postings on lazy loading here, here and here.

Comments

Popular posts from this blog

TFS Reports Out of Date

You may have noticed it takes a while for Team Foundation Server (TFS) reports to reflect changes you've made to work items or builds. Let me guess...about an hour, right? Out of the box, TFS is set to refresh the data warehouse from its transactional store every 60 minutes. How do I change the frequency of the data warehouse refresh? Browse to the TFS Controller Web Service on your TFS application tier server within IE at: http://localhost:8080/Warehouse/v1.0/warehousecontroller.asmx Select the ChangeSetting option Enter RunIntervalSeconds for the settingId and the desired number of seconds for newValue (300 for 5 minutes...5*60) Select Invoke How do I force a data warehouse refresh? Two methods here: either via the above web service or using SQL Server Management Studio. Via the web service: Browse to the TFS Controller Web Service within IE at: http://localhost:8080/Warehouse/v1.0/warehousecontroller.asmx Select the Run option Click Invo...

Rollback a Ooops in TFS with TFPT Rollback

Rhut roe, Raggie. You just checked in a merge operation affecting 100's of files in TFS against the wrong branch. Ooops. Well, you can simply roll it back, right? Select the folder in Source Control Explorer and...hey, where's the Rollback? Rollback isn't supported in TFS natively. However, it is supported within the Power Tools leveraging the command-line TFPT.exe utility. It's fairly straightforward to revert back to a previous version--with one caveot. First, download and install the Team Foundation Power Tools 2008 on your workstation. Before proceeding, let's create a workspace dedicated to the rollback. To "true up" the workspace, the rollback operation will peform a Get Latest for every file in your current workspace. This can consume hours (and many GB) with a broad workspace mapping. To work around this, I create a temporary workspace targeted at just the area of source I need to roll back. So let's drill down on our scenario... I'm worki...

VSTS 2008 Data-Driven Web Test

During a client demo this afternoon, I mind-blanked on creating a data-driven web test. Sure enough, I stepped out of the client offices and figured it out. Super. But, I think it's counterintuitive enough to outline in a blog post (...trying to make myself feel better ;-). Somewhat in my defense, I was thinking about data-driven unit tests. These, IMO, are easy to implement. Data-driven web tests are a bit more effort and challenging. Create a web test: Right-mouse on your test project selecting Add >> Web Test. Use the browser recorder to capture your web test scenario. Let's assume you're submitting an age and credit rating to an insurance rate calculator...similar to the screen capture below: Create a table or CSV file (comma-delimited...just use Excel and save as CSV) with relevant headings for your inputs and expected values: From Solution Explorer, double-click on your web test. In the main window, right-mouse on your web test selecting Add Data Sou...