Saturday, December 4, 2010
XTIVIA Colorado Springs Holiday Party
http://picasaweb.google.com/drobincorvette/20101204#
We had some great food and even home made beer that was really tasty! Thank you to all the Xtivia employees in Colorado Springs!
We look forward to seeing pictures and stories from the other offices.
Happy Holidays!
Friday, December 3, 2010
Larry Vs. HP
http://online.wsj.com/article/SB10001424052748703377504575651260196782120.html?mod=WSJ_business_whatsNews
The Data Quality Challenge
According to a study published by The Data Warehousing Institute (TDWI) entitled Taking Data Quality to the Enterprise through Data Governance, some issues are primarily technical in nature, such as the extra time required for reconciling data (85%) or delays in deploying new systems (52%). Other problems are closer to business issues, such as customer dissatisfaction (69%), compliance problems (39%) and revenue loss (35%). Poor-quality data can also cause problems with costs (67%) and credibility (77%).
Another recurring statement that we come across frequently is : Only less than 5% of our data is “bad”.
If you consider a fact table with 10 million rows, then about 500K rows are bad – which is a huge problem for analytics, operations or anything else you may want to do with the data!!
Even if you choose to ignore all the statistics and studies, it comes down to just a few question –
- How do we get the information that we need and where do we get it from?
- Can we trust this information?
- What does it mean and how can we get this information in the format we need?
Data Quality Definition: Commonly known, data is deemed to be of high quality if it correctly represents the real-world construct to which they refer.
The implementation of Data Quality processes and procedures can dramatically affect the quality and therefore usability of the information in your Data Warehouse.
The Goal: The reason for building a Data Warehouse is to provide consolidating data from multiple sources, both internal and external, into an enterprise solution. This consolidation involves ensuring data consistency, integrity and quality. The data warehouse is the medium to provide enriched analytical data to empower better decision making. The goal is not to copy the same old bad information from multiple locations and provide little or no benefit to the enterprise.
We must consider two key factors to achieve this goal.
• Data Profiling identifies the problems. It provides snap-shots of a company‘s data quality and measures the evolution of data quality over time.
• Data Cleansing corrects, or ―cleanses incomplete or inconsistent data by crosschecking against other databases and reference data. It also enriches data by providing value-add information that actually improves the quality and usefulness of existing data.
I will post more on the two factors soon...
DB2 table reorg performance - Problem Solved
The row size of each table was near the 4k pagesize limit. So as a test, I moved these tables from a 4k pagesize tablespace to an 8k pagesize tablespace. This process included creating the new 8k bufferpool and tablespace, creating the clone tables with a slightly different name in the new tablespace, loading from cursor from the old table to the new table, dropping all constraints on the old table, renaming the old table to *_OLD, then renaming the new table to the correct name and recreating all the indexes, constraints and views. Finally a runstats was required on the tables and all indexes.
The INPLACE reorg now takes 10 minutes! This made the table reorgs simple to run at any time because they are never offline. The reorg helps performance and helps to limit disk space usage.
Tuesday, November 30, 2010
Informix Replication
Here are my observations and recommendations.
1.) The session thread was doing quite a bit of sleeping. This normally
occurs when control is handed off to another thread. The session information did not show any other sub-threads for this session. It appears that the thread was waiting for the commit response from the secondary.
Currently, the two servers are configured for synchronous mode. That means that the transaction will not commit on the primary until it gets the commit response back from the secondary. With any network latency, this can be quite slow during bulk processes. It is more apt to be up to date though.
2.) I saw the session wait for a logical log buffer quite often. Not
near as it was sleeping (which was 90% of the time), but it happened often enough to be looked at.
My recommendations:
Change DRINTERVAL from -1 to 0. This will disable synchronous mode with HDR and will go to asynchronous. This is much better for performance as the primary will no longer have to wait for the secondary server to acknowledge to transaction before moving on. Checkpoints are still done synchronously.
Increase LOGBUFF setting from 32 to 64. This will increase the logical log buffers and prevent processes from waiting for them as often as they did.
New SQL Server Customer
This customer chose Xtivia as the best of breed support organization to assist them as they require 24x7 availability as well as needing assistance with implementing and managing replication.
Monday, November 29, 2010
Xtivia is getting ready for the Holidays
This Friday we have the Xtivia Colorado Potluck Dinner and as I understand it, Santa himself is coming from the North Pole.
Practice Manager Tony Salerno visits Colorado Springs
I think he is staying through Friday night so he will be able to join us for the potluck dinner on Friday.