Agile Data Warehouse Design for Big Data

21 Big Data Master Data Management Best Practices
21 Big Data Master Data Management Best Practices

On Nov 14th 2013 Big Data Analytics, Discovery & Visualization meetup hosted “Agile Data Warehouse Design for Big Data” by Jim Stagnitto & John Di Pietro from A2C.

Here’s the synopsis:


Jim Stagnitto and John DiPietro of consulting firm a2c) will discuss Agile Data Warehouse Design – a step-by-step method for data warehousing / business intelligence (DW/BI) professionals to better collect and translate business intelligence requirements into successful dimensional data warehouse designs.


The method utilizes BEAM✲ (Business Event Analysis and Modeling) – an agile approach to dimensional data modeling that can be used throughout analysis and design to improve productivity and communication between DW designers and BI stakeholders. BEAM✲ builds upon the body of mature “best practice” dimensional DW design techniques, and collects “just enough” non-technical business process information from BI stakeholders to allow the modeler to slot their business needs directly and simply into proven DW design patterns.


BEAM✲ encourages DW/BI designers to move away from the keyboard and their entity relationship modeling tools and begin “white board” modeling interactively with BI stakeholders.  With the right guidance, BI stakeholders can and should model their own BI data requirements, so that they can fully understand and govern what they will be able to report on and analyze.


The BEAM✲ method is fully described in

Agile Data Warehouse Design – a text co-written by Lawrence Corr and Jim Stagnitto.


About the speaker:

Jim Stagnitto Director of a2c Data Services Practice

Data Warehouse Architect: specializing in powerful designs that extract the maximum business benefit from Intelligence and Insight investments.

Master Data Management (MDM) and Customer Data Integration (CDI) strategist and architect.

Data Warehousing, Data Quality, and Data Integration thought-leader: co-author with Lawrence Corr of “Agile Data Warehouse Design”, guest author of Ralph Kimball’s “Data Warehouse Designer” column, and contributing author to Ralph and Joe Caserta’s latest book: “The DW ETL Toolkit”.


John DiPietro Chief Technology Officer at A2C IT Consulting

John DiPietro is the Chief Technology Officer for a2c. Mr. DiPietro is responsible
for setting the vision, strategy, delivery, and methodologies for a2c’s Solution
Practice Offerings for all national accounts. The a2c CTO brings with him an
expansive depth and breadth of specialized skills in his field.


Sponsor Note:

Thanks to:

Microsoft NERD for providing awesome venue for the event.

A2C IT Consulting for providing the food/drinks.

Cognizeus for providing book to give away as raffle.

Here’s the youtube link for the presentation:

And Slideshare:

Originally Posted at: Agile Data Warehouse Design for Big Data by v1shal

Nov 02, 17: #AnalyticsClub #Newsletter (Events, Tips, News & more..)


Pacman  Source


More WEB events? Click Here

[ AnalyticsWeek BYTES]

>> Media firms are excelling at social: Reach grows by 236% by analyticsweekpick

>> Big Data Provides Big Insights for U.S. Hospitals by bobehayes

>> The Evolution of Master Data Management: Connecting the Big Data Dots by jelaniharper

Wanna write? Click Here


 Free analytics tool lets developers troubleshoot app performance – BetaNews Under  Analytics

 Goldman Sachs Leads $38M Investment in Cloud Storage Firm … – Xconomy Under  Cloud

 Zodiac Aerospace flies in Birst to support local data needs – Diginomica Under  Business Analytics

More NEWS ? Click Here


Intro to Machine Learning


Machine Learning is a first-class ticket to the most exciting careers in data analysis today. As data sources proliferate along with the computing power to process them, going straight to the data is one of the most stra… more


Storytelling with Data: A Data Visualization Guide for Business Professionals


Storytelling with Data teaches you the fundamentals of data visualization and how to communicate effectively with data. You’ll discover the power of storytelling and the way to make data a pivotal point in your story. Th… more


Data Have Meaning
We live in a Big Data world in which everything is quantified. While the emphasis of Big Data has been focused on distinguishing the three characteristics of data (the infamous three Vs), we need to be cognizant of the fact that data have meaning. That is, the numbers in your data represent something of interest, an outcome that is important to your business. The meaning of those numbers is about the veracity of your data.


Q:Explain what a local optimum is and why it is important in a specific context,
such as K-means clustering. What are specific ways of determining if you have a local optimum problem? What can be done to avoid local optima?

A: * A solution that is optimal in within a neighboring set of candidate solutions
* In contrast with global optimum: the optimal solution among all others

* K-means clustering context:
It’s proven that the objective cost function will always decrease until a local optimum is reached.
Results will depend on the initial random cluster assignment

* Determining if you have a local optimum problem:
Tendency of premature convergence
Different initialization induces different optima

* Avoid local optima in a K-means context: repeat K-means and take the solution that has the lowest cost



#FutureOfData Podcast: Peter Morgan, CEO, Deep Learning Partnership

 #FutureOfData Podcast: Peter Morgan, CEO, Deep Learning Partnership

Subscribe to  Youtube


You can have data without information, but you cannot have information without data. – Daniel Keys Moran


#BigData @AnalyticsWeek #FutureOfData #Podcast with @DavidRose, @DittoLabs

 #BigData @AnalyticsWeek #FutureOfData #Podcast with @DavidRose, @DittoLabs


iTunes  GooglePlay


Every person in the world having more than 215m high-resolution MRI scans a day.

Sourced from: Analytics.CLUB #WEB Newsletter