Data Vault Games with Cindi

By No Comments

Cynthia Meyersohn

C​indi has worked in a variety of IT realms over the past 35 years and, as of 2018, had spent the last 17 years working in applications and data engineering development within the U.S. DoD. As a Data Vault 2.0 (DV2) Solution Architect and Certified Instructor, her responsibilities and expertise range from the design, development, implementation, and technical guidance of Enterprise Data Warehouse/Big Data builds to crafting processes surrounding data acquisition and ingest, data governance and Master Data Management policy and compliance, development and team leadership.

Cindi has spent the past seven years leading the architectural design, implementation, and development of Data Vault 2.0 solutions at the U.S. DoD and Department of State. She is a Certified Authorized Data Vault 2.0 Instructor.

Cindi holds a MS in Systems Engineering from George Washington University and a BS in Information Systems from Strayer University.

Christian Kurze

By No Comments

MongoDB: A general purpose, distributed, and highly scalable data platform for modern applications

 

ABSTRACT

The database for modern applications: MongoDB is a general purpose, document-based, distributed database built for modern application developers and for the cloud era. No database is more productive to use.
MongoDB emerged into a general purpose database that easily allows to build globally distributed data platforms that are highly available and scale almost indefinitely. While NoSQL is still considered as a “new” technology, many of the fortune 1000 companies already migrated mission-critical workloads and decided to use MongoDB as a strategic data platform.

Due to its flexibility, the JSON-based document model support a bandwidth of use cases, like Single View, Internet of Things, Mobile, Real-Time Applications, Personalization, Content Management, Catalogs and Mainframe Offloading.

This presentation provides and overview of MongoDB, the document model, and how data can be accessed in many different ways via native drivers in almost any programming language, but also connectors like Spark or R and even SQL. A practical example shows how to use MongoDB for Data Vault creation in the insurance industry.

Bio

Christian spent the last couple of years on data management and data integration in order to generate value out of data. In MongoDB he works as a Principal Solutions Architect. Prior to joining MongoDB, he worked on data virtualization, data warehousing and active metadata management. He holds a PhD in data warehouse automation.

What you will learn

  • Comparison of the document model vs. the relational model
  • Native high availability, horizontal scalability, workload isolation and data locality
  • Deployment agnostic: on-prem, hybrid, cloud, Kubernetes
  • Additional features for rich data usage like S3-based data lake access, full-text search, access by analytical tools, etc.
  • Example how to build a Data Vault in MongoDB

 

 

Neil Strange

By No Comments

What’s so scary? Safely migrating to a Big Data, Data Vault Solution from a legacy Kimball data warehouse

ABSTRACT

A frequent question we get asked is “how can I migrate from my existing Kimball data warehouse to a big data Data Vault solution?”
But what do we mean by migration? And what are the implications of choosing a big data architecture? Can we use Snowflake or a Azure SQL Data Warehouse to run our new system? Where do we start?
This presentation will explore the migration question and suggest some good practice for designing a big data Data Vault target architecture.

BIO

Neil is the founder and managing director of Datavault UK, a consultancy specialising in Data Vault 2.0 and Information Governance implementations and coaching. He has many years experience working with a diverse range of clients and industries helping organisations make the best strategic use of their IT systems and data services. Neil has presented at the previous three WWDVC events in the USA.

WHAT YOU WILL LEARN

  • How to define your migration project.
  • Architecting your big data Data Vault target solution.
  • Working on the migration process.
  • Migration good practice.

André Dörr

By No Comments

Data Vault in sports analytics

Abstract:

Everything started with Moneyball in 2002. It’s the first well know use case, where a sports team used a data-driven approach to measure player value. In the meantime, many sports clubs tried to copy this method. And with more and more  technology entering sports, more and more data is collected and analyzed to get an edge in the competition.

PRESENTATION

This presentation will take a look at different sports analytics use cases for football clubs.
– Technical challenges in football clubs
– Building a compact analytical architecture based on Data Vault & Exasol
– Data Science in sports analytics with Data Vault & Exasol

Matthias Wegner

By No Comments

Data Vault + GDPR at mobile.de

ABSTRACT

Matthias Wegner is senior technical consultant for Data Warehouse platforms. He initiated and implemented Data Wareouse platforms for multiple projects and customers in Germany using Data Vault and Talend as the main toolsets.  Providing a tailored set of standards and best practices for all aspects of a Data Warehouse project is one of his main missions.
Matthias is Head of BI at cimt AG – IT consulting since 5 years.
Currently he works as the architect for the Data Warehouse of mobile.de where the concept of encryption of data for GDPR was developed and implemented.

 

WHAT YOU WILL LEARN

In this case study we will give you an overview of the data warehouse migration project at mobile.de. You will see how we address GDPR requirements and which role Talend plays in this project. We’ll also show how easy it is to virtualize the access layer through the database switch to Exasol.

· Data Warehouse state-of-the-art, overview and source landscape
· Full Data Vault architecture
· Team setting
· Toolset (Talend, Exasol, Confluence)
· Loading procedures with Talend / Exasol – ELT
· GDPR requirements
· Encryption architecture and decryption approach on the fly in Exasol
· Lessons learned

Matthias Reiß

By No Comments

A day at the data lake

Matthias Reiß is a Senior Client Technical Professional within the IBM Cloud and Cognitive Technical Sales Team in Germany.He has more than 15 years experience in Analytics and data integration projects in heterogeneous environments.

A day at the Data Lake – Get your data working in your Data Lake and beyond
Catch the big fish faster. Get the most out of your data in your Data Lake and all the data stores connected to it.Imagine how you can easily combine the different data formats in your lake with other relational and non-relational Data Stores within one single query.

– IBMs common and hybrid SQL Engine
– Data Virtualization
– Data Caching
– Polymorphic Table Functions (i. e. Apache Spark Integration)

Kent Graziano

By No Comments

Kent Graziano

Making Sense of Schema-On-Read

 

Abstract

Kent Graziano is the Chief Technical Evangelist for Snowflake Computing. His is award winning author, speaker, and trainer, in the areas of data modeling, data architecture, and data warehousing. He is a certified Data Vault Master and Data Vault 2.0 Practitioner (CDVP2), an Oracle ACE Director (Alumni), member of the OakTable Network, expert data modeler and solution architect with more than 30 years of experience, including over two decades doing data warehousing and business intelligence (in multiple industries). He is an internationally recognized expert in Data Vault, Oracle SQL Developer Data Modeler, Agile Data Warehousing, and Cloud-based Data Warehousing. Mr. Graziano has created and led many successful software and data warehouse implementation teams, including multiple agile DW/BI teams. He has written numerous articles, authored three Kindle book (available on Amazon.com), co-authored four books (including the 1st Edition of The Data Model Resource Book), and has given hundreds of presentations, nationally and internationally. He was a co-author on the first book on Data Vault, and the technical editor for Super Charge Your Data Warehouse. You can follow Kent on twitter @KentGraziano or on his blog The Data Warrior (http://kentgraziano.com).

presentation

Making Sense of Schema-On-Read

With the increasing prevalence of semi-structured data from IoT devices, web logs, and other sources, data architects and modelers have to learn how to interpret and project data from things like JSON. While the concept of loading data without upfront modeling is appealing to many, ultimately, in order to make sense of the data and use it to drive business value, we have to turn that schema-on-read data into a real schema! That means data modeling! In this session I will walk through both simple and complex JSON documents, decompose them, then turn them into a representative data model using Oracle SQL Developer Data Modeler. I will show you how they might look using both traditional 3NF and data vault styles of modeling.

  1. See what a JSON document looks like
  2. Understand how to read it
  3. Learn how to convert it to a standard data model

Milan Steinhauer & Annekathrin Meinzer

By No Comments

POC converting an existing EDW into new technologies

Abstract

Like many other insurance companies in the market, ROLAND is facing increasing requirements for a data warehouse, e.g. fulfilment of GDPR as well as more flexibility and scalabiltiy. ROLAND takes this challenge and is working on transforming the existing data warehouse into new (cloud based) technologies. Data vault also takes a key role in this context. For choosing the right technologies, ROLAND tried three different solutions within a proof of concept.

Presentation:

In this presentation we will talk about how we set up our project for choosing the technology and what challenges we met on our way.

About us

Milan Steinhauer and Annekathrin Meinzer are responsible for the ongoing project at ROLAND Rechtsschutz-Versicherungs-AG.
Milan is a member within the BICC Team and Annekathrin is a project manager at ROLAND Rechtsschutz.

Scott Ambler

By

Choose Your WoW! How Agile Software Teams Can Optimize Their Way of Working (WoW)

We like to say that agile teams own their own process by choosing their way of working, their “WoW.” This of course is easier said than done because there are several aspects to WoW. First, our team needs to know how to choose the appropriate lifecycle for the situation that we face. Should we take a Scrum-based approach, a lean/Kanban-based approach, a continuous delivery approach, or an exploratory/lean startup approach? Second, what practices should the team adopt? How do they fit together? When should we apply them? Third, what artifacts should the team create? When should they be created? To what level of detail? Finally, how do we evolve our WoW as we experiment and learn?

There are several strategies that we could choose to follow when we tailor and evolve our WoW. A common approach is to adopt a prescriptive method or framework such as Scrum or SAFe and follow it, but our team will soon find itself in “method prison” with no guidance for how to improve from there. A second approach is to bootstrap our WoW, to figure it out on our own. This works, but it is a very slow and expensive strategy in practice. A third strategy is to hire an agile coach, but sadly in practice the majority of coaches seem to be like professors who are only a chapter or two ahead of their students. Or we could take a more disciplined, streamlined approach and leverage the experiences of the thousands of teams who have already struggled through the very issues that our team currently faces. This is something we call Guided Continuous Improvement (GCI), which is effectively a Kaizen loop on steroids.

Prerequisites:

  • An understanding of agile fundamentals
  • Some experience on an agile team
  • An open mind

Agenda:

  1. Being Agile: Principles for agile and lean in the enterprise
  2. Context counts: The Software Development Context Framework (SDFC)
  3. Guided Continuous Improvement (GCI)
  4. Identify your initial Way of Working (WoW) – Tailoring your approach for the right outcomes
  5. Breaking out of “method prison” – How to evolve beyond Scrum, SAFe, and others
  6. Choosing an appropriate lifecycle – Scrum? Kanban? Continuous Delivery? Exploratory?
  7. Improve your WoW – Evolve your approach through thoughtful experiments

Learning Outcomes:

  • Every agile/lean team will form a unique way of working (WoW)
  • How to choose between Scrum-based, Kanban-based, Experimental, and Continuous Delivery lifecycles yet still enable consistent governance between them
  • How to use light-weight process guidance to enable teams to choose and evolve their WoW

Dr. Alexander Brunner

By

Analysis and Presentation of Blockchain Data in Real-time

Abstract:

In this state-of-the art use case, the Data Vault 2.0 Business Intelligence is applied to extract Information from Etherium-blockchain data in near real-time. Thereby, the ingestion of semi-structured data is covered as well its processing on an MPP-system. In addition, the talk provides tips and tricks for Greenplum databases.

Bio:

With extensive experience in Information Technology stretching over more than 10 years, Dr. Alexander Brunner is a Scalefree Partner for the finance sector since October 2018. As a senior IT- and financial risk management professional, he has expertise in conducting technical- as well as business-driven projects at investment banks, credit banks, state-owned banks and real-estate banks. He has focused on the introduction of risk management systems, financial engineering (robo-advisory) and advising financial institutions to meet regulatory requirements regarding liquidity- and market risk.

Sven Börjesson

By

SCALEFREE PROJECT PROCESS MODEL

HOW TO SET FREE YOUR CAPABILITIES

 

Abstract

Empowering your team is essential for successfully implementing your enterprise data warehousing strategy. The Scalefree Project Process Model is the key for enabling the capabilities of your organisation and building solutions successfully.

 

Bio

Sven Börjesson has more than 15 years of experience in Information Technology with a focus on Business Intelligence solutions. He mainly worked in end-to-end BI projects from the strategy definition and requirements analysis all the way to the design and implementation of BI solutions. After managing projects following different models and standards like Waterfall and PMI, he recognized the charm of an agile approach to adapt to fast-changing requirements. Since then Sven engages himself in agile initiatives based on Data Vault 2.0. Since June 2018 Sven is a Scalefree Partner for the insurance sector.