SlideShare a Scribd company logo
1 of 27
TORQUE
IT SOLUTIONS
                                       BTECH 451
Empowering Automotive Finance                   Data Integration


          Final Presentation
                                Shawn D’souza
                                   Oct2012
DEFINITION     NEED FOR DI       CHALLENGES FOR DI     APPROACHES




 PREVIOUSLY   TECHNICAL DETAILS        DEMO           SOLUTION ANALYSIS




FUTURE WORK     CONCLUSION        EXPERIENCE GAINED      THANK YOU
Data integration involves combining
data residing in different sources and
providing users with a unified view of
these data.[1]




           Maurizio Lenzerini (2002). "Data Integration: A Theoretical Perspective". PODS 2002. pp. 233–246
Data warehouse                                                  Live Reporting
    Pros:                                                         Pros:
    • Reports run against the Data Warehouse rather than          • Less costly
      your production database so your production                 • Less complicated
      database can be dedicated to transactional                  • “IT Lite” with much less reliance on IT resources
      processing rather than reporting
                                                                  • Reports run against live production data rather
    • Reporting can be faster                                       than a Data Warehouse so you know all data
    • Static Metadata is provided in the Data Warehouse             returned in reports is guaranteed to be the most
                                                                    recent data in DPMS environment
                                                                  • Reports may run up to 10 to 30 times faster
                                                                    with Live Data reporting than with existing
Cons:                                                               DPMS
• Building or buying pre-built Data Warehouses is more
  expensive than a Live Data strategy
• “IT intensive” with heavy reliance on IT support              Cons:
• Resources intensive to manage, maintain, and provide          • If POS tables are purged then tables often you will have to
  additional content on an ongoing basis                          be copied first if you want to report historical information
• The frequency of data being refreshed in the Data               with a Live Data strategy
  Warehouse may impact reporting                                • Report processing is shared with transactional processing
• Requires additional database software to store data and ETL     on DPMS database
  software to populate your Data Warehouse
• Querying on business activities, for statistical
 analysis, online analytical processing (OLAP), and
 data mining in order to en-able forecasting, decision
 making, enterprise-wide planning, and, in the end,
• To gain sustainable competitive advantages.
• Requirements for improved customer service or self-
 service
• Data quality
  •   The data integration team must promote data quality to a first-class citizen.
• Transparency and auditability
  •   Even high-quality results will be questioned by business consumers.
      Providing complete transparency into how the data results were produced
      will be necessary to relieve business consumers’ concerns around data
      quality.
• Tracking history
  •   The ability to correctly report results at a particular period in time is an on-
      going challenge, particularly when there are adjustments to historical data.
• Reducing processing times
  •   Efficiently processing very large volumes of data within ever shortening
      processing windows is an on-going challenge for the data integration team
[Dittrich and Jonscher, 1999], All Together Now — Towards Integrating the World’s Information Systems
• Manual Integration
  •   users directly interact with all relevant information systems and manually integrate
      selected data
• Common User Interface
  •   the user is supplied with a common user interface (e.g., a web browser) that provides a
      uniform look and feel.
• Integration by Applications
  •   Applications that access various data sources and return integrated results to the user
• Integration by Middleware
  •   reusable functionality that is generally used to solve dedicated aspects of the integration
      problem
• Uniform Data Access
  •   a logical integration of data is accomplished at the data access level
• Common Data Storage
  •   physical data integration is performed by transferring data to a new data storage


             [Dittrich and Jonscher, 1999], All Together Now — Towards Integrating the World’s Information Systems
•   Enterprise Information Integration (EII) – This pattern loosely couples multiple data stores by creating
    a semantic layer above the data stores and using industry-standard APIs such as ODBC, OLE-DB, and
    JDBC to access the data in real time.
•   Enterprise Application Integration (EAI) – This pattern supports business processes and workflows
    that span multiple application systems. It typically works on a message-/event-based model and is not
    data-centric (i.e., it is parameter-based and does not pass more than one “record” at a time).


• Extract, Transform, and Load (ETL) – This pattern extracts data from
    sources, transforms the data in memory and then loads it into a destination.

•   Extract, Load, and Transform (ELT) – This pattern first extracts data from sources and loads it into a
    relational database. The transformation is then performed within the relational database and not in
    memory.
•   Replication – This is a relational database feature that detects changed records in a source and
    pushes the changed records to a destination or destinations. The destination is typically a mirror of the
    source, meaning that the data is not transformed on the way from source to destination.
Torque IT Solutions
          • Provides I.T Solutions For Automotive
              Finance Companies And Car Dealerships

    The   •
          •
              Start-up
              Dealer Performance Management System
          •
Company       Show Profit Potential
The Goal
Implement an Interface that will allow users to
    Import Data from external databases
Dealer’s Vehicle
   Database



                                                                    P.O.S

                                                                                               The

                                               Ad-hoc Pull
                                                                                       Architecture
 Database Dump
   Overnight




                          D.P.M.S

                                                             Pull


    Export File
                                                                    Reference Tables

                   Customised import process

                                                                                         Standardised
                                                                                       D.P.M.S Database
PosImportService

                                                   +SetExternalLogImportSource
                                                   +GetSearchParameters();
                     << ILogImportService >>       +GetSearchResults

                     +SetExternalLogImportSource
  Request from       +GetSearchParameters();
Presentation Layer   +GetSearchResults
                                                   DpmsImportService

                                                   +SetExternalLogImportSource
                                                   +GetSearchParameters();
                                                   +GetSearchResults
• Limitations
• Pros
   •   Flexibility – allows new external data sources to be easily
       configured
• Cons
   •   Exact match
• Bulk Import
• Edge server caching
• database caching at edge servers enables dynamic
  content to be replicated at the edge of the
  network, thereby improving the scalability and the
  response time of Web applications.
• Integrates data service technology and edge server
  data replication architecture, in order to improve Web
  services‟ data performance and address a variety of
  data issues in the SOA network.
• Provide data services with edge
  server data replication to clients
• Increase data service
  performance
• Reduce client-perceived
  response time
• Ensure data consistency is more
  easily achieved
• Importance of DI
• Issues for DI
• How You can improve DI
• Scalability considerations for DI
• NHibernate
• MVC .NET
• jQuery
• SQL Server
Xinfeg Ye
  Academic Advisor


Frederik Dinkelaker
   Industrial Mentor

More Related Content

What's hot

Storage simplicity value_110810
Storage simplicity value_110810Storage simplicity value_110810
Storage simplicity value_110810rjmurphyslideshare
 
Comparison of MPP Data Warehouse Platforms
Comparison of MPP Data Warehouse PlatformsComparison of MPP Data Warehouse Platforms
Comparison of MPP Data Warehouse PlatformsDavid Portnoy
 
Hadoop in the Enterprise: Legacy Rides the Elephant
Hadoop in the Enterprise: Legacy Rides the ElephantHadoop in the Enterprise: Legacy Rides the Elephant
Hadoop in the Enterprise: Legacy Rides the ElephantDataWorks Summit
 
Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...
Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...
Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...Cloudera, Inc.
 
How Real TIme Data Changes the Data Warehouse
How Real TIme Data Changes the Data WarehouseHow Real TIme Data Changes the Data Warehouse
How Real TIme Data Changes the Data Warehousemark madsen
 
Green Plum IIIT- Allahabad
Green Plum IIIT- Allahabad Green Plum IIIT- Allahabad
Green Plum IIIT- Allahabad IIIT ALLAHABAD
 
A Time Traveller’s Guide to DB2: Technology Themes for 2014 and Beyond
A Time Traveller’s Guide to DB2: Technology Themes for 2014 and BeyondA Time Traveller’s Guide to DB2: Technology Themes for 2014 and Beyond
A Time Traveller’s Guide to DB2: Technology Themes for 2014 and BeyondSurekha Parekh
 
FAS2240: An Inside Look
FAS2240: An Inside LookFAS2240: An Inside Look
FAS2240: An Inside LookNetApp
 
IBM Spectrum Scale and Its Use for Content Management
 IBM Spectrum Scale and Its Use for Content Management IBM Spectrum Scale and Its Use for Content Management
IBM Spectrum Scale and Its Use for Content ManagementSandeep Patil
 
Gartner magic quadrant for data warehouse database management systems
Gartner magic quadrant for data warehouse database management systemsGartner magic quadrant for data warehouse database management systems
Gartner magic quadrant for data warehouse database management systemsparamitap
 
OpenPOWER Roadmap Toward CORAL
OpenPOWER Roadmap Toward CORALOpenPOWER Roadmap Toward CORAL
OpenPOWER Roadmap Toward CORALinside-BigData.com
 
SnapLogic corporate presentation
SnapLogic corporate presentationSnapLogic corporate presentation
SnapLogic corporate presentationpbridges
 
BarbaraZigmanResume 2016
BarbaraZigmanResume 2016BarbaraZigmanResume 2016
BarbaraZigmanResume 2016bzigman
 
Software Defined Storage - Open Framework and Intel® Architecture Technologies
Software Defined Storage - Open Framework and Intel® Architecture TechnologiesSoftware Defined Storage - Open Framework and Intel® Architecture Technologies
Software Defined Storage - Open Framework and Intel® Architecture TechnologiesOdinot Stanislas
 
Oracle en Entel Summit 2010
Oracle en Entel Summit 2010Oracle en Entel Summit 2010
Oracle en Entel Summit 2010Entel
 
A-B-C Strategies for File and Content Brochure
A-B-C Strategies for File and Content BrochureA-B-C Strategies for File and Content Brochure
A-B-C Strategies for File and Content BrochureHitachi Vantara
 
Big data analytics beyond beer and diapers
Big data analytics   beyond beer and diapersBig data analytics   beyond beer and diapers
Big data analytics beyond beer and diapersKai Zhao
 
Data Ninja Webinar Series: Realizing the Promise of Data Lakes
Data Ninja Webinar Series: Realizing the Promise of Data LakesData Ninja Webinar Series: Realizing the Promise of Data Lakes
Data Ninja Webinar Series: Realizing the Promise of Data LakesDenodo
 

What's hot (19)

Storage simplicity value_110810
Storage simplicity value_110810Storage simplicity value_110810
Storage simplicity value_110810
 
Comparison of MPP Data Warehouse Platforms
Comparison of MPP Data Warehouse PlatformsComparison of MPP Data Warehouse Platforms
Comparison of MPP Data Warehouse Platforms
 
Hadoop in the Enterprise: Legacy Rides the Elephant
Hadoop in the Enterprise: Legacy Rides the ElephantHadoop in the Enterprise: Legacy Rides the Elephant
Hadoop in the Enterprise: Legacy Rides the Elephant
 
Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...
Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...
Hadoop World 2011: I Want to Be BIG - Lessons Learned at Scale - David "Sunny...
 
How Real TIme Data Changes the Data Warehouse
How Real TIme Data Changes the Data WarehouseHow Real TIme Data Changes the Data Warehouse
How Real TIme Data Changes the Data Warehouse
 
Green Plum IIIT- Allahabad
Green Plum IIIT- Allahabad Green Plum IIIT- Allahabad
Green Plum IIIT- Allahabad
 
Teradata - Architecture of Teradata
Teradata - Architecture of TeradataTeradata - Architecture of Teradata
Teradata - Architecture of Teradata
 
A Time Traveller’s Guide to DB2: Technology Themes for 2014 and Beyond
A Time Traveller’s Guide to DB2: Technology Themes for 2014 and BeyondA Time Traveller’s Guide to DB2: Technology Themes for 2014 and Beyond
A Time Traveller’s Guide to DB2: Technology Themes for 2014 and Beyond
 
FAS2240: An Inside Look
FAS2240: An Inside LookFAS2240: An Inside Look
FAS2240: An Inside Look
 
IBM Spectrum Scale and Its Use for Content Management
 IBM Spectrum Scale and Its Use for Content Management IBM Spectrum Scale and Its Use for Content Management
IBM Spectrum Scale and Its Use for Content Management
 
Gartner magic quadrant for data warehouse database management systems
Gartner magic quadrant for data warehouse database management systemsGartner magic quadrant for data warehouse database management systems
Gartner magic quadrant for data warehouse database management systems
 
OpenPOWER Roadmap Toward CORAL
OpenPOWER Roadmap Toward CORALOpenPOWER Roadmap Toward CORAL
OpenPOWER Roadmap Toward CORAL
 
SnapLogic corporate presentation
SnapLogic corporate presentationSnapLogic corporate presentation
SnapLogic corporate presentation
 
BarbaraZigmanResume 2016
BarbaraZigmanResume 2016BarbaraZigmanResume 2016
BarbaraZigmanResume 2016
 
Software Defined Storage - Open Framework and Intel® Architecture Technologies
Software Defined Storage - Open Framework and Intel® Architecture TechnologiesSoftware Defined Storage - Open Framework and Intel® Architecture Technologies
Software Defined Storage - Open Framework and Intel® Architecture Technologies
 
Oracle en Entel Summit 2010
Oracle en Entel Summit 2010Oracle en Entel Summit 2010
Oracle en Entel Summit 2010
 
A-B-C Strategies for File and Content Brochure
A-B-C Strategies for File and Content BrochureA-B-C Strategies for File and Content Brochure
A-B-C Strategies for File and Content Brochure
 
Big data analytics beyond beer and diapers
Big data analytics   beyond beer and diapersBig data analytics   beyond beer and diapers
Big data analytics beyond beer and diapers
 
Data Ninja Webinar Series: Realizing the Promise of Data Lakes
Data Ninja Webinar Series: Realizing the Promise of Data LakesData Ninja Webinar Series: Realizing the Promise of Data Lakes
Data Ninja Webinar Series: Realizing the Promise of Data Lakes
 

Similar to Fi nf068c73aef66f694f31a049aff3f4

3d9f068c73d9f068c73aef66f694f31a049aff3f43a3d9f068c73aef66f694f31a049aff3f4ef...
3d9f068c73d9f068c73aef66f694f31a049aff3f43a3d9f068c73aef66f694f31a049aff3f4ef...3d9f068c73d9f068c73aef66f694f31a049aff3f43a3d9f068c73aef66f694f31a049aff3f4ef...
3d9f068c73d9f068c73aef66f694f31a049aff3f43a3d9f068c73aef66f694f31a049aff3f4ef...Shawn D'souza
 
ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...
ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...
ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...DATAVERSITY
 
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)Denodo
 
ADV Slides: When and How Data Lakes Fit into a Modern Data Architecture
ADV Slides: When and How Data Lakes Fit into a Modern Data ArchitectureADV Slides: When and How Data Lakes Fit into a Modern Data Architecture
ADV Slides: When and How Data Lakes Fit into a Modern Data ArchitectureDATAVERSITY
 
Bridging the Last Mile: Getting Data to the People Who Need It
Bridging the Last Mile: Getting Data to the People Who Need ItBridging the Last Mile: Getting Data to the People Who Need It
Bridging the Last Mile: Getting Data to the People Who Need ItDenodo
 
Data Mesh using Microsoft Fabric
Data Mesh using Microsoft FabricData Mesh using Microsoft Fabric
Data Mesh using Microsoft FabricNathan Bijnens
 
Journey to the Programmable Data Center
Journey to the Programmable Data CenterJourney to the Programmable Data Center
Journey to the Programmable Data CenterToby Weiss
 
The Shifting Landscape of Data Integration
The Shifting Landscape of Data IntegrationThe Shifting Landscape of Data Integration
The Shifting Landscape of Data IntegrationDATAVERSITY
 
Building a business intelligence architecture fit for the 21st century by Jon...
Building a business intelligence architecture fit for the 21st century by Jon...Building a business intelligence architecture fit for the 21st century by Jon...
Building a business intelligence architecture fit for the 21st century by Jon...Mark Tapley
 
Data Warehousing, Data Mining & Data Visualisation
Data Warehousing, Data Mining & Data VisualisationData Warehousing, Data Mining & Data Visualisation
Data Warehousing, Data Mining & Data VisualisationSunderland City Council
 
La creación de una capa operacional con MongoDB
La creación de una capa operacional con MongoDBLa creación de una capa operacional con MongoDB
La creación de una capa operacional con MongoDBMongoDB
 
Building a Single Logical Data Lake: For Advanced Analytics, Data Science, an...
Building a Single Logical Data Lake: For Advanced Analytics, Data Science, an...Building a Single Logical Data Lake: For Advanced Analytics, Data Science, an...
Building a Single Logical Data Lake: For Advanced Analytics, Data Science, an...Denodo
 
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)Denodo
 
Denodo DataFest 2017: Conquering the Edge with Data Virtualization
Denodo DataFest 2017: Conquering the Edge with Data VirtualizationDenodo DataFest 2017: Conquering the Edge with Data Virtualization
Denodo DataFest 2017: Conquering the Edge with Data VirtualizationDenodo
 
Product overview 6.0 v.1.0
Product overview 6.0 v.1.0Product overview 6.0 v.1.0
Product overview 6.0 v.1.0Gianluigi Riccio
 
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization Denodo
 
Which Change Data Capture Strategy is Right for You?
Which Change Data Capture Strategy is Right for You?Which Change Data Capture Strategy is Right for You?
Which Change Data Capture Strategy is Right for You?Precisely
 
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)Denodo
 
Migrer vos bases Oracle vers du SQL, le tout dans Azure !
Migrer vos bases Oracle vers du SQL, le tout dans Azure !Migrer vos bases Oracle vers du SQL, le tout dans Azure !
Migrer vos bases Oracle vers du SQL, le tout dans Azure !Microsoft Technet France
 

Similar to Fi nf068c73aef66f694f31a049aff3f4 (20)

3d9f068c73d9f068c73aef66f694f31a049aff3f43a3d9f068c73aef66f694f31a049aff3f4ef...
3d9f068c73d9f068c73aef66f694f31a049aff3f43a3d9f068c73aef66f694f31a049aff3f4ef...3d9f068c73d9f068c73aef66f694f31a049aff3f43a3d9f068c73aef66f694f31a049aff3f4ef...
3d9f068c73d9f068c73aef66f694f31a049aff3f43a3d9f068c73aef66f694f31a049aff3f4ef...
 
4512012
45120124512012
4512012
 
ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...
ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...
ADV Slides: Platforming Your Data for Success – Databases, Hadoop, Managed Ha...
 
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
Bridging the Last Mile: Getting Data to the People Who Need It (APAC)
 
ADV Slides: When and How Data Lakes Fit into a Modern Data Architecture
ADV Slides: When and How Data Lakes Fit into a Modern Data ArchitectureADV Slides: When and How Data Lakes Fit into a Modern Data Architecture
ADV Slides: When and How Data Lakes Fit into a Modern Data Architecture
 
Bridging the Last Mile: Getting Data to the People Who Need It
Bridging the Last Mile: Getting Data to the People Who Need ItBridging the Last Mile: Getting Data to the People Who Need It
Bridging the Last Mile: Getting Data to the People Who Need It
 
Data Mesh using Microsoft Fabric
Data Mesh using Microsoft FabricData Mesh using Microsoft Fabric
Data Mesh using Microsoft Fabric
 
Journey to the Programmable Data Center
Journey to the Programmable Data CenterJourney to the Programmable Data Center
Journey to the Programmable Data Center
 
The Shifting Landscape of Data Integration
The Shifting Landscape of Data IntegrationThe Shifting Landscape of Data Integration
The Shifting Landscape of Data Integration
 
Building a business intelligence architecture fit for the 21st century by Jon...
Building a business intelligence architecture fit for the 21st century by Jon...Building a business intelligence architecture fit for the 21st century by Jon...
Building a business intelligence architecture fit for the 21st century by Jon...
 
Data Warehousing, Data Mining & Data Visualisation
Data Warehousing, Data Mining & Data VisualisationData Warehousing, Data Mining & Data Visualisation
Data Warehousing, Data Mining & Data Visualisation
 
La creación de una capa operacional con MongoDB
La creación de una capa operacional con MongoDBLa creación de una capa operacional con MongoDB
La creación de una capa operacional con MongoDB
 
Building a Single Logical Data Lake: For Advanced Analytics, Data Science, an...
Building a Single Logical Data Lake: For Advanced Analytics, Data Science, an...Building a Single Logical Data Lake: For Advanced Analytics, Data Science, an...
Building a Single Logical Data Lake: For Advanced Analytics, Data Science, an...
 
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
 
Denodo DataFest 2017: Conquering the Edge with Data Virtualization
Denodo DataFest 2017: Conquering the Edge with Data VirtualizationDenodo DataFest 2017: Conquering the Edge with Data Virtualization
Denodo DataFest 2017: Conquering the Edge with Data Virtualization
 
Product overview 6.0 v.1.0
Product overview 6.0 v.1.0Product overview 6.0 v.1.0
Product overview 6.0 v.1.0
 
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
DAMA & Denodo Webinar: Modernizing Data Architecture Using Data Virtualization
 
Which Change Data Capture Strategy is Right for You?
Which Change Data Capture Strategy is Right for You?Which Change Data Capture Strategy is Right for You?
Which Change Data Capture Strategy is Right for You?
 
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
Logical Data Lakes: From Single Purpose to Multipurpose Data Lakes (APAC)
 
Migrer vos bases Oracle vers du SQL, le tout dans Azure !
Migrer vos bases Oracle vers du SQL, le tout dans Azure !Migrer vos bases Oracle vers du SQL, le tout dans Azure !
Migrer vos bases Oracle vers du SQL, le tout dans Azure !
 

Fi nf068c73aef66f694f31a049aff3f4

  • 1. TORQUE IT SOLUTIONS BTECH 451 Empowering Automotive Finance Data Integration Final Presentation Shawn D’souza Oct2012
  • 2. DEFINITION NEED FOR DI CHALLENGES FOR DI APPROACHES PREVIOUSLY TECHNICAL DETAILS DEMO SOLUTION ANALYSIS FUTURE WORK CONCLUSION EXPERIENCE GAINED THANK YOU
  • 3.
  • 4. Data integration involves combining data residing in different sources and providing users with a unified view of these data.[1] Maurizio Lenzerini (2002). "Data Integration: A Theoretical Perspective". PODS 2002. pp. 233–246
  • 5. Data warehouse Live Reporting Pros: Pros: • Reports run against the Data Warehouse rather than • Less costly your production database so your production • Less complicated database can be dedicated to transactional • “IT Lite” with much less reliance on IT resources processing rather than reporting • Reports run against live production data rather • Reporting can be faster than a Data Warehouse so you know all data • Static Metadata is provided in the Data Warehouse returned in reports is guaranteed to be the most recent data in DPMS environment • Reports may run up to 10 to 30 times faster with Live Data reporting than with existing Cons: DPMS • Building or buying pre-built Data Warehouses is more expensive than a Live Data strategy • “IT intensive” with heavy reliance on IT support Cons: • Resources intensive to manage, maintain, and provide • If POS tables are purged then tables often you will have to additional content on an ongoing basis be copied first if you want to report historical information • The frequency of data being refreshed in the Data with a Live Data strategy Warehouse may impact reporting • Report processing is shared with transactional processing • Requires additional database software to store data and ETL on DPMS database software to populate your Data Warehouse
  • 6. • Querying on business activities, for statistical analysis, online analytical processing (OLAP), and data mining in order to en-able forecasting, decision making, enterprise-wide planning, and, in the end, • To gain sustainable competitive advantages. • Requirements for improved customer service or self- service
  • 7. • Data quality • The data integration team must promote data quality to a first-class citizen. • Transparency and auditability • Even high-quality results will be questioned by business consumers. Providing complete transparency into how the data results were produced will be necessary to relieve business consumers’ concerns around data quality. • Tracking history • The ability to correctly report results at a particular period in time is an on- going challenge, particularly when there are adjustments to historical data. • Reducing processing times • Efficiently processing very large volumes of data within ever shortening processing windows is an on-going challenge for the data integration team
  • 8. [Dittrich and Jonscher, 1999], All Together Now — Towards Integrating the World’s Information Systems
  • 9. • Manual Integration • users directly interact with all relevant information systems and manually integrate selected data • Common User Interface • the user is supplied with a common user interface (e.g., a web browser) that provides a uniform look and feel. • Integration by Applications • Applications that access various data sources and return integrated results to the user • Integration by Middleware • reusable functionality that is generally used to solve dedicated aspects of the integration problem • Uniform Data Access • a logical integration of data is accomplished at the data access level • Common Data Storage • physical data integration is performed by transferring data to a new data storage [Dittrich and Jonscher, 1999], All Together Now — Towards Integrating the World’s Information Systems
  • 10. Enterprise Information Integration (EII) – This pattern loosely couples multiple data stores by creating a semantic layer above the data stores and using industry-standard APIs such as ODBC, OLE-DB, and JDBC to access the data in real time. • Enterprise Application Integration (EAI) – This pattern supports business processes and workflows that span multiple application systems. It typically works on a message-/event-based model and is not data-centric (i.e., it is parameter-based and does not pass more than one “record” at a time). • Extract, Transform, and Load (ETL) – This pattern extracts data from sources, transforms the data in memory and then loads it into a destination. • Extract, Load, and Transform (ELT) – This pattern first extracts data from sources and loads it into a relational database. The transformation is then performed within the relational database and not in memory. • Replication – This is a relational database feature that detects changed records in a source and pushes the changed records to a destination or destinations. The destination is typically a mirror of the source, meaning that the data is not transformed on the way from source to destination.
  • 11.
  • 12. Torque IT Solutions • Provides I.T Solutions For Automotive Finance Companies And Car Dealerships The • • Start-up Dealer Performance Management System • Company Show Profit Potential
  • 13. The Goal Implement an Interface that will allow users to Import Data from external databases
  • 14. Dealer’s Vehicle Database P.O.S The Ad-hoc Pull Architecture Database Dump Overnight D.P.M.S Pull Export File Reference Tables Customised import process Standardised D.P.M.S Database
  • 15.
  • 16. PosImportService +SetExternalLogImportSource +GetSearchParameters(); << ILogImportService >> +GetSearchResults +SetExternalLogImportSource Request from +GetSearchParameters(); Presentation Layer +GetSearchResults DpmsImportService +SetExternalLogImportSource +GetSearchParameters(); +GetSearchResults
  • 17.
  • 18.
  • 19.
  • 20.
  • 21.
  • 22. • Limitations • Pros • Flexibility – allows new external data sources to be easily configured • Cons • Exact match • Bulk Import • Edge server caching
  • 23. • database caching at edge servers enables dynamic content to be replicated at the edge of the network, thereby improving the scalability and the response time of Web applications. • Integrates data service technology and edge server data replication architecture, in order to improve Web services‟ data performance and address a variety of data issues in the SOA network.
  • 24. • Provide data services with edge server data replication to clients • Increase data service performance • Reduce client-perceived response time • Ensure data consistency is more easily achieved
  • 25. • Importance of DI • Issues for DI • How You can improve DI • Scalability considerations for DI
  • 26. • NHibernate • MVC .NET • jQuery • SQL Server
  • 27. Xinfeg Ye Academic Advisor Frederik Dinkelaker Industrial Mentor

Editor's Notes

  1. I’ll be giving you a brief overview of the project I am going to be working on this year,I will talk to you a bit a about the company, what they doand what I will be working on with the company.
  2. Definition of data integration
  3. Di can be use for querrying ..
  4. Start-upStarted last year, the current system has the bare nProvides I.T Solutions For Automotive Finance Companies And Dealersby Assisting Car Dealerships, distributor/ manufacturers etc… Make The Most Profit From Car Sales, Finance And Income And After Sales I.E. Servicing And Parts.Dealer Performance Management SystemI will be working on the DPMS which stands for Dealer Performance Management System, this is the web application that will be used by the car dealers to help them make decisions.. Show Profit PotentialShow Profit Potential And Where business and operationalProcess Improvements Can Be Made, Using KPI or(key performance indicators) that will evaluate the success of deals in car sales, insurance ... ROI rate of income – how much profit u make on a saleF&amp;I Hand over time How many deals have been convertedTurn Data Into InformationThese dealers already have records of sales, however these records are date not information. The purpose of this system is to turn this data into information. This will be done finding trends in the data for example.. Turning data into informationInformation is quite simply an understanding of the relationships between pieces of data, or between pieces of data and other information. – this is done in the system through KPIs such as improved performance, competitive advantage, innovation, the sharing of lessons learned and continuous improvement of the organization.
  5. Currently the only way to input data into DPMS is by manual data entry that is: manually fill 50 or so fields, this leaves room for user to make mistake or enter inaccurate data and also takes up a lot of time.Avoid Manual Data EntryHow?Use Existing DataSo we Need a way for users to import their existing data from external systems into the application to help auto complete fields.The system need to be automatedA way to extract Data DPMS needs a way to use vehicle details informationfrom currently operated systems i.e. Point of Sale (POS), (Vehicle database)Allow users to auto-populate vehicle detail fields Automated Import of DataAuto-populate Form Fields Design and implement an interface that will allow user to load the data.Reconcile With Existing Data The system will convert different data formats into a standardised format that will be consumed by the system Flexible interface The interface need to be as Flexible as possible so the will it be a be to interface with as much systems as possible
  6. DPMS User in DPMS to have choice regarding where to pull data from (Mototrack or POS for example)P.O.S – point of sale - Presume Point of Sale System will provide Web service searchData dump - which will contain a record of the table structure and the data from a vehicle database and is usually in the form of a list of SQL statements.A new set of tables for reference dataETLExtract: I will extract data from the export, Transform: Then transform the data, e.g. combine or split columns, format the data eg. If the data in a different data format, the idea of this is to get all the data into be consistent with our database.Load: finally the data will be loaded into the reference tableCustomised import process will convert the data from data dump
  7. With a rise in the complexity of data and business demands, enterprises today find it challenging to handle a mass of application data and the relevant data issues in an efficient and flexible way. In this thesis, we proposed a Web-based data service system –
  8. A scalable solution