Vendor Landscape Plus: Data Integration Tools

Transcription

Vendor Landscape Plus: Data Integration Tools
Vendor Landscape Plus: Data Integration Tools
Integration is the name of the game.
Info-Tech Research Group
1
Introduction
Many enterprises are unfamiliar with the benefits of data integration tools.
Use this research to get a handle on how you can best integrate diverse
source data, and which software vendors will help you manage your data.
This Research Is Designed For:
This Research Will Help You:
 CIOs and IT managers of medium to large
 Understand the capabilities of data integration
firms who have informal Data Integration (DI)
processes.
 Enterprises with 2+ major systems (e.g. CRM,
ERP), a document repository, and increasing
Analytics needs.
 All data integration architectures.
tools, and their potential use cases.
 Differentiate between vendor offerings and
align potential solutions with your
organization‟s requirements.
 Shortlist DI vendors, prepare an RFP, and
score RFP responses.
 Develop an implementation strategy and
optimize your investment in data integration
tools.
Info-Tech Research Group
2
The integration space consists of data and process integration
as well as middleware; this set focuses on data integration
Info-Tech predicts that the tooling's capabilities will merge across all types
into single product family offerings within the next three to five years.
Data Integration
•
•
•
•
Reflects the convergence of
Enterprise Application
Integration (EAI), Data
Integration (DI) and Extract,
Transform, load (ETL) vendors.
Encompasses data
consolidation, federation,
propagation and access.
Is gaining momentum as
enterprises move from handcoding to tools.
Prevents silos of information
and enables an enterprise-wide
view of data.
For information on
Application Integration Strategy,
refer to Info-Tech‟s
Develop an Application Integration
Strategy
Process Integration
SOA and Middleware
• Is workflow automation for
processes, e.g. online ordering.
• Emphasizes a top-down process
design approach to a greater
extent than the other two,
starting from a process model.
• Allows organizations to
streamline process and creates
opportunity for coordinated
escalation
• Is used to connect enterprise
apps without transforming the
data.
• Creates access to data in
distributed architectures.
• Requires advanced architecture
to achieve success.
For information on
Business Process Management,
refer to Info-Tech‟s
Develop a Business Process
Management Strategy
For information on
Application Integration Middleware,
refer to Info-Tech‟s
Application Integration Middleware
Vendor Landscape
Info-Tech Research Group
3
Executive Summary
Understand Data Integration Trends & Considerations
•
•
•
•
•
DI tools are intended to free up developer resources by speeding up development, testing, and deployment.
Tools offer data federation, replication, and synchronization, not just ETL.
DI as a Service is ideal for the less mature who are ready to give up hand-coding.
Big Data and Analytics insights can provide valuable information to organizations. DI tools assist in managing new and
emerging types of data that are resulting from the data explosion.
Make the case to the CIO and the CFO and get budget approval for DI tools.
Develop a DI Tool Selection Strategy
•
•
•
Your data integration tool selection strategy is a function of your information architecture maturity, data architecture and
developer staff size.
DI tools aren‟t for everyone, review your required features and determine which solution is right for you.
Not all vendor offerings are equal. Choose the right one to suit your needs.
Evaluate Data Integration Vendor Offerings
•
•
•
Platform standardization doesn‟t lock you in with that vendor for tools; looking elsewhere is often more cost-effective.
Choose your solution carefully as products evaluated offer a tremendous range of functionality and target different needs.
If gaining budget approval is difficult due to data quality issues; consider tooling with data profiling and cleansing
capabilities.
Develop a Data Integration Implementation Strategy
•
•
•
Prioritize integration, include people, process and technology. And align tool deployment with existing architecture.
Develop your timeline in preparation for integration.
Introduce training early and paint a compelling picture of process ease with tools to convince developers that change is
good.
Info-Tech Research Group
4
DI is no longer a nice to have but a need to have to enable
companies to effectively manage multiple data sources
Info-Tech evaluated eight competitors in the data
integration market, including the following notable
performers:
Champions:
• Informatica provides a unified platform that is a one-stop shop for
data integration.
• IBM offers a highly capable and comprehensive integration
product that is backed by the flanking “Sphere” platform.
• SAP’s platform has evolved and matured into a comprehensive
portfolio with a strong product focus.
Value Award:
• SAP offers the most affordable data integration tool solution with a
high-performing product at the absolute lowest price.
Innovation Award:
• Talend’s open-source, subscription offering provides a quality
product at a reasonable price. They offer a free, downloadable
version for smaller companies, and a subscription based product
for larger enterprises.
Info-Tech Insight
1. Overcoming a hand-coding monopoly:
Developer resistance and prevalent inhouse hand-coding continue to be data
integration‟s greatest competitor even
though there are a number of viable
vendor tool solutions in the marketplace.
2. It’s the Era of Big Data and Business
Intelligence (BI):
The benefits of exploiting BI and Big Data
insights are sweeping though companies
of all sizes. Data integration software has
come under scrutiny recently as a solution
to the data mining challenge.
3. The move towards unified platforms:
Tools don‟t just integrate anymore. Many
vendors are now incorporating a mix of
technologies such as Extract Transform
Load (ETL), data quality, replication,
metadata, and Master Data Management
(MDM) data federation into their DI tools
as they move their solutions towards a
single unified platform.
Info-Tech Research Group
5
The Info-Tech Data Integration Tools Vendor Landscape
Champions receive high scores for most
evaluation criteria and offer excellent value.
They have a strong market presence and
are usually the trend setters for the industry.
Innovators have demonstrated innovative
product strengths that act as their
competitive advantage in appealing to niche
segments of the market.
Market Pillars are established players with
very strong vendor credentials, but with
more average product scores.
Emerging players are newer vendors who
are starting to gain a foothold in the
marketplace. They balance product and
vendor attributes, though score lower
relative to market Champions.
For an explanation of how the Info-Tech Vendor Landscape is created, please see Vendor Evaluation Methodology in the appendices.
Info-Tech Research Group
6
Every vendor has its strengths & weaknesses;
pick the one that works best for you
Product
Overall
Features
Usability
Vendor
Affordability
Architecture
Overall
Viability
Strategy
Reach
Channel
SAS/DataFlux
IBM
Informatica
Microsoft
Oracle
Pervasive
SAP
Talend
Legend
=Exemplary
= Good
= Adequate
=Inadequate
= Poor
For an explanation of how the Info-Tech Harvey Balls are calculated please see Vendor Evaluation Methodology in the appendices.
Info-Tech Research Group
7
The Data Integration Tools Value Index
What is a Value Score?
The Value Score indexes each vendor‟s product
offering and business strength relative to their
price point. It does not indicate vendor ranking.
Vendors that score high offer more bang-for-thebuck (e.g. features, usability, stability, etc.) than
the average vendor, while the inverse is true for
those that score lower.
Price-conscious enterprises may wish to give the
Value Score more consideration than those who
are more focused on specific vendor/product
attributes.
For an explanation of how the Info-Tech Value Index is calculated, please see Value Index Ranking Methodology in the appendices.
For an explanation of how normalized pricing is determined, please see Product Pricing Scenario & Methodology in the appendices.
Info-Tech Research Group
8
Table Stakes represent the minimum standard; without these
a product doesn’t even get reviewed
The Table Stakes
What Does This Mean?
Feature
Description
Batch Integration
Schedule integration processes to run on a
periodic basis to control the movement of
data from a source to target.
Data
Migration/Conversion
(ETL)
Extract, Transform and Load data from a
source to a target. Basic functionality for
date, string and numeric data manipulation.
Exception Reporting
and Notification
Ability to catch, report and handle
exceptions as they occur during
integration.
Metadata Management
Ability to manage the metadata associated
with data stores in the enterprise.
Administrative
Console
Software allowing users to see the current
state of integration processes during
execution, and intercede if necessary.
The products assessed in this Vendor
LandscapeTM meet, at the very least, the
requirements outlined as Table Stakes.
Many of the vendors go above and beyond the
outlined Table Stakes, some even do so in
multiple categories. This section aims to highlight
the products capabilities in excess of the criteria
listed here.
If Table Stakes are all you need from your Data Integration Tool, the only true differentiator for the
organization is price. Otherwise, dig deeper to find the best price to value for your needs.
Info-Tech Research Group
9
Advanced Features are the market differentiators that make or
break a product
Scoring Methodology
Info-Tech scored each vendor‟s features
offering as a summation of their
individual scores across the listed
advanced features. Vendors were given
one point for each feature the product
inherently provided. Some categories
were scored on a more granular scale
with vendors receiving half points.
Advanced Features
Feature
What We Looked For
Real Time Integration
Ability to trigger integration processes in
near real time as data changes.
Data
Cleaning/Cleansing/
Quality
Data profiling, cleansing and reconciliation
across multiple sources, and subsequent
monitoring of new data creation.
Recovery of
Integration after
Failure
Ability to successfully rollback a
transaction, regardless of size or
distribution.
Performance
Monitoring
Ability to see performance metrics of
executing processes without degradation.
Middleware
Compatibility
Compatibility with industry leading data,
application and messaging middleware.
Data
Semantics/Context
Ability to resolve semantic and context
conflicts between numerous data sources.
Synchronization and
Data Replication
Ability to accurately reflect data changes in
data across multiple data stores.
Info-Tech Research Group
10
Each vendor offers a different feature set; concentrate on what
you need
Real Time
Integration
Data Quality
Recovery
after Failure
Perf.
Monitoring
Middleware
Compatible
Data
Semantics
Synchronize
Replicate
SAS/DataFlux
IBM
Informatica
Microsoft
Oracle
Pervasive
SAP
Talend
Legend
= Feature fully present
= Feature partially present / pending
= Feature unsatisfactory
Info-Tech Research Group
11
Informatica is a focused data integration vendor that offers a
deep product portfolio
Overview
Champion
Product:
Employees:
Headquarters:
Website:
Founded:
Presence:
Informatica PowerCenter
2,550+
Redwood City, CA
informatica.com
1993
NASDAQ: INFA
FY11 Revenue: $783.3M
• Unified data integration platform designed to access, integrate,
and manage any type of data on any processing platform.
Utilizes a set of data mapping, ETL, and information life-cycle
management tools that extend and scale to enterprise needs.
Strengths
• Newest release focuses on Big Data, MDM, data quality and
self-service tools. Features include support to pull in data from
social network feeds such as Twitter, Facebook and LinkedIn,
and a universal connector to the Hadoop file system.
• Cloud offering provides the ability to move integration
workflows from Cloud to on-premise with no redevelopment
required.
Challenges
$1
$1M+
3 Year TCO: Priced between $100K and $250K
• Platform historically focused on power users; Big Data and
complex integrations require the on-premise product rather
than the cloud offering.
• The social media and Big Data connectors are sold separately
from the core platform.
Info-Tech Recommends:
The Informatica platform is a good fit for enterprises looking for an efficient, reliable, and easy to
manage data integration solution.
Info-Tech Research Group
12
IBM InfoSphere Information Server marches into the spotlight
with a solid product offering
Overview
Champion
Product:
Employees:
Headquarters:
Website:
Founded:
Presence:
InfoSphere Information Server
427,000
Armonk, NY
ibm.com
1911
NASDAQ: IBM
FY10 Revenue: $99.9B
• Comprehensive platform that supports key enterprise
initiatives, such as warehousing, Big Data, MDM, and
information governance, and provides the ability to cleanse,
transform, and deliver information for immediate action.
Strengths
• SaaS offering through Amazon EC2.
• Metadata driven platform provides information in real-time, on
demand, or in batch to the data warehouse for immediate
action regardless of location.
• Advanced message handling supporting complex XML.
• BI tools provide powerful data analysis for competitive insights
and advantage.
Challenges
$1
$1M+
3 Year TCO: Priced between $250K and $500K
• InfoSphere Information Server is a family of products that has
been integrated into a unified platform. However, each product
in the family may be sold separately, and may also require
additional installation and configuration to enable its
functionality.
• Difficult learning curve for larger, more complex integrations.
Info-Tech Recommends:
InfoSphere Information Sever offers great integration functionality and impressive extensibility into
other data domains.
Info-Tech Research Group
13
SAP is an established player in the industry and they offer a
solid product platform
Overview
Champion
Product:
Employees:
Headquarters:
Website:
Founded:
Presence:
BusinessObjects Data
Services
54,000
Walldorf, Germany
sap.com
1972
NYSE, TecDAX: SAP
FY10 Revenue: $16.8B
• Agile and intuitive environment that combines all data
integration and quality requirements, including development,
metadata and ETL functionality, and is fully web-services
enabled to support an SOA architecture.
Strengths
• Intuitive, codeless, drag-and-drop IDE rapidly develops
integration projects with the option to include data quality.
• Transfer large quantities of data using parallelism, caching,
and grid computing approaches.
• Trending and advanced text data processing analysis features.
• Powerful data quality capabilities and support for unstructured
data.
Challenges
$1
$1M+
3 Year TCO: Priced between $50K and $100K
• The SAP data integration technology was acquired from an
independent software vendor. Therefore, its roots are not SAP
centric and can be effective in non-SAP environments. Look to
the vendor to demonstrate examples of the technology
working effectively in non-SAP environments.
Info-Tech Recommends:
BusinessObjects Data Integrator is an accessible, functional solution geared towards the SMB
market.
Info-Tech Research Group
14
Pervasive Data Integrator offers impressive price and
performance, especially for SMBs
Overview
Innovator
Product:
Employees:
Headquarters:
Website:
Founded:
Presence:
Data Integrator
250
Austin, TX
pervasive.com
1996
NASDAQ: PVSW
FY10 Revenue: $47.2M
• Mature integration suite that provides transformation and flow
of almost any kind of data between sources throughout the
organization – on a continuous, event-driven, or scheduled
basis in a full range of usage scenarios.
Strengths
• Offers intranet SaaS and a broad offering of connectors.
• Highly configurable. Design and deploy on-premises or in the
cloud using reusable metadata stored in an open, XML-based
design repository.
• IDE interfaces are feature-rich and easy to use and administer.
• Galaxy – exchange marketplace for previously custom only
integration products.
Challenges
$1
$1M+
• Platform is not as robust for high performance requirements
that come with large, complex projects, however it is ideal for
smaller companies that have smaller-scale data integration
needs.
3 Year TCO: Priced between $250K and $500K
Info-Tech Recommends:
Pervasive‟s product is a strong offering. While the suite is not as hearty as other offerings, Galaxy and
Data Integrator‟s add-on library are strong selling points for SMBs.
Info-Tech Research Group
15
Talend is ideal for enterprises that need to make the DI tool
business case with little or no budget
Overview
Innovator
Product:
Employees:
Headquarters:
Website:
Founded:
Presence:
Integration Solutions
400
Los Altos, CA & Suresnes,
France
talend.com
2005
Privately held company
• Subscription based, real-time, open source, integration
platform that uses ETL functionality for data integration,
cleansing, migration and synchronization, MDM, and business
intelligence.
Strengths
• Talend Cloud supports all forms of cloud-based computing.
• A real-time data integration platform that supports multi-user
development and employs a unified data system for seamless
integration.
• Advanced event-based scheduling, additional execution
features, error recovery management, and a full set of
connectivity adapters.
Challenges
$1
$1M+
3 Year TCO: Priced between $100K and $250K
• Doesn‟t have the bells & whistles of the larger offerings, but is
a viable option for smaller companies with repeatable DI tasks
who are constrained by budget.
• Graphical User Interface is written in “developer-speak” so
there is a slight learning curve for less experienced users.
• No hosted offering, but a full set of connectors is included.
Info-Tech Recommends:
Talend is an open-source, scalable product that can handle data integration requirements in
organizations of all sizes.
Info-Tech Research Group
16
Oracle Data Integrator is a comprehensive solution, especially
when used to populate large data warehouses
Overview
Market Pillar
Product:
Employees:
Headquarters:
Website:
Founded:
Presence:
Data Integrator
108,000
Redwood Shores, CA
oracle.com
1977
NASDAQ: ORCL
FY10 Revenue: $2.99B
• Unified solution that encompasses all data integration
requirements: from high-volume, high-performance batch
loads, to event-driven, real-time integration processes, to
SOA-enabled data services.
Strengths
• Advanced debugging, diagnoses, and error reporting.
• Open, integrated ETL architecture delivers high-performance
data movement and transformation across complex systems.
• Seamlessly integrates within an SOA infrastructure.
• Enhanced connectivity to all third-party databases, data
warehouses and applications.
• IDE contains mapping wizards and integrates with JDeveloper.
Challenges
$1
$1M+
Vendor Declined to Provide Pricing
• A comprehensive enterprise-level solution, but not designed
for smaller organizations that do not have high-level
integration needs.
• Not as viable for non-Oracle shops; the platform works best
when used with other Oracle-supplied components.
Info-Tech Recommends:
Oracle Data Integrator offers a ton of functionality for Oracle-aligned organizations. The platform is
best suited to those with complex data integration requirements.
Info-Tech Research Group
17
Microsoft SSIS is an obvious choice for SQL Server users, but
it may not reduce hand-coding in the long run
Overview
Market Pillar
Product:
Employees:
Headquarters:
Website:
Founded:
Presence:
SQL Server Integration
Services (SSIS)
90,000
Redmond, WA
microsoft.com/sqlserver
1975
NASDAQ: MSFT
FY10 Revenue: $875.4M
• Offers a fully scalable enterprise-class data integration
platform that includes a high-performance ETL tool that
supports integration and workflow apps, and can extract and
transform data from a wide variety of sources.
Strengths
• Very strong BI functionality, particularly for populating Data
Warehouses and Data Marts.
• High performance, customizable integration functionality that
scales according to need.
• Includes a rich set of built-in tasks and transformations; works
as a stand-alone or in combination with other packages to
address complex integration requests.
Challenges
$1
$1M+
Vendor Declined to Provide Pricing
• There can be compatibility issues in non-Windows
environments.
• Configuration, deployment, and maintenance can be difficult
and time consuming; steep learning curve.
• Robust IDE features drag-and-drop functionality, but the need
for hand-coding has not been eliminated.
Info-Tech Recommends:
For SQL Server owners, SSIS is free, making it the most viable option, but be aware that life may not
be simpler for your hand-coders.
Info-Tech Research Group
18
SAS/DataFlux takes on the industry giants with its unified
data management suite
Overview
Emerging Player
Product:
Employees:
Headquarters:
Website:
Founded:
Presence:
SAS Enterprise Data
Integration Server, DataFlux
Data Management Platform
12,479
Cary, NC
dataflux.com; sas.com
1976
Privately held company
• Single, integrated suite that manages everything from up-front
data discovery, MDM, business process integration, data
governance and federation, and event processing. The
platform encompasses batch, real-time, and virtual integration.
Strengths
• Unique platform that combines DataFlux and SAS
technologies; the framework bridges the gap between data
quality, profiling, monitoring, and ETL tools from a centralized
location.
• Data Management Studio provides extensive functionality for
developing rules and profiles that are reusable in other
projects; interface appeals to IT and business users.
Challenges
$1
$1M+
3 Year TCO: Priced at 1M+*
*Vendor only provided list pricing, no discounts were
incorporated into the pricing scenario .
• Processing large volumes of data can be slow.
• MDM component is fairly new, so that piece of the platform
may not be fully integrated.
• SAS offering provides strong capability for data management
relative to BI, but the DataFlux offering lacks the BI focus.
Info-Tech Recommends:
An excellent choice for organizations that are seeking broad data management functionality.
However, if data integration is your only focus, you may want to look elsewhere.
Info-Tech Research Group
19
Rich Development Environments are key to the effectiveness
of your data integration tool
Usability, debugging, robustness of tools, and developer appeal are key
factors to consider in the decision making process.
1
2
3
4
Exemplary Performers
Development
Environment
Vendor
Comprehensiveness
Platform Independence
Viable Performers
Adequate Performers
Info-Tech Research Group
20
Consider vendor comprehensiveness when evaluating data
integration tools
Are there a number of complementary products that can be used as standalone products and scale to integrate within a unified solution when needed?
1
2
3
4
Exemplary Performers
Development
Environment
Vendor
Comprehensiveness
Platform Independence
Viable Performers
Adequate Performers
Info-Tech Research Group
21
Ensure the vendors offering can be platform independent if
required
Not everyone works in a stack environment. Avoid vendor lock-in by
selecting a solution that is platform, ERP, and middleware independent.
1
Exemplary Performers
Development
Environment
2
3
4
Vendor
Comprehensiveness
Platform Independence
Viable Performers
Adequate Performers
Info-Tech Research Group
22
Data is an investment and every company’s greatest asset; use
DI tooling to leverage that asset to get valuable returns
Don’t just evaluate the financial ROI, but also the Return on Data. Info-Tech
research has concluded that DI tooling has measurable benefits.
DI is not an IT only initiative, it requires
collaboration with business to drive value and ensure
success.
% of Survey respondents that reported
positive impacts of using DI tools
70
60
Using DI Tools Increases:
50
40





Developer effectiveness
Development estimates
Documentation quality
Report accuracy
Data quality
30
20
10
0
Using DI Tools Decreases:
 Data Integration Errors
 Hand Coding
N=117 Source: Info-Tech Research Group
Data and report accuracy are factors that can be improved early in the lifecycle of the tool, whereas code reuse
and development time will improve over time.
Info-Tech Research Group
23
Understand DI Trends and Considerations
What’s in this Section:
• Hand-coding and point-to-point integrations stifle growth.
• DI tools are ready for primetime and address many data
management issues.
• Investigate Data Integration as a Service (DIaaS) if you
have lighter integration needs.
• DI tools help with Big Data, BI and development resources.
• Make the case for budget approval.
• Learn from your peers through case studies.
Sections:
Understand Data Integration
Trends and Considerations
Develop a DI Tool Selection
Strategy
Evaluate DI Vendors
Develop Your DI Implementation
Strategy
Appendices
Info-Tech Research Group
24
Point-to-point integration stifles growth; stop hand-coding
your own demise
Point-to-point integration architectures cannot keep up with business growth
and represent weak links in mission critical integration scenarios.
Rework, quality
problems, manual keying,
and endless rewriting are
needless costs.
Hand-coding is unreliable
and hard to manage
across data architecture
changes.
P2P Interfaces
Tools can ensure your
changes don‟t lead to
data mismatches.
Tools eliminate the need
to rekey, simplify rework,
offer quality and
metadata.
Implement tools to
increase predictability
and data management
maturity.
Tooling Interfaces
200
180
# of Integration Scenarios
One wrong change in
multiple integrations, can
bring down the entire
system.
Avoid
Exponential
Chaos!
160
140
120
100
P2P=n(n-1)/2
80
Tooling=n*2
60
40
20
Simple coding requests
frequently spiral into
tedious, never-ending
tasks.
Automation in tools free
resources from cyclical
development.
0
2
4
6
8
10
12
14
16
18
20
n # of Applications
For an efficient IT shop, the tool is much more versatile and easier to maintain.
- Manager of Enterprise Services, Entertainment
Info-Tech Research Group
25
If you haven’t looked at tools lately you should look again
Tools offer integration functionality beyond traditional ETL*: data federation,
replication, synchronization, and design are increasingly appealing to
business users.
The Problems
Data is everywhere. Data is found across the enterprise in
disparate systems and formats (e.g. transactional systems,
legacy systems, spreadsheets, and flat files).
Big Data and BI insights are useful, but the challenge is
manipulating the volume of data for quality and usability.
Proliferation of isolated solutions, as different teams
create custom code for each integration project. This
degrades data quality, creates redundancy, impairs
reusability, and increases TCO due to the time and effort
needed to support multiple tools.
Mergers & acquisitions require tools to combine data from
disparate systems to eliminate the need to create customcode to extract data from each system.
Legacy systems are outdated and difficult to integrate.
Data integration tools have built-in connectivity capabilities
enabling extraction from many source technologies.
The Solution
Data Integration Tools integrate and consolidate all data
types from internal and external sources into target
destinations such as databases, data-marts, datawarehouses, and files.
Contrary to popular belief, these tools are no longer in
their infancy. A lot of tool functionality is now embedded.
Intuitive user interfaces have made them very easy to
use.
Having a handful of tools to achieve your integration
needs is no longer necessary. They are now
consolidated and platform based so you can pick the
solution that‟s right for you.
*Extract, Transform, Load
Info-Tech Research Group
26
You’re probably integrating anyway, so why not use tools to do
it cheaper, faster, and better
Developers that are embracing tools improve their development process,
which leads to higher quality results, code re-use, and reliable estimates.
Benefits of Tools
•
•
•
•
•
•
•
•
•
•
•
•
Creates consistent approach, simplifying
development and debugging.
No need to re-code to adapt function of
integration.
Manual re-keying not necessary.
Self-documenting.
Accommodates changes in the architecture.
Knowledge is out of developers heads, not lost
with turnover.
On-boards new developers faster.
Create metadata to enable BI.
Easier reporting.
Decreased data quality issues.
Shorter timelines and reduced complexity of
business projects.
Better access to data.
•
•
•
•
•
•
Costs Reduced
Creates consistent approach, simplifying
Maintaining multiple skill sets and technologies.
Redevelopment, rework, integration updates , and
project times.
Debugging and exception handling development
time.
Resource ramp-up time.
Data mapping and modeling time.
Two-thirds of survey respondents
are already using DI tools
Hand-coded
Vendor Solution
N=131 Source: Info-Tech Research Group
Info-Tech Research Group
27
DI tooling improves IT delivery success rates on integration
projects
Pitfalls of Poor Data Integration
Avoid the headaches. A recent Info-Tech survey shows that using DI
tools provides significant success rates in the following areas:
Data integration tools may not be as
versatile as coding by hand, but they
provide a standard that can be easily
picked up by others and remove the
inconsistencies that are often
introduced by hand-coding.
- Info-Tech Survey Respondent
Minimized Hand-Coding
Quality of Documentation
Validity of Development Estimates
Code reusability
Development Process Consistency
0
10
20
30
40
50
% of increased success due to tool usage
60
For hundreds of thousands of lines,
you can’t do it by hand. You’re going
to miss something. If you use tools
it’s doing it all for you, and going to
save you hours of work.
- Data Warehouse Manager,
Distribution Services
N=117 Source: Info-Tech Research Group
Costs
Customer Experience
Marketing
Operations
Hardware costs
Customer churn
Poor decision making
Slow service delivery
Software costs
Reactive decisions
Poor positioning
Slow time-to-market
Operational costs
Poor marketing
Poor investment
Rigid
Opportunity costs
Slow reaction times
Strategy
Labor-intensive
Info-Tech Research Group
28
If you haven’t made an investment in on-premise DI,
investigate DI as a Service as an alternative
DI as a Service (DIaaS) offerings are becoming more robust and offer a quick
ROI to organizations that are looking to give up hand-coding
A number of industry experts are utilizing the acronym
IaaS in reference to Integration as a Service.
To reduce the confusion between Integration as a
Service and Infrastructure as a Service short forms,
Info-Tech refers to Data Integration as a Service as
DIaaS.
Consider DIaaS if you…
 Are adopting SalesForce.com or other SaaS apps
 Have Informal MDM process
 Are intimidated by DI Tools
 Operate in a SaaS-friendly industry
Adopting SaaS introduces the need for integration,
and while custom coding is common and generally
feasible, it can be very resource intensive and difficult
to scale as SaaS applications multiply and more pointto-point connections are demanded.
DIaaS, an on-demand offering itself, addresses SaaS
integration challenges. It is cloud-based and designed
to work with SaaS offerings, significantly reducing
implementation time and overall costs.
Many DIaaS vendors utilize pricing models designed
to scale from small to large organizations, e.g.
providing free development tools and charging
customers on a monthly, per-connection basis, rather
than license a major software package.
Organizations who adopt SaaS solutions without a careful data management and integration strategy will
result in degraded Data Quality. DI as a Service offers the fast path to a solution.
Info-Tech Research Group
29
Big Data is big news; determine how large of an impact it will
have on your organization
Big Data, and how to manage it, is being hailed, or hyped, depending on your
opinion, as THE key IT strategy of the future.
 Substantial smartphone and social network usage,
along with increasing use of video and media files
are creating enormous amounts of structured
and unstructured data.
 Many organizations are finding that they need to
keep more data longer to meet legal and
regulatory compliance, which just adds to the evergrowing pile.
 Enterprises are now facing significant
technology challenges in managing this data
overflow.
 Big Data isn‟t going away. It will continue to impact
the IT world and organizations should begin to
consider the implications of Big Data.
Big Data spans the three following areas:
Variety – Big Data not only includes structured data,
but all types of unstructured data: text, email, audio,
video, click streams, log files and more.
Velocity – Big Data tends to be time sensitive, so in
order to maximize its value, it must be captured as it is
streaming into the enterprise.
Volume – Big Data only has one size: large.
Enterprises are flooded with data, easily accumulating
gigabytes, terabytes and sometimes petabytes of
information.
Data volumes continue to expand exponentially
and the size and scope of this data makes it
difficult to manage with today’s current
technologies.
While the size of Big Data is debatable, one thing is for certain; Big Data is only important to an organization
if it can be utilized to provide results and offer legitimate insights for future direction. Big Data only becomes
relevant when an organization understands what to do with it.
Info-Tech Research Group
30
Decide if Big Data will influence your Data Integration needs
 While Hadoop, Pig, and NoSQL are garnering a lot
of attention, the management of Big Data is still
evolving. However, a number of DI vendors have
begun to incorporate Big Data solutions within their
platforms.
 Blogs, social media, smartphone usage, and
enterprise applications are producing a mountain of
data that, when properly handled and analyzed,
can assist organizations to uncover concealed
opportunities that weren‟t recognized in the past.
 These Big Data insights have to be guided by realtime predictive intelligence in order to be useful.
 However, most IT departments do not currently
employ the right talent to support a Big Data
strategy. If your organization has chosen to go
down the Big Data path, be prepared to hire or
retrain staff to support the skill set that Big Data
requires.
When considering DI technology for Big Data,
consider these factors:
• Do you have a Big Data problem, or merely a
data problem? If you‟ve simply run out of
storage, processing power, backup window, or
network capacity, it may be a classic data
problem. DI tools can help to optimize your
existing data management.
• Have you optimized the data into tiers for
performance? As you optimize storage
performance vs. cost, your DI tools can help
by getting data in and out more quickly.
• Big Data has one critical attribute: it will get
much bigger very quickly. The right DI tool will
support rapidly evolving needs that may be
unforeseen during product selection.
In order to take advantage of the insights hidden within Big Data, IT will create a management architecture
that guarantees effortless data flow between systems. Data integration, middleware and business
process flow technologies will become critical platforms to support this evolution.
Info-Tech Research Group
31
Enterprise Business Intelligence (BI) depends on Data
Integration for quality data
BI is only as useful as the underlying data. With
data coming from numerous sources, data
cleansing has become a primary goal of the data
integration process. IT‟s role is shifting from
“developer of the reports” to “provider of usable and
useful data”. IT departments need to ensure
continual data quality as BI increases its role in
decision making.
Data integration is a core component of any
enterprise BI solution. While a single data source
may contain high quality data, that quality will
degrade when multiple data sets are simply
aggregated. Data Integration tools are key to the
ongoing aggregation of data sources used by
Business Intelligence technologies.
High
BI success is
directly related to
the quality of the
underlying data
BI Success
Organizations are now mining huge data sets and
using BI-driven insights, however BI solutions are
often still constrained by data quality issues.
Low
Low
Overall Data Quality
High
Data Quality is a lack of data conflicts, data duplication,
outdated data, incomplete data, and inaccurate data.
N = 41 Source: Info-Tech Research Group
In today‟s competitive business environment, BI and Analytics are expected to drive business results in
areas where the bottom line can easily be measured.
Info-Tech Research Group
32
Getting budget approval is tough, and it can be even tougher if
the value of the tooling is not fully understood
Successful data integration initiatives are a source of competitive advantage,
allowing enterprises to reduce costs, derive insights and out-perform
competitors.
N=84 Source: Info-Tech Research Group
70
Realizing the business and IT benefits of tools
requires an ability to justify the costs of a data
integration tool to the CFO or budget authority.
Many vendors offer a family of products in which
an organization can start small and grow. This
may result in a lower up front investment and a
better chance to gain budget approval.
If budget approval is a real challenge, consider
looking at DIaaS versions of the vendor‟s product
offerings that may offer a lower cost solution and
subsequently, a better chance of budget approval.
60
% of Respondents
Opportunities are missed for tool adoption
because of failure to make the business case and
gain budget approval.
50
40
30
20
10
0
Gaining
Budget
Approval
Selecting
Vendor
Platform
Tools aren't
Robust
Enough
Developer
Resistance
Budget Approval is the Highest
Friction Point to DI Adoption
Organizations with poor data quality have difficulty getting budget approval. Integrating bad data results in more
bad data, and the cost of cleaning up data before attempting to integrate it is prohibitive. DI tools can help
organizations improve their data quality and therefore should be a worthwhile investment where this is an issue.
Info-Tech Research Group
33
DI tools pay for themselves by freeing up development
resources
Hard costs won’t always justify tool purchases. But soft savings from a
faster, easier process often validate the decision.
Using a tool does not mean firing developers.
Developers can often be used for more directly
beneficial business functions when Data Integration
costs go down.
Data Integration tools can reduce the skill set
requirements for developers, making it possible to
reduce staffing costs over time.
Our cost is in the six figures annually just for
the integration point. It’s excessive, there’s no
question about it.
- CIO, Construction
Look beyond resources; tools offer hard cost savings
from decreased operational development and
maintenance costs.
The cost of licensing a tool could be anywhere from
zero* to $500,000 depending on the product and your
needs.
For the mid-market, several recommended options are
available in the $30-$50,000 range.
This does not include:
 Implementation costs
 Support - 10-15% of licensing cost
 Developer training - $2000/developer
 Purchase of extra application licenses
 Any new hardware required
Often at six-figure salaries, developers can cost you more than many tools on the market today.
*Talend offers a free open-source data integration tool, though Info-Tech does not recommend that version
for enterprise use.
Info-Tech Research Group
34
Case Study: Thirty days is great, but how about twenty
minutes
Mid-sized insurance firm selecting a data integration tool to
decrease business order time.
One of the findings was if we wanted
to just keep our company where it is
today then we don’t need to do
anything. If you want to grow, then
we have to do something.
- IT Director, Insurance
Industry: Financial Services
Segment: Insurance
Situation
•
•
Enterprise applications,
including Salesforce.com are
integrated through hand-coding,
resulting in high development
times and effort.
Need an FTE to run every
integration batch manually.
Action
•
Brought in consultants to
document the business case for
integration tools.
•
Presented case to management
and developers, gaining full
support.
Results
•
Business order times will be
reduced to 20 minutes from 30
days because tooling allowed
for a more flexible architecture
and near real-time integration
automation.
•
Client Website will be
dramatically more useable.
•
Eliminated one FTE, saving
over $200,000 annually.
Info-Tech Research Group
35
Case Study: DI tool as the enabler of a virtual Master Data
Management strategy
Master Data Management is beyond the reach of many
companies because of cost, but DI can enable a simple
solution.
Industry: Manufacturing
Segment: Office Furnishings
Situation: Unreliable Data
Action: Virtual MDM
• A mid-sized manufacturer of
office furnishings had unreliable
analytics after their CRM system
was migrated to a cloud-based
solution.
• The IT department decided on a
“Virtual Master Data
Management” strategy based on
their Data Integration tool.
• To encourage adoption of the
new system, UI edit rules were
relaxed and restrictions on
account record creation were
lifted.
• The rules-driven solution uses a
registry model for MDM as the
DI tool dynamically queries
multiple data sources to create a
real time high-quality data
source.
• The relationships between
customer, financial, and product
data became unstable.
• A small amount of human
intervention is required daily to
resolve data conflicts.
Results: Improved BI
• After a relatively small
development effort, the IT
department produced a reliable
analytics data source for
marketing and product
development users.
• People throughout the
organization did not have to
change their „silo‟ mentality. The
CRM and Supply Chain
Management systems stayed
as-is, optimized for their users
rather than for data quality.
Info-Tech Research Group
36
Employ DI Tools to measurably improve efficiency and
operational success
Success divides into two main statistical factors, but you can skip the heavy
math. Efficiency and operational factors are significantly improved when
integration tools are adopted.
Impact of Data Integration Tools
Efficiency Success
Operational Success
Development Time
MDM Structure
Code Reusability
Process Predictability
Process Efficiency
Data Accuracy
Versatility
Cost Effectiveness
Process Satisfaction
Solution Satisfaction
Compared to organizations that handcode, organizations that use tools are
10% more successful when it comes
to efficiency of data integration
Organizations using tools rather than
hand-coding for data integration are
11% more successful when it comes
to operational success
N=63 Source: Info-Tech Research Group
Info-Tech Research Group
37
Develop a DI Tool Selection Strategy
What’s in this Section:
•
•
•
•
Decide if DI tools are right for you.
Determine which features you need.
Decide on your licensing model.
Select the right vendor for appropriate tool functionality.
Sections:
Understand Data Integration
Trends and Considerations
Develop a DI Tool Selection
Strategy
Evaluate DI Vendors
Develop Your DI Implementation
Strategy
Appendices
Info-Tech Research Group
38
You’ve heard the pitch, now decide if tools are right for you
Data Integration Tools aren’t for everyone, and within the space there are
many options for many issues. Determine which, if any, are right for you.
Consideration
Build
Buy
Initial Investment
Lower
Higher
Operations Cost
Higher
Lower
Support &
Maintenance
In-house IT
Vendor-managed
In-house IT Staff Skill
Requires high
maturity
Less required
Data Cleansing
Limited
Often included
Metadata Capture &
Impact Analysis
Limited
Often included
Source-to-Target
Connections
Single/Single
only
Multiple/Multiple,
Multiple/Single,
Single/Multiple
Data Source Changes
IT effort
Vendor-managed
Complex
Transformations
Limited,
taxing
Comprehensive,
supports user code
One-Time Integrations
Ideal
Overkill
Based on your answers to a series of questions about
your environment, the Info-Tech Data Integration Tool
Appropriateness Assessment will recommend an
appropriate strategy for your organization.
Info-Tech Research Group
39
Determine which data integration features you need
Use…
To…
When…
Data Migration/ Conversion
Tool
Migrate/ convert data in bulk to a new
application.
A source system is renewed or because of
application consolidation.
Batch Integration Tool
Process data changes in bulk at
scheduled intervals.
Loading data marts/ warehouses and application
integration.
Real Time Integration Tool
Update data in multiple systems as it
is changed.
Requirements necessitate data be updated in
real time – operational dashboards, etc.
Metadata Management Tool
Track data lineage (source, history,
transformation).
Transforming/consolidating data from multiple
sources into a data store, (warehouse or mart).
Data Quality Tool
Cleanse and consolidate conflicting
data.
Cleansing/consolidating within the same
application or as part of a transformation routine.
Data Modeler
Create well-formed data structures
that respect integrity.
Building/revising a data architecture for an
application or enterprise structure.
Selection Tips
•
Most data integration tool
suites encompass all listed
features to varying degrees.
Selection Tips
•
Use your feature requirements
as a guide as you develop
your data integration tool
strategy.
Selection Tips
• Purchasing a full suite may be
more costly in the short term but
the level of support and
functionality is higher than
with cheap and easy tools.
Info-Tech Research Group
40
Four factors govern your data integration tool selection
strategy
Developing a strategy is an art, not a science; use these factors to direct your
selection, but don’t expect any vendor to deliver a silver bullet.
Answer…
So you can…
How well do you understand your
enterprise‟s data flow?
Determine which tools will be
comfortable for your current level
or take you to the next one.
Are you using point-to-point
integrations or have you moved
to a hub?
Determine which tools help you
develop your architecture, or just
make your life easier.
Functionality
The data integration tool umbrella
encompasses many features.
Which do you need?
Determine which tools offer you
the „need to have‟, not the „nice
to have‟ features.
Number of
Developers
How many developers will be
working with the data integration
tools?
Determine which licensing
models work best for your
enterprise.
Information
Architecture
Maturity
Data Architecture
Info-Tech Research Group
41
When you are ready to take a more holistic approach to
integration… use tools
DI tool adoption should be a function of information architecture maturity.
If you have any of the following scenarios, tools are for you.
Information
Lite
Enterprise
Situation
Informal
Maturity
No MDM processes or
tools are in place. Data
integrations are fully
point-to-point.
Data flow is understood
and documented but no
formal system exists.
MDM is part of a BI, DI,
or EII suite. Data is
captured when moving
into a BI report or
warehouse.
Enterprise data modeling
– every piece of data is
accounted for in a data
dictionary.
Next Steps
Nothing
Architecture
Start developing an
understanding of your
data integration needs
before jumping into
tools.
Pick an architecture
based on your
processing load
bottleneck and select
a tool to support it.
Use tools to achieve
application integration
before it hits central hub.
You are likely already
using advanced DI tools.
Good job!
Data Integration is necessary for providing consolidated and accurate data for decision making. Without
integration tools, MDM and BI initiatives will fail.
Info-Tech Research Group
42
Use your developer team size to pick the licensing model that’s
right for you
Lower numbers of developers may be less costly to license by seat than a
server/processor based licensing model
1. Pick the Right License
Per Developer Licensing:
Tools designed for smaller shops typically use per
developer licensing. All else equal, these are
typically most cost effective for shops with four or
less integration developers.
Per Processor Licensing:
Stack tools, typically targeting enterprise clients, license
per processor. Measures of enterprise size aside, these
typically become cost-effective if they will be used by
more than four integration developers.
If up-front costs are too high to make tool
use feasible, don’t give up!
Check out DI as a Service
2. Minimize Tool Count
Having a handful of tools to achieve your integration needs
is no longer necessary. They are now consolidated and
platform based so you can pick the stack that‟s right for
you.
Standardize and consolidate tools to drive down costs.
Paying for software licenses and support for multiple tools
is much less cost effective than purchasing a robust tool or
suite.
Consider what tools offer in terms of price-to-performance.
It is better to invest in one flexible tool than purchase
several as needed, which inflates software costs.
With market consolidation, a single tool may
now be used for data integration, migration, data
quality, and even MDM.
Info-Tech Research Group
43
Not all vendors are created equal; pick the right one for your
circumstances
70
Effectiveness is highly vendor dependent
Reduction in hand-coding is also variable, with some
tools failing to provide a significant reduction at all.
60
% of Respondents
While tool satisfaction is uniformly above handcoding, it ranges dramatically across stack vendors.
Primary Inhibitors of DI Tool Success
50
40
30
20
10
0
A recent Info-Tech survey indicates that almost 50% of
respondents cite selecting a standard vendor platform as
one of the primary inhibitors of DI tool success.
Gaining Budget
Approval
Selecting
Vendor
Platform
Tools aren't
Robust Enough
Developer
Resistance
N=84 Source: Info-Tech Research Group
I want…
Info-Tech Recommends
To reduce hand-coding without abandoning point-to-point
Informatica Cloud
To get tools as cheaply as possible
Talend, Informatica Cloud, Pervasive Data Integrator
Stick with a stack vendor
SAP BusinessObjects Data Integrator, Oracle Data
Integrator, IBM InfoSphere, Microsoft SSIS
Integration with a comprehensive data management platform
DataFlux , IBM InfoSphere
To try before I buy to help make the business case
Talend
To improve my SQL Server data integration
Microsoft SSIS
Info-Tech Research Group
44
Evaluate Data Integration Tools Vendors
What’s in this Section:
• Info-Tech‟s Vendor Landscape for eight DI Tool vendors.
• Shortlisting DI Tool vendors through scenario analysis.
• Developing and executing a DI Tool RFP.
Sections:
Understand Data Integration
Trends and Considerations
Develop a DI Tool Selection
Strategy
Evaluate DI Vendors
Develop Your DI Implementation
Strategy
Appendices
Info-Tech Research Group
45
Market Overview
How it got here
• Data integration continues to be an evolving practice even
though it came into existence more than 30 years ago, when
computing became distributed. The objective of data
integration is to collect data from numerous, different sources,
merge it, and display it to the user in such a way that it
appears to be a unified whole.
• The need for data integration arose from mergers,
acquisitions, and organizations maintaining numerous
databases and applications that housed different data
aspects necessary to business operations. For many years,
organizations tried to consolidate their data into one system
by relying on developers to hand-code scripts to connect the
various sources, but that practice often resulted in quality
issues, lengthy development times, and unreliability – one
error in a line of code can bring an entire system down.
• Data integration isn't just a component of an IT project
anymore. It has become a core practice that requires
consideration from the start if an organization is going to
make the most use of their data.
Where it’s going
• Data integration tools have become a hot commodity in
recent years. With BI and Big Data analysis moving to the
forefront in organizational strategy, DI tools are becoming
geared towards business stakeholders as well as IT
developers. Rich development environments, reporting
analytics, and efficient data management are becoming
key factors in the design of DI tools.
• The integration area currently consists of middleware, data
and process integration. As DI solutions move towards
becoming a comprehensive, unified platform, the lines
distinguishing these areas will become increasingly
blurred.
• In the future, as data integration processes evolve to meet
the growing needs of the organization, they will require the
capacity to accommodate a broader range of data types.
Increasing use of SaaS will require DI to reside in the
cloud. Organizations wishing to capture social networking
data will require their DI solutions to integrate with various
media feeds such as Twitter and Facebook.
As the market evolves, capabilities that were once cutting-edge become default, and new functionality
becomes differentiating. Batch integration has become a Table Stakes capability and should no longer be
used to differentiate solutions. Instead, focus on data quality and real-time integration to get the best fit for
your requirements.
Info-Tech Research Group
46
DI tools Vendor Landscape selection/knock-out criteria:
market share, mind share, and platform unification
• Until recently, the data integration market was a fairly conservative space with most solutions only offering ETL
functionality. However, changes to vendor offerings, with the inclusion of data quality (profiling, analysis and
cleansing), synchronization, and MDM are bringing about a movement towards unified platforms in DI technologies.
For this Vendor Landscape, Info-Tech focused on those vendors that have a strong market presence and/or reputational
presence among small to mid-sized enterprises.
•
Included in the Vendor Landscape:
•
SAS/DataFlux: The platform is a unified design, development and execution DI solution that enables data
quality, integration and master data management (MDM) from a single interface.
•
IBM: InfoSphere Information Server is a comprehensive platform that provides seamless data integration to
support initiatives across MDM, data warehousing, Big Data and migration projects.
•
Informatica: An inclusive, unified, open software platform to access, integrate, and manage any type of data on
any processing platform.
•
Microsoft: SQL Server Integration Services (SSIS) platform features a flexible data warehousing tool that aids
in data integration and workflow applications.
•
Oracle: Data Integrator is a broad integration platform that addresses all data integration requirements,
including seamless batch and real-time integration and data warehousing.
•
Pervasive: Data Integrator is a highly configurable integration platform that performs extraction,
transformation and flow of nearly any kind of data between sources throughout the organization.
•
SAP: BusinessObjects allows the organization to profile, extract, transform, and move data in real time and at
any interval anywhere across the enterprise.
•
Talend: Provides open source, subscription based data integration solutions designed for business intelligence
and data migration and synchronization.
Info-Tech Research Group
47
Data Integration Tools Criteria & Weighting Factors
Product Evaluation
Features
Features
The solution provides basic
and advanced feature/functionality.
Affordability
The five year TCO of the solution is
economical.
Usability
Architecture
Architecture
30%
30%
20%
The solution‟s dashboard and reporting tools
are intuitive and easy to use.
Affordability
Product
50%
The delivery method of the solution aligns with
what is expected within the space.
Vendor Evaluation
50%
Viability
Vendor is profitable, knowledgeable, and will
be around for the long-term.
Strategy
Vendor is committed to the space and has a
future product and portfolio roadmap.
Reach
Vendor offers global coverage and is able to
sell and provide post-sales support.
Channel
20%
Usability
Vendor channel strategy is appropriate and the
channels themselves are strong.
Vendor
Viability
25%
30%
Strategy
15%
Channel
30%
Reach
Info-Tech Research Group
48
Identify leading candidates with the Data Integration Tool
Vendor Shortlist tool
The Info-Tech Data Integration Tool Vendor Shortlist tool is designed to
generate a customized shortlist of vendors based on your key priorities.
This tool offers the ability to modify:
• Overall Vendor vs. Product Weightings
• Top-level weighting of product vs. vendor
criteria
• Individual product criteria weightings:
 Features
 Usability
 Affordability
 Architecture
Custom Vendor Landscape™ and Vendor Shortlist
Your customized Vendor Shortlist is sorted based on the priorities identified on the Data Entry tab. Scores are calculated using
the Client Weightings and the assigned Info-Tech Vendor Landscape scores. Vendors are ranked based on the computed
Average Score. The Average Score is the average of the weighted average Vendor Score and the weighted average Product
Score. A custom Vendor Landscape™ has been generated as well, plotting the weighted average Vendor Score against the
weighted average Product Score.
Custom Vendor Landscape™ for [Enterprise Name Here]
Informatica
• Individual vendor criteria weightings:
 Viability
 Strategy
 Reach
 Channel
IBM
Talend
Pervasive
SAP
Oracle
SAS/DataFlux
Microsoft
Info-Tech Research Group
49
Issue an RFP to ensure that Data Integration Tools vendors fit
your needs, and not the other way around
Use Info-Tech’s Data Integration Solution RFP Template to conduct this
critical step in your vendor selection process.
Info-Tech‟s DI Tools RFP Template is
populated with critical elements, including:
 The Statement of Work
 Proposal Preparation Instructions
 Scope of Work
 Functional Requirements
 Technical Specifications
 Operations & Support
 Sizing & Implementation
 Vendor Qualifications & References
 Budget & Estimated Pricing
 Vendor Certification
An RFP implies stable requirements and an intent to buy – use this tool to help select a supplier, not to
develop a shortlist.
Info-Tech Research Group
50
To get the most value out of the RFP process, use the
Data Integration Tools RFP Scoring Tool
A standard and transparent process for scoring individual vendor RFP
responses will help ensure that internal team biases are minimized.
Use Info-Tech‟s
DI Tools RFP Scoring Tool to:
The Info-Tech Data Integration Solution
Evaluation & RFP Response Tool comes
pre-built with important scoring criteria for
vendor RFP responses.
This tool includes modifiable criteria across
the following categories:
•
Features (real-time integration)
•
Operational Requirements
(debugging, exception reporting)
•
Architecture (hosted deployment,
connector volume)
•
Support
Adjust the individual category weightings to
customize this tool to business priorities.
Info-Tech Research Group
51
Take charge of vendor finalist demonstrations with a
Vendor Demo Script
An on-site product demonstration will help enterprise decision-makers better
understand the capabilities and constraints of various solutions.
This tool is designed to provide vendors
with a consistent set of instructions for
demonstrating key scenarios for the DI
tools implementation.
The Info-Tech Data Integration Vendor
Demonstration Script is designed to provide
vendors with a consistent set of instructions
for key scenarios.
This template includes examples for the
following scenarios:
•
Planning & Deployment
•
Meeting Setup & Operation
•
Software Installation
•
Initial Configuration
•
Diagnose/Repair/Uninstall
Adjust the individual category weightings to
customize this tool to business priorities.
Info-Tech Research Group
52
Develop Your DI Tool Implementation Strategy
What’s in this Section:
• Include people, process, and technology in your
implementation preparation.
•
•
•
•
Sections:
Understand Data Integration
Trends and Considerations
Determine which tool attributes will drive your tool solution.
Develop a DI Tool Selection
Strategy
Develop your timeline in preparation for integration.
Evaluate DI Vendors
Introduce training early to avoid resistance.
Develop Your DI
Implementation Strategy
Align tool deployment with existing architecture.
Appendices
Info-Tech Research Group
53
Prepare for implementation to ensure a smooth process
1. Manage the Technology
Flexibility is King
Where possible, reduce future costs by
selecting a tool that is reusable and
replace other single-function tools to
minimize licensing and support. Consider
present and future needs.
When possible, automate
The closer your tool aligns with your
desired functionality, the less hand-coding
you will have to do, so capturing precise
technical requirements is critical. Properly
chosen, data integration tools are flexible
and can lend themselves to complex
integration scenarios.
Prune data first
2. Don’t Forget People & Processes
Integration size, complexity,
and available internal staff
determine consulting need.
For straightforward implementations,
most enterprises should be able to
proceed with their own staff or
resources provided by their selected
vendor or channel partners.
Complex, large scale integrations, or
enterprises with very limited IT staff
may benefit from introducing
consultants as soon as possible.
Info-Tech Insight
The goal of any data
integration tool should
be to make source
data systems plug n’
play. Keep this in mind
when developing your
implementation
process, as processes
that don’t enable this
will require on-going
management, limiting
cost and time savings,
and causing additional
grief.
Storing duplicate and unneeded data is
expensive; the cost to integrate or
migrate it is worse. Remove this data
before integration to drive down costs
and the timeframe.
Info-Tech Research Group
54
Look to your existing data architecture to direct your tool
implementation architecture
DI tools fit into multiple data architectures. Align your tool deployment with
your existing data architecture to derive maximum benefit.
Next Steps
Situation
011001
010010
100010
110
Point-to-Point
Tightly coupled
connections between
software applications.
Start with lighter
solutions, such as DI as
a Service options to
streamline the process.
Hub and Spoke
Centralized architecture
that uses a data store
(hub) to populate
systems (spoke).
Federated
The data store does not
actually store any data –
only rules to direct and
coordinate data between
systems.
Distributed
Uses agents positioned
between the system and
data store to reduce
system processing load.
Most DI Tools support all three structures. If you are looking to upgrade to
one of the structures, tools will enable and simplify this process. The decision to
pick a hub, rules engine, or intermediaries will depend on where in your system
you would like to process load.
Info-Tech Research Group
55
Not all DI projects are created equal, but they share common
attributes that may drive different solutions
Real-time or
Batch?
Though real-time and online integration is supplanting batch-oriented, this is not always
appropriate and can significantly increase costs and bog down performance. Unless upto-date real-time data will drive business value, batch integration remains entirely
appropriate.
How do I address
huge data flow?
More advanced tools may provide parallel processing and workload balancing. Both
are useful for enterprises with large data volumes and overcome performance
issues stemming from bottlenecks in peak frequency times.
What about Data
Quality?
Don’t overlook data quality; it is key to data integration success, as integrating bad
data is ineffective. Many offerings include this functionality directly or through a third
party extension; in most cases the difference is negligible. Keep in mind the ratio of costto-effectiveness, as 100% accuracy is rarely worth the investment.
What’s Impact
Analysis?
When do I
Debug?
Impact analysis (i.e. tracing the impact of source data changes on reports or analysis)
requires metadata integration and data lineage and allows for development of „what-if‟
scenarios. This functionality is recommended for enterprises with complex
integration architectures. (e.g. a target field is derived from multiple source fields).
Debugging environments, often included, are useful for error-checking when
developing integrations, especially if staffing less experienced developers.
Info-Tech Research Group
56
Your timeline length depends on your integration, but the
steps are the same
Vendors setup takes between 1 day and 1 month depending on your
environment. Ensure you are internally prepared for integration.
Implement
Operate
If you are purchasing new hardware systems in
conjunction with your tool, begin by mapping and
connecting the new physical infrastructure before worrying
about the tools themselves.
Tools that will run on existing software typically feature
straightforward installations for the agent and
demonstration environment.
Prioritize integrations.
Start automating crucial integrations that are deemed
mission critical to the business. If you are using legacy
systems, start there, as these integrations are in most
urgent need of replacing.
Updating other lower priority integration points should be
undertaken as a subsequent phase.
Start Integrating Data with a 5 Step Process:
1.
Implement
Quality Control
2.
Create
Connections
3.
Automate
Integrations
4.
Test
Procedures
5.
Execute
Plan accordingly: More complex integrations will have a lower reduction in development time.
Hand-coded Integration: 8 weeks
Hand-coded Integration: 24 weeks
Complex
Simple
With Tools: 8 days – 80% reduction
With Tools: 9 weeks – 62% reduction
Info-Tech Research Group
57
Avoid the pitfalls of using DI as a Service and integrating SaaS
applications
SaaS vendors offer professional services to assist with data transfer, which
may be helpful for more complex integration projects.
Know When to Create a Custom Adapter
Some enterprises may elect to write their own connector
to facilitate bulk uploads. There are two triggers for this
scenario:
1.
2.
Real-time integration with other
systems. Some enterprises may require near
real-time integration with other applications such
as ERP, marketing automation, or customer
service applications.
Excessive scale and complexity. Third-party
applications for bulk uploads will baulk with
transfers characterized by:
Over 200,000 records.
Over 1,000 data fields.
Most large SaaS vendors run off a single database
instance, which, while easier to maintain, presents
customer challenges:
It takes time. Moving batches of millions of
records takes a considerable amount of time.
Tweak the interface to improve performance.
Persistent connections are a must. Negotiating
thousands or millions of individual connections for
a batch transfer will cripple performance. Use the
persistent connection feature.
Multiple threads don't help. Starting multiple
sessions to transmit batched data will not help
performance due to authentication complications
and may even slow down performance.
Lock out unnecessary features. During batch
updates some usability features will constantly
update, leading to poor performance.
Overcoming negative SaaS perceptions is still an issue. When surveyed, 61% of Info-Tech respondents
believed SaaS to be more difficult to integrate than on-premise applications.
Info-Tech Research Group
58
Technology is rarely the issue that causes integration projects
to fail; effective scoping & training will promote success
Before diving in, scope out the project
thoroughly:
Integration projects usually involve multiple
stakeholders:
 There could be one or more stakeholder for each
database that is being integrated.
 An owner of a database and/or application may be
hesitant to allow updates from an external process.
Don’t boil the ocean:
 Integrating multiple systems is best achieved in small
steps.
 Smaller, discrete scoping can limit the number of
stakeholders that need to be involved. Too many
cooks in the kitchen can spoil the dinner.
Make sure stakeholders are engaged early in the
project and get their buy-in:
 Leverage management resources as necessary to get
stakeholders on side.
Getting stakeholders involved in the testing early and
often can help add to the success of the project:
 Use an iterative approach to test early and test often
during implementation.
Introduce training early to avoid user
resistance:
 DI technologies continue to evolve. As many IT staffers
have relied on hand-coding, DI technologies may not
be universally understood or accepted.
 Business requirements drive data needs. Review the
internal infrastructure. Realize that not every system will
need to be integrated and integration will be dependent
upon the scope of the business and data requirements.
 Develop the internal culture for project acceptance,
include non-IT stakeholders to cover off all areas of data
usage.
 Understand that there may be major shifts to be made
in terms of traditional data management practices
before DI technologies can be successfully implemented
within an organization.
 Determine if there is a major gap in skills prior to
execution, as data integration requires staff with very
specific skills sets. Prepare to hire or retrain staff if that
gap exists.
Info-Tech Research Group
59
Expect developer resistance and overcome it through dialog
Developers want to Develop
Tools are Friendly
The market is moving away from custom coding but
some stubborn developers continue to hold their
organizations back.
Managers don‟t train developers on data integration
tools. Address resistance through developer
managers.
Tools are strongly resisted by developers who feel the
tools take away control and freedom, yet the tools are
essential for data management maturity.
Don‟t expect resistance to be demographically uniform,
or even logical, e.g. newer developers could expect
tools to be in place or fiercely resist them, feeling the
need to prove themselves through coding.
Developers will say the tool isn‟t robust enough which
comes from insecurity and fear of being replaced.
Programming is a part of the job they like doing so they
don‟t want to reduce it. They think tools mean they just
do data entry.
Yet system mapping and knowledge is more valuable.
They need to expand their thinking of what it means to
be a developer.
Regardless of developer reasoning, the key to
convincing developers is process: show them how
tools will make their lives easier and they will be
wowed.
I think getting a third party, a second voice that
they’re not used to hearing all the time, saying the
same thing in different words,
I think that made a significant impact.
- IT Director, Insurance
Research has shown that developer resistance is fading. Developers are finding that the tooling improves
their development processes, estimates, and outcomes. It is also freeing them up for more innovative
activities.
Info-Tech Research Group
60
Summary
At some point data integration tools become necessary for business growth.
They improve overall operations, marketing, and customer experience.
• Tools now provide broader functionality beyond traditional ETL and measurably improve efficiency and operational
success.
• DI as a Service (DIaaS) offerings are a viable option for organizations that are beginning to look at data integration and
have lighter integration needs.
• Big Data, and how to manage it, has become a key IT strategy; however, organizations first need to determine if they
actually have a Big Data problem, or simply a classic data problem.
• The overall value of BI is being recognized, but many solutions are still constrained by data quality issues. IT departments
need to utilize DI tools to ensure continual data quality as BI increases its role in decision making.
• If a tool re-directs at least one developer from maintenance to innovation, it has probably paid for itself.
• Organizations with poor data quality have difficulty getting budget approval. Make the case with the CIO and CFO that
data is an organization‟s best investment, and implementing DI tools will provide significant ROI.
• Developing a data integration tool selection strategy is a function of your information architecture maturity, data
architecture, features needed, and developer staff size.
•
•
•
•
Tool effectiveness is highly vendor-dependent as many tools don‟t actually reduce necessary hand-coding.
Prioritize integration, include people, process and technology. And align tool deployment with existing architecture.
Develop your timeline in preparation for integration, and automate as much as possible.
Get user and stakeholder buy-in by scoping the project thoroughly and providing early training.
Info-Tech Research Group
61
Appendices
What’s in this Section:
• Glossary
• Vendor Landscape methodology
• DI tools survey demographics
Sections:
Understand Data Integration
Trends and Considerations
Develop a DI Tool Selection
Strategy
Evaluate DI Vendors
Develop Your DI Implementation
Strategy
Appendices
Info-Tech Research Group
62
Glossary
Business Intelligence (BI)
Processes and applications that provide a holistic view of business operations that are
used to support decision making.
Business Process
Management (BPM)
The continuous improvement of business processes to align them with customer
needs.
Data Agility
The ability of a piece or group of data to be easily manipulated and adapted as needed
Data Architecture
The design of data, systems and tools that support an organization or business
function.
Enterprise Application
Integration (EAI)
Linking applications together through data integration.
Enterprise Information
Integration (EII)
Creating a single interface to view data from more than one source.
Extract, Transform and Load
(ETL)
The process of extracting data from an outside source, transforming it to meet needs
and loading it into the end target.
Legacy System
Old hardware or software that is still used and contain valuable data.
Master Data Management
(MDM)
Creating a single and inclusive view of data.
Software-as-a-Service (SaaS)
Enables a company to rent an application from a vendor that hosts it.
Service-Oriented Architecture
(SOA)
The process of building scalable distributed systems that treat all software
components as services.
Data Modeling
The process of defining and analyzing enterprise data to better manage data and
improve business processes.
Info-Tech Research Group
63
Vendor Evaluation Methodology
Info-Tech Research Group‟s Vendor Landscape market evaluations are a part of a larger program of vendor evaluations, which includes
Solution Sets that provide both Vendor Landscapes and broader Selection Advice.
From the domain experience of our analysts, as well as through consultation with our clients, a vendor/product shortlist is established.
Product briefings are requested from each of these vendors, asking for information on the company, products, technology, customers,
partners, sales models, and pricing.
Our analysts then score each vendor and product across a variety of categories, on a scale of 0-10 points. The raw scores for each
vendor are then normalized to the other vendors‟ scores to provide a sufficient degree of separation for a meaningful comparison. These
scores are then weighted according to weighting factors that our analysts believe represent the weight that an average client should
apply to each criteria. The weighted scores are then averaged for each of two high level categories: vendor score and product score. A
plot of these two resulting scores is generated to place vendors in one of four categories: Champion, Innovator, Market Pillar, and
Emerging Player.
For a more granular category by category comparison, analysts convert the individual scores (absolute, non-normalized) for each
vendor/product in each evaluated category to a scale of zero to four whereby exceptional performance receives a score of four and poor
performance receives a score of zero. These scores are represented with “Harvey Balls,” ranging from an open circle for a score of zero
to a filled-in circle for a score of four. Harvey Ball scores are indicative of absolute performance by category but are not an exact
correlation to overall performance.
Individual scorecards are then sent to the vendors for factual review, and to ensure no information is under embargo. We will make
corrections where factual errors exist (e.g. pricing, features, technical specifications). We will consider suggestions concerning benefits,
functional quality, value, etc.; however, these suggestions must be validated by feedback from our customers. We do not accept
changes that are not corroborated by actual client experience or wording changes that are purely part of a vendor‟s market messaging or
positioning. Any resulting changes to final scores are then made as needed, before publishing the results to Info-Tech clients.
Vendor Landscapes are refreshed every 12 to 24 months, depending upon the dynamics of each individual market.
Info-Tech Research Group
64
Value Index Ranking Methodology
Info-Tech Research Group‟s Value Index is part of a larger program of vendor evaluations, which includes Solution Sets that
provide both Vendor Landscapes and broader Selection Advice.
The Value Index is an indexed ranking of value per dollar as determined by the raw scores given to each vendor by analysts.
To perform the calculation, Affordability is removed from the Product score and the entire Product category is reweighted to
represent the same proportions. The Product and Vendor scores are then summed, and multiplied by the Affordability raw
score to come up with Value Score. Vendors are then indexed to the highest performing vendor by dividing their score into
that of the highest scorer, resulting in an indexed ranking with a top score of 100 assigned to the leading vendor.
The Value Index calculation is then repeated on the raw score of each category against Affordability, creating a series of
indexes for Features, Usability, Viability, Strategy and Support, with each being indexed against the highest score in that
category. The results for each vendor are displayed in tandem with the average score in each category to provide an idea of
over and under performance.
The Value Index, where applicable, is refreshed every 12 to 24 months, depending upon the dynamics of each individual
market.
Info-Tech Research Group
65
Product Pricing Scenario & Methodology
Info-Tech Research Group provided each vendor with a common pricing scenario to enable normalized scoring of Affordability,
calculation of Value Index rankings, and identification of the appropriate solution pricing tier as displayed on each vendor scorecard.
Vendors were asked to provide list costs for Data Integration Tools and/or licensing to address the needs of a reference organization
described in the pricing scenario.
Additional consulting, deployment, and training services were within the scope of the pricing request, as was the cost of enhanced
support options, though vendors were encouraged to highlight any such items included with the base product acquisition. The annual
maintenance rate was also requested, allowing a three-year total acquisition cost to be calculated for each vendor‟s data integration
solution. This three-year total acquisition cost is the basis of the solution pricing tier indicated for each vendor.
Finally, the vendors‟ three-year total acquisition costs were normalized to produce the Affordability raw scores and calculate Value Index
ratings for each solution.
Key elements of the common pricing scenario provided to Data Integration vendors included:
• Includes development and production, licensing and annual maintenance for 3 years. In addition, any additional benefits that may be
associated with the integration scenario i.e. ancillary licensing and/or training and professional services.
• Company information:
 10,000 employees , 15 locations, 120 IT staff that includes 20 internal developers dedicated to development/integration
• Infrastructure information:
 6 dual processor mission critical application servers located in a central data center and 2 quad processor database servers – 1
runs SQL, the other runs Oracle.
 Separate backup site that includes redundancy for all mission critical solutions.
 Teradata Data Warehouse Appliance
 Salesforce.com licensed for 3000 users within the company
 Other applications in use include: an ERP system, an HR system, a Finance system, Exchange , 5 additional internally
developed software solutions (including the company website, which is hosted internally).
• All 10,000 employees have access to certain components of each of the systems.
Info-Tech Research Group
66
Data Integration Tools Survey Demographics
•
•
•
•
•
Industry
Country
Revenue
Full-time Employees
IT Employees
Info-Tech Research Group
67
Industry
Info-Tech Research Group
68
Country
Info-Tech Research Group
69
Revenue
Info-Tech Research Group
70
Full-time Employees
Info-Tech Research Group
71
IT Employees
Info-Tech Research Group
72