<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>CDAP</title><link>/</link><description>Recent content on CDAP</description><generator>Hugo -- gohugo.io</generator><atom:link href="/index.xml" rel="self" type="application/rss+xml"/><item><title/><link>/community/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/community/</guid><description/></item><item><title>Accelerators</title><link>/accelerators/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/accelerators/</guid><description/></item><item><title>Analytics</title><link>/accelerators/analytics/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/accelerators/analytics/</guid><description/></item><item><title>BI migration</title><link>/case-studies/bi-migration/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/bi-migration/</guid><description>Scenario The customer, a Health Insurance company, was using Netezza to aggregate and report on multiple dimensions on different health care and services. They were looking for alternatives to offload the data from Netezza to reduce the workload for reducing the cost. Data Administrators, Data Analysts, Data Scientist and Engineers were operating and supporting the loading, cleansing and reporting efforts. They were facing a lot of difficulties in the following areas that was hindering their transition into Hadoop:</description></item><item><title>Data lake</title><link>/case-studies/data-lake/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/data-lake/</guid><description>Scenario The customer, a Health Insurance company, was using Netezza to aggregate and report on multiple dimensions on different health care and services. They were looking for alternatives to offload the data from Netezza to reduce the workload for reducing the cost. Data Administrators, Data Analysts, Data Scientist and Engineers were operating and supporting the loading, cleansing and reporting efforts. They were facing a lot of difficulties in the following areas that was hindering their transition into Hadoop:</description></item><item><title>Finance</title><link>/case-studies/finance/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/finance/</guid><description>Scenario The customer, a Fortune 50 financial institution, created a pipeline that aggregates batched data into a secured on-premise cluster to create daily aggregates and reports. The current system performed multiple transformations, which created new datasets. The customer faced multiple issues:
The data pipeline was inefficient, took 6 hours to run, and required manual intervention almost on a daily basis Reports were not aligning correctly with day boundaries. Any points of failure require reconfiguring and restarting the pipeline, a time-consuming and frustrating task.</description></item><item><title>Finance</title><link>/case-studies/finance2/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/finance2/</guid><description>Scenario The customer, a Fortune 500 company in the Financial sector, had custom-built a data pipeline to perform data validation and correction transforms. The pipeline was constructed using multiple complex technologies. Examples performed on the 3 billion records included:
Standardization, verification, and cleansing of USPS codes Domain set validation, Null Checks, Length Checks Regular expression validation (email, SSN, dates, etc.) The legacy pipeline ran overnight, required multiple teams to keep it operating, and costly experts to maintain it.</description></item><item><title>Fraud detection</title><link>/case-studies/fraud-detection/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/fraud-detection/</guid><description>Scenario The customer, a Health Insurance company, was using Netezza to aggregate and report on multiple dimensions on different health care and services. They were looking for alternatives to offload the data from Netezza to reduce the workload for reducing the cost. Data Administrators, Data Analysts, Data Scientist and Engineers were operating and supporting the loading, cleansing and reporting efforts. They were facing a lot of difficulties in the following areas that was hindering their transition into Hadoop:</description></item><item><title>Get Started</title><link>/get-started/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/get-started/</guid><description/></item><item><title>Healthcare</title><link>/case-studies/healthcare/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/healthcare/</guid><description>Scenario The customer, a Health Insurance company, was using Netezza to aggregate and report on multiple dimensions on different health care and services. They were looking for alternatives to offload the data from Netezza to reduce the workload for reducing the cost. Data Administrators, Data Analysts, Data Scientist and Engineers were operating and supporting the loading, cleansing and reporting efforts. They were facing a lot of difficulties in the following areas that was hindering their transition into Hadoop:</description></item><item><title>Healthcare</title><link>/case-studies/sports-and-entertainment/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/sports-and-entertainment/</guid><description>Scenario The customer, a Sports and Entertainment company has arriving in XML files and the data scientists couldn&amp;rsquo;t query for the information they needed. They weren&amp;rsquo;t sure what the final format would be for the tables and needed to be able to adjust quickly. They required access to the final data from their existing BI tools such as Tableau and R.
CDAP value proposition(s) CDAP enables customers to ingest and transform the XML data into multiple tables without writing any code.</description></item><item><title>Pipelines</title><link>/accelerators/pipelines/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/accelerators/pipelines/</guid><description/></item><item><title>Plugins</title><link>/resources/plugins/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/resources/plugins/</guid><description/></item><item><title>Retail</title><link>/case-studies/retail/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/retail/</guid><description>Scenario The customer, a Health Insurance company, was using Netezza to aggregate and report on multiple dimensions on different health care and services. They were looking for alternatives to offload the data from Netezza to reduce the workload for reducing the cost. Data Administrators, Data Analysts, Data Scientist and Engineers were operating and supporting the loading, cleansing and reporting efforts. They were facing a lot of difficulties in the following areas that was hindering their transition into Hadoop:</description></item><item><title>Retail</title><link>/case-studies/telco/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/telco/</guid><description>Scenario The customer, a Fortune 50 company in the Telecom sector, developed a legacy custom data pipeline that performed format-preserving encryption and data masking. The pipeline extracted data from Teradata to HDFS, performed transformations, and loaded the results back into Teradata on a daily basis. This pipeline, built by a third-party service, was operationally unstable and required constant, costly intervention to keep it running.
CDAP value proposition(s) The self-service, code-free interface allowed the in-house team to reproduce and replace the existing pipeline.</description></item><item><title>Risk analytics</title><link>/case-studies/risk-analytics/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/risk-analytics/</guid><description>Scenario The customer, a Health Insurance company, was using Netezza to aggregate and report on multiple dimensions on different health care and services. They were looking for alternatives to offload the data from Netezza to reduce the workload for reducing the cost. Data Administrators, Data Analysts, Data Scientist and Engineers were operating and supporting the loading, cleansing and reporting efforts. They were facing a lot of difficulties in the following areas that was hindering their transition into Hadoop:</description></item><item><title>Rules Engine</title><link>/accelerators/rules-engine/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/accelerators/rules-engine/</guid><description/></item><item><title>Telecommunications</title><link>/case-studies/e-commerce/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/e-commerce/</guid><description>Scenario A Fortune 500 e-commerce company built a data pipeline that ingested their Twitter stream in real-time. The data was cleansed and transformed prior to conducting multi-dimensional aggregation and sentiment analysis on marketing campaigns based on tweets. The results were updated twice daily to HBase. However, the legacy pipeline suffered on two fronts: first, latency in the existing pipeline delayed the decision making process. Second, the existing data movement process proved to be costly in time and money.</description></item><item><title>Telecommunications</title><link>/case-studies/telecommunications/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/case-studies/telecommunications/</guid><description>Scenario The customer, a Health Insurance company, was using Netezza to aggregate and report on multiple dimensions on different health care and services. They were looking for alternatives to offload the data from Netezza to reduce the workload for reducing the cost. Data Administrators, Data Analysts, Data Scientist and Engineers were operating and supporting the loading, cleansing and reporting efforts. They were facing a lot of difficulties in the following areas that was hindering their transition into Hadoop:</description></item><item><title>Videos</title><link>/accelerators/videos/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/accelerators/videos/</guid><description/></item><item><title>Videos</title><link>/resources/videos/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/resources/videos/</guid><description/></item><item><title>Wrangler</title><link>/accelerators/wrangler/</link><pubDate>Mon, 01 Jan 0001 00:00:00 +0000</pubDate><guid>/accelerators/wrangler/</guid><description/></item></channel></rss>