SlideShare a Scribd company logo
1 of 53
Pop-up Loft
Streaming Data Analytics with Amazon Kinesis Firehose and
Redshift
Joyjeet Banerjee
Enterprise Solutions Architect
What to Expect From This Session
Amazon Kinesis streaming data on the AWS cloud
• Amazon Kinesis Streams
• Amazon Kinesis Firehose
• Amazon Kinesis Analytics
What to Expect From This Session
Amazon Kinesis streaming data on the AWS cloud
• Amazon Kinesis Streams
• Amazon Kinesis Firehose (focus of this session)
• Amazon Kinesis Analytics
What to Expect From This Session
Amazon Kinesis and Streaming data on the AWS cloud
Amazon Kinesis Firehose
• Key concepts
• Experience for S3 and Redshift
• Putting data using the Amazon Kinesis Agent and Key APIs
• Pricing
• Data delivery patterns
• Understanding key metrics
• Troubleshooting
Amazon Kinesis: Streaming Data Done the AWS Way
Makes it easy to capture, deliver, and process real-time data streams
Pay as you go, no up-front costs
Elastically scalable
Right services for your specific use cases
Real-time latencies
Easy to provision, deploy, and manage
Amazon Web Services
AZ AZ AZ
Durable, highly consistent storage replicates data
across three data centers (availability zones)
Aggregate and
archive to S3
Millions of
sources producing
100s of terabytes
per hour
Front
End
Authentication
Authorization
Ordered stream
of events supports
multiple readers
Real-time
dashboards
and alarms
Machine learning
algorithms or
sliding window
analytics
Aggregate analysis
in Hadoop or a
data warehouse
Inexpensive: $0.028 per million puts
Real-Time Streaming Data Ingestion
Custom-built
Streaming
Applications
(KCL)
Inexpensive: $0.014 per 1,000,000 PUT Payload Units
Amazon Kinesis Streams
Fully managed service for real-time processing of streaming data
Select Kinesis Customer Case Studies
Ad Tech Gaming IoT
We have listened to our customers…
Amazon Kinesis Streams, features from customer…
Kinesis Producer Library
PutRecords API, 500 records or 5 MB payload
Kinesis Client Library in Python, Node.JS, Ruby…
Server-Side Timestamps
Increased individual max record payload 50 KB to 1 MB
Reduced end-to-end propagation delay
Extended Stream Retention from 24 hours to 7 days
Amazon Kinesis Streams
Build your own data streaming applications
Easy administration: Simply create a new stream, and set the desired level of
capacity with shards. Scale to match your data throughput rate and volume.
Build real-time applications: Perform continual processing on streaming big data
using Kinesis Client Library (KCL), Apache Spark/Storm, AWS Lambda, and more.
Low cost: Cost-efficient for workloads of any scale.
Sushiro: Kaiten Sushi Restaurants
380 stores stream data from sushi plate sensors and stream to Kinesis
Sushiro: Kaiten Sushi Restaurants
380 stores stream data from sushi plate sensors and stream to Kinesis
Sushiro: Kaiten Sushi Restaurants
380 stores stream data from sushi plate sensors and stream to Kinesis
Amazon Kinesis Firehose
Load massive volumes of streaming data into Amazon S3 and Amazon Redshift
Zero administration: Capture and deliver streaming data into S3, Redshift, and
other destinations without writing an application or managing infrastructure.
Direct-to-data store integration: Batch, compress, and encrypt streaming data
for delivery into data destinations in as little as 60 secs using simple configurations.
Seamless elasticity: Seamlessly scales to match data throughput w/o intervention
Capture and submit
streaming data to Firehose
Firehose loads streaming data
continuously into S3 and Redshift
Analyze streaming data using your favorite
BI tools
Amazon Kinesis Firehose Customer Experience
1. Delivery Stream: The underlying entity of Firehose. Use Firehose by
creating a delivery stream to a specified destination and send data to it.
• You do not have to create a stream or provision shards.
• You do not have to specify partition keys.
2. Records: The data producer sends data blobs as large as 1,000 KB to a
delivery stream. That data blob is called a record.
3. Data Producers: Producers send records to a Delivery Stream. For
example, a web server sends log data to a delivery stream is a data
producer.
Amazon Kinesis Firehose
3 Simple Concepts
Amazon Kinesis Firehose Console Experience
Unified Console Experience for Firehose and Streams
Amazon Kinesis Firehose Console Experience (S3)
Create fully managed resources for delivery without building an app
Amazon Kinesis Firehose Console Experience
Configure data delivery options simply using the console
Amazon Kinesis Firehose to Redshift
Amazon Kinesis Firehose to Redshift
A two-step process
Use customer-provided S3 bucket as an intermediate destination
• Still the most efficient way to do large-scale loads to Redshift.
• Never lose data, always safe, and available in your S3 bucket.1
Amazon Kinesis Firehose to Redshift
A two-step process
Use customer-provided S3 bucket as an intermediate destination
• Still the most efficient way to do large scale loads to Redshift.
• Never lose data, always safe, and available in your S3 bucket.
Firehose issues customer-provided COPY command
synchronously. It continuously issues a COPY command once the
previous COPY command is finished and acknowledged back from
Redshift.
1
2
Amazon Kinesis Firehose Console
Configure data delivery to Redshift simply using the console
Amazon Kinesis Agent
Software agent makes submitting data to Firehose easy
• Monitors files and sends new data records to your delivery stream
• Handles file rotation, checkpointing, and retry upon failures
• Delivers all data in a reliable, timely, and simple manner
• Emits AWS CloudWatch metrics to help you better monitor and
troubleshoot the streaming process
• Supported on Amazon Linux AMI with version 2015.09 or later, or Red Hat
Enterprise Linux version 7 or later. Install on Linux-based server
environments such as web servers, front ends, log servers, and more
• Also enabled for Streams
• CreateDeliveryStream: Create a DeliveryStream. You specify the S3 bucket
information where your data will be delivered to.
• DeleteDeliveryStream: Delete a DeliveryStream.
• DescribeDeliveryStream: Describe a DeliveryStream and returns configuration
information.
• ListDeliveryStreams: Return a list of DeliveryStreams under this account.
• UpdateDestination: Update the destination S3 bucket information for a
DeliveryStream.
• PutRecord: Put a single data record (up to 1000KB) to a DeliveryStream.
• PutRecordBatch: Put multiple data records (up to 500 records OR 5MBs) to a
DeliveryStream.
Amazon Kinesis Firehose API Overview
Amazon Kinesis Firehose Pricing
Simple, pay-as-you-go and no up-front costs
Dimension Value
Per 1 GB of data ingested $0.035
Amazon Kinesis Firehose or Amazon Kinesis Streams?
Amazon Kinesis Streams is a service for workloads that requires
custom processing, per incoming record, with sub-1 second
processing latency, and a choice of stream processing frameworks.
Amazon Kinesis Firehose is a service for workloads that require
zero administration, ability to use existing analytics tools based
on S3 or Redshift, and a data latency of 60 seconds or higher.
Amazon Kinesis Firehose
Deep Dive
Amazon Kinesis Firehose to S3
Data Delivery Specifics
Data Delivery Methodology for Amazon S3
Controlling size and frequency of S3 objects
• Single delivery stream delivers to single S3 bucket
• Buffer size/interval values control size + frequency of data
delivery
• Buffer size – 1 MB to 128 MBs or
• Buffer interval - 60 to 900 seconds
• Firehose concatenates records into a single larger object
• Condition satisfied first triggers data delivery
• (Optional) compress data after data is buffered
• GZIP, ZIP, SNAPPY
• Delivered S3 objects might be smaller than total data Put into Firehose
• For Amazon Redshift, GZIP is only supported format currently
Data Delivery Methodology for Amazon S3
AWS IAM Roles and Encryption
• Firehose needs IAM role to access to your S3 bucket
• Delivery stream assumes specified role to gain access to target
S3 bucket
• (Optional) encrypt data with AWS Key Management Service
• AWS KMS is a managed service that makes it easy for you to create and
control the encryption keys used to encrypt your data
• Uses Hardware Security Modules (HSMs) to protect keys
• Firehose passes KMS-id to S3 which uses existing integration
• S3 uses bucket and object name as encryption context
Data Delivery Methodology for Amazon S3
S3 Object Name Format
• Firehose adds a UTC time prefix in the format YYYY/MM/DD/HH before
putting objects to Amazon S3. Translates to an Amazon S3 folder
structure. Each ‘/’ is a sub-folder
• Specify S3-prefix to get your own top-level folder
• Modify this by adding your own top-level folder with ‘/’ to get
myApp/YYYY/MM/DD/HH
• Or simply pre-pend text without a ‘/’ to get myapp YYYY/MM/DD/HH
• Expect following naming pattern “DeliveryStreamName-
DeliveryStreamVersion-YYYY-MM-DD-HH-MM-SS-RandomString”
• DeliveryStreamVersion == 1 if delivery stream config is not updated
• DeliveryStreamVersion increases by 1 for every config change
Data Delivery Methodology for Amazon S3
If Firehose cannot reach your S3 bucket
• If your Amazon S3 bucket is not reachable by Amazon
Kinesis Firehose, Firehose tries to access the S3 bucket
every 5 seconds until the issue is resolved.
• Firehose durably stores your data for 24 hours
• It resumes delivery (per configurations) as soon as your S3
bucket becomes available and catches up
• Data can be lost if bucket is not accessible for > 24 hours
Amazon Kinesis Firehose to Redshift
Data Delivery Specifics
Data Delivery Methodology for Amazon Redshift
2-step process executed on your behalf
• Use customer-provided S3 bucket as an intermediate
destination
• Still the most efficient way to do large scale loads to Redshift
• Never lose data, always safe, and available in your S3 bucket
• Firehose issues customer-provided Redshift COPY
command synchronously
• Loads data from intermediate-S3 bucket to Redshift cluster
• Single delivery stream loads into a single Redshift cluster,
database, and table
Data Delivery Methodology for Amazon Redshift
Frequency of loads into Redshift Cluster
• Based on delivery to S3 – buffer size and interval
• Redshift COPY command frequency
• Continuously issues the Redshift COPY command once the
previous one is acknowledged from Redshift.
• Frequency of COPYs from Amazon S3 to Redshift is determined
by how fast your Amazon Redshift cluster can finish the COPY
command
• For efficiency, Firehose uses manifest copy
• Intermediate S3 bucket contains ‘manifests’ folder – that holds a
manifest of files that are to be copied into Redshift
Data Delivery Methodology for Amazon Redshift
Redshift COPY command mechanics
• Firehose issues the Redshift COPY command with zero
error tolerance
• If a single record within a file fails, the whole batch of files
within the COPY command fails
• Firehose does not perform partial file or batch copies to
your Amazon Redshift table
• Skipped objects' information is delivered to S3 bucket as
manifest file in the errors folder
Data Delivery Methodology for Amazon Redshift
If Redshift cluster is not accessible
• If Firehose cannot reach cluster it tries every 5 minutes
for up to a total of 60 minutes
• If cluster is still not reachable after 60 minutes, it skips
the current batch of S3 objects and moves on to next
• Like before, skipped objects' information is delivered to
S3 bucket as manifest file in the errors folder
• Use info for manual backfill as appropriate
Amazon Kinesis Firehose
Metrics and Troubleshooting
Amazon Kinesis Firehose Monitoring
Load to S3 specific metrics
Metrics Description
Incoming.
Bytes
The number of bytes ingested into the Firehose delivery stream.
Incoming.
Records
The number of records ingested into the Firehose delivery stream.
DeliveryToS3.
Bytes
The number of bytes delivered to Amazon S3
DeliveryToS3.
DataFreshness
Age of the oldest record in Firehose. Any record older than this
age has been delivered to the S3 bucket.
DeliveryToS3.
Records
The number of records delivered to Amazon S3
DeliveryToS3.
Success
Sum of successful Amazon S3 put commands over sum of all
Amazon S3 put commands.
Amazon Kinesis Firehose Monitoring
Load to Redshift specific metrics
Metrics Description
DeliveryToRedshift.
Bytes
The number of bytes copied to Amazon Redshift
DeliveryToRedshift.
Records
The number of records copied to Amazon Redshift
DeliveryToRedshift.
Success
Sum of successful Amazon Redshift COPY
commands over the sum of all Amazon Redshift
COPY commands.
Amazon Kinesis Firehose Troubleshooting
Is data flowing end-to-end?
• Key question: Is data flowing through Firehose from
producers to the S3 destination end-to-end?
• Key metrics: Incoming.Bytes, and Incoming.Records
metrics along with DeliveryToS3.Success and
DeliveryToRedshift.Success can be used to confirm
that data is flowing end-to-end
• Additionally, check the Put* APIs Bytes, Latency, and
Request metrics for Firehose and or S3
Amazon Kinesis Firehose Troubleshooting
What is being captured, and what is being delivered?
• Key question: How much data is being captured and delivered to the
destination?
• Key metrics: Put* Records, Bytes, and Requests to determine the
data volume captured by Firehose. Next check
DeliveryToS3.Bytes/ Records and
DeliverytoRedshift.Bytes/Records
• Additionally, check the S3 bucket related metrics above confirm the
data volume delivered to S3. Verify with Incoming.Records and
Incoming.Bytes
Amazon Kinesis Firehose Troubleshooting
Is my data showing ‘on time’?
• Key question: Is Firehose delivering data in a timely
way?
• Key metrics: DeliveryToS3.DataFreshness in seconds
helps determine freshness of data. It shows the
maximum age in seconds of oldest undelivered record in
Firehose.
• Any record that is older than this age has been delivered
into S3.
Amazon Kinesis Firehose Troubleshooting
Something wrong with my Firehose or bucket or cluster?
• Key question: Is something wrong with Firehose or with the
customers’ own configuration, say S3 bucket?
• Key metrics: The DeliveryToS3.Success metric computes the
sum of successful S3 Puts over sum of all S3 puts managed by
Firehose. Similarly DeliveryToRedshift.Success metric
computes sum of successful COPY commands over sum of all
COPY commands.
• If this trends below “1” it suggests potential issues with the S3
bucket configuration such that Firehose puts to S3 are failing.
• Check Incoming.Bytes and Incoming.Records metrics to ensure
that data is Put successfully.
• Check DeliveryToS3.Success metric to ensure that Firehose is
putting data to your S3 bucket.
• Ensure that the S3 bucket specified in your delivery stream still
exists.
• Ensure that the IAM role specified in your delivery stream has
access to your S3 bucket.
Amazon Kinesis Firehose Troubleshooting
Something key steps if data isn’t showing up in S3
• Check Incoming.Bytes and Incoming.Records metrics to ensure
that data is Put successfully, and the DeliveryToS3.Success
metric to ensure that data is being staged in the S3 bucket.
• Check DeliveryToRedshift.Success metric to ensure that
Firehose has copied data from your S3 bucket to the cluster.
• Check Redshift STL_LOAD_ERRORS table to verify the reason
of the COPY failure.
• Ensure that Redshift config in Firehose is accurate
Amazon Kinesis Firehose Troubleshooting
Something key steps if data isn’t showing up in Redshift
Amazon Kinesis Analytics
Amazon Kinesis Analytics
Analyze data streams continuously with standard SQL
Apply SQL on streams: Easily connect to data streams and apply
existing SQL skills.
Build real-time applications: Perform continual processing on
streaming big data with sub-second processing latencies
Scale elastically: Elastically scales to match data throughput without
any operator intervention.
Announcement
Only!
Amazon Confidential
Connect to Kinesis streams,
Firehose delivery streams
Run standard SQL queries
against data streams
Kinesis Analytics can send processed
data to analytics tools so you can create
alerts and respond in real-time
Amazon Kinesis
Streaming Data Made Easy
Amazon Kinesis
Streams
Build your own custom
applications that
process or analyze
streaming data
Amazon Kinesis
Firehose
Easily load massive
volumes of streaming
data into Amazon S3
and Redshift
Amazon Kinesis
Analytics
Easily analyze data
streams using
standard SQL queries
Amazon Kinesis: Streaming data made easy
Services make it easy to capture, deliver, and process streams on AWS
Amazon Confidential
Q&A

More Related Content

What's hot

기술 지원 사례로 알아보는 마이그레이션 이슈 및 해결 방안 모음-김용기, AWS Storage Specialist SA / 한소영, AWS...
기술 지원 사례로 알아보는 마이그레이션 이슈 및 해결 방안 모음-김용기, AWS Storage Specialist SA / 한소영, AWS...기술 지원 사례로 알아보는 마이그레이션 이슈 및 해결 방안 모음-김용기, AWS Storage Specialist SA / 한소영, AWS...
기술 지원 사례로 알아보는 마이그레이션 이슈 및 해결 방안 모음-김용기, AWS Storage Specialist SA / 한소영, AWS...Amazon Web Services Korea
 
컨테이너 및 서버리스를 위한 효율적인 CI/CD 아키텍처 구성하기 - 현창훈 데브옵스 엔지니어, Flex / 송주영 데브옵스 엔지니어, W...
컨테이너 및 서버리스를 위한 효율적인 CI/CD 아키텍처 구성하기 - 현창훈 데브옵스 엔지니어, Flex / 송주영 데브옵스 엔지니어, W...컨테이너 및 서버리스를 위한 효율적인 CI/CD 아키텍처 구성하기 - 현창훈 데브옵스 엔지니어, Flex / 송주영 데브옵스 엔지니어, W...
컨테이너 및 서버리스를 위한 효율적인 CI/CD 아키텍처 구성하기 - 현창훈 데브옵스 엔지니어, Flex / 송주영 데브옵스 엔지니어, W...Amazon Web Services Korea
 
Aws glue를 통한 손쉬운 데이터 전처리 작업하기
Aws glue를 통한 손쉬운 데이터 전처리 작업하기Aws glue를 통한 손쉬운 데이터 전처리 작업하기
Aws glue를 통한 손쉬운 데이터 전처리 작업하기Amazon Web Services Korea
 
실시간 스트리밍 분석 Kinesis Data Analytics Deep Dive
실시간 스트리밍 분석  Kinesis Data Analytics Deep Dive실시간 스트리밍 분석  Kinesis Data Analytics Deep Dive
실시간 스트리밍 분석 Kinesis Data Analytics Deep DiveAmazon Web Services Korea
 
Migrate Enterprise Applications Framework and Guiding Principles.pdf
Migrate Enterprise Applications Framework and Guiding Principles.pdfMigrate Enterprise Applications Framework and Guiding Principles.pdf
Migrate Enterprise Applications Framework and Guiding Principles.pdfAmazon Web Services
 
Introduction to Cloud Computing with Amazon Web Services
Introduction to Cloud Computing with Amazon Web ServicesIntroduction to Cloud Computing with Amazon Web Services
Introduction to Cloud Computing with Amazon Web ServicesAmazon Web Services
 
Infrastructure as Code in AWS using Cloudformation
Infrastructure as Code in AWS using CloudformationInfrastructure as Code in AWS using Cloudformation
Infrastructure as Code in AWS using CloudformationJohn Reilly Pospos
 
Serverless computing with AWS Lambda
Serverless computing with AWS Lambda Serverless computing with AWS Lambda
Serverless computing with AWS Lambda Apigee | Google Cloud
 
AWS Compute Evolved Week: High Performance Computing on AWS
AWS Compute Evolved Week: High Performance Computing on AWSAWS Compute Evolved Week: High Performance Computing on AWS
AWS Compute Evolved Week: High Performance Computing on AWSAmazon Web Services
 
AWS Storage Services - AWS Presentation - AWS Cloud Storage for the Enterpris...
AWS Storage Services - AWS Presentation - AWS Cloud Storage for the Enterpris...AWS Storage Services - AWS Presentation - AWS Cloud Storage for the Enterpris...
AWS Storage Services - AWS Presentation - AWS Cloud Storage for the Enterpris...Amazon Web Services
 
Introduction to Amazon Elastic File System (EFS)
Introduction to Amazon Elastic File System (EFS)Introduction to Amazon Elastic File System (EFS)
Introduction to Amazon Elastic File System (EFS)Amazon Web Services
 
AWS Lake Formation을 통한 손쉬운 데이터 레이크 구성 및 관리 - 윤석찬 :: AWS Unboxing 온라인 세미나
AWS Lake Formation을 통한 손쉬운 데이터 레이크 구성 및 관리 - 윤석찬 :: AWS Unboxing 온라인 세미나AWS Lake Formation을 통한 손쉬운 데이터 레이크 구성 및 관리 - 윤석찬 :: AWS Unboxing 온라인 세미나
AWS Lake Formation을 통한 손쉬운 데이터 레이크 구성 및 관리 - 윤석찬 :: AWS Unboxing 온라인 세미나Amazon Web Services Korea
 
Architecting security and governance across your AWS environment
Architecting security and governance across your AWS environmentArchitecting security and governance across your AWS environment
Architecting security and governance across your AWS environmentAmazon Web Services
 
Scaling your analytics with Amazon EMR
Scaling your analytics with Amazon EMRScaling your analytics with Amazon EMR
Scaling your analytics with Amazon EMRIsrael AWS User Group
 
Introduction to AWS Cost Management
Introduction to AWS Cost ManagementIntroduction to AWS Cost Management
Introduction to AWS Cost ManagementAmazon Web Services
 
AWS 12월 웨비나 │성공적인 마이그레이션을 위한 클라우드 아키텍처 및 운영 고도화
AWS 12월 웨비나 │성공적인 마이그레이션을 위한 클라우드 아키텍처 및 운영 고도화AWS 12월 웨비나 │성공적인 마이그레이션을 위한 클라우드 아키텍처 및 운영 고도화
AWS 12월 웨비나 │성공적인 마이그레이션을 위한 클라우드 아키텍처 및 운영 고도화Amazon Web Services Korea
 

What's hot (20)

기술 지원 사례로 알아보는 마이그레이션 이슈 및 해결 방안 모음-김용기, AWS Storage Specialist SA / 한소영, AWS...
기술 지원 사례로 알아보는 마이그레이션 이슈 및 해결 방안 모음-김용기, AWS Storage Specialist SA / 한소영, AWS...기술 지원 사례로 알아보는 마이그레이션 이슈 및 해결 방안 모음-김용기, AWS Storage Specialist SA / 한소영, AWS...
기술 지원 사례로 알아보는 마이그레이션 이슈 및 해결 방안 모음-김용기, AWS Storage Specialist SA / 한소영, AWS...
 
컨테이너 및 서버리스를 위한 효율적인 CI/CD 아키텍처 구성하기 - 현창훈 데브옵스 엔지니어, Flex / 송주영 데브옵스 엔지니어, W...
컨테이너 및 서버리스를 위한 효율적인 CI/CD 아키텍처 구성하기 - 현창훈 데브옵스 엔지니어, Flex / 송주영 데브옵스 엔지니어, W...컨테이너 및 서버리스를 위한 효율적인 CI/CD 아키텍처 구성하기 - 현창훈 데브옵스 엔지니어, Flex / 송주영 데브옵스 엔지니어, W...
컨테이너 및 서버리스를 위한 효율적인 CI/CD 아키텍처 구성하기 - 현창훈 데브옵스 엔지니어, Flex / 송주영 데브옵스 엔지니어, W...
 
Aws glue를 통한 손쉬운 데이터 전처리 작업하기
Aws glue를 통한 손쉬운 데이터 전처리 작업하기Aws glue를 통한 손쉬운 데이터 전처리 작업하기
Aws glue를 통한 손쉬운 데이터 전처리 작업하기
 
실시간 스트리밍 분석 Kinesis Data Analytics Deep Dive
실시간 스트리밍 분석  Kinesis Data Analytics Deep Dive실시간 스트리밍 분석  Kinesis Data Analytics Deep Dive
실시간 스트리밍 분석 Kinesis Data Analytics Deep Dive
 
Migrate Enterprise Applications Framework and Guiding Principles.pdf
Migrate Enterprise Applications Framework and Guiding Principles.pdfMigrate Enterprise Applications Framework and Guiding Principles.pdf
Migrate Enterprise Applications Framework and Guiding Principles.pdf
 
AWS ELB
AWS ELBAWS ELB
AWS ELB
 
AWS 101
AWS 101AWS 101
AWS 101
 
Introduction to Cloud Computing with Amazon Web Services
Introduction to Cloud Computing with Amazon Web ServicesIntroduction to Cloud Computing with Amazon Web Services
Introduction to Cloud Computing with Amazon Web Services
 
Infrastructure as Code in AWS using Cloudformation
Infrastructure as Code in AWS using CloudformationInfrastructure as Code in AWS using Cloudformation
Infrastructure as Code in AWS using Cloudformation
 
Serverless computing with AWS Lambda
Serverless computing with AWS Lambda Serverless computing with AWS Lambda
Serverless computing with AWS Lambda
 
AWS Compute Evolved Week: High Performance Computing on AWS
AWS Compute Evolved Week: High Performance Computing on AWSAWS Compute Evolved Week: High Performance Computing on AWS
AWS Compute Evolved Week: High Performance Computing on AWS
 
AWS Storage Services - AWS Presentation - AWS Cloud Storage for the Enterpris...
AWS Storage Services - AWS Presentation - AWS Cloud Storage for the Enterpris...AWS Storage Services - AWS Presentation - AWS Cloud Storage for the Enterpris...
AWS Storage Services - AWS Presentation - AWS Cloud Storage for the Enterpris...
 
Introduction to Amazon Elastic File System (EFS)
Introduction to Amazon Elastic File System (EFS)Introduction to Amazon Elastic File System (EFS)
Introduction to Amazon Elastic File System (EFS)
 
Deep Dive Amazon EC2
Deep Dive Amazon EC2Deep Dive Amazon EC2
Deep Dive Amazon EC2
 
AWS Lake Formation을 통한 손쉬운 데이터 레이크 구성 및 관리 - 윤석찬 :: AWS Unboxing 온라인 세미나
AWS Lake Formation을 통한 손쉬운 데이터 레이크 구성 및 관리 - 윤석찬 :: AWS Unboxing 온라인 세미나AWS Lake Formation을 통한 손쉬운 데이터 레이크 구성 및 관리 - 윤석찬 :: AWS Unboxing 온라인 세미나
AWS Lake Formation을 통한 손쉬운 데이터 레이크 구성 및 관리 - 윤석찬 :: AWS Unboxing 온라인 세미나
 
Architecting security and governance across your AWS environment
Architecting security and governance across your AWS environmentArchitecting security and governance across your AWS environment
Architecting security and governance across your AWS environment
 
Scaling your analytics with Amazon EMR
Scaling your analytics with Amazon EMRScaling your analytics with Amazon EMR
Scaling your analytics with Amazon EMR
 
Amazon ECS Deep Dive
Amazon ECS Deep DiveAmazon ECS Deep Dive
Amazon ECS Deep Dive
 
Introduction to AWS Cost Management
Introduction to AWS Cost ManagementIntroduction to AWS Cost Management
Introduction to AWS Cost Management
 
AWS 12월 웨비나 │성공적인 마이그레이션을 위한 클라우드 아키텍처 및 운영 고도화
AWS 12월 웨비나 │성공적인 마이그레이션을 위한 클라우드 아키텍처 및 운영 고도화AWS 12월 웨비나 │성공적인 마이그레이션을 위한 클라우드 아키텍처 및 운영 고도화
AWS 12월 웨비나 │성공적인 마이그레이션을 위한 클라우드 아키텍처 및 운영 고도화
 

Viewers also liked

Spiderストレージエンジンの使い方と利用事例 他ストレージエンジンの紹介
Spiderストレージエンジンの使い方と利用事例 他ストレージエンジンの紹介Spiderストレージエンジンの使い方と利用事例 他ストレージエンジンの紹介
Spiderストレージエンジンの使い方と利用事例 他ストレージエンジンの紹介Kentoku
 
AndApp開発における全て #denatechcon
AndApp開発における全て #denatechconAndApp開発における全て #denatechcon
AndApp開発における全て #denatechconDeNA
 
神に近づくx/net/context (Finding God with x/net/context)
神に近づくx/net/context (Finding God with x/net/context)神に近づくx/net/context (Finding God with x/net/context)
神に近づくx/net/context (Finding God with x/net/context)guregu
 
golang.tokyo #6 (in Japanese)
golang.tokyo #6 (in Japanese)golang.tokyo #6 (in Japanese)
golang.tokyo #6 (in Japanese)Yuichi Murata
 
MongoDBの可能性の話
MongoDBの可能性の話MongoDBの可能性の話
MongoDBの可能性の話Akihiro Kuwano
 
Operations: Production Readiness Review – How to stop bad things from Happening
Operations: Production Readiness Review – How to stop bad things from HappeningOperations: Production Readiness Review – How to stop bad things from Happening
Operations: Production Readiness Review – How to stop bad things from HappeningAmazon Web Services
 
ScalaからGoへ
ScalaからGoへScalaからGoへ
ScalaからGoへJames Neve
 
Apache Spark Streaming + Kafka 0.10 with Joan Viladrosariera
Apache Spark Streaming + Kafka 0.10 with Joan ViladrosarieraApache Spark Streaming + Kafka 0.10 with Joan Viladrosariera
Apache Spark Streaming + Kafka 0.10 with Joan ViladrosarieraSpark Summit
 
An introduction and future of Ruby coverage library
An introduction and future of Ruby coverage libraryAn introduction and future of Ruby coverage library
An introduction and future of Ruby coverage librarymametter
 
AWS X-Rayによるアプリケーションの分析とデバッグ
AWS X-Rayによるアプリケーションの分析とデバッグAWS X-Rayによるアプリケーションの分析とデバッグ
AWS X-Rayによるアプリケーションの分析とデバッグAmazon Web Services Japan
 
Fast and Reliable Swift APIs with gRPC
Fast and Reliable Swift APIs with gRPCFast and Reliable Swift APIs with gRPC
Fast and Reliable Swift APIs with gRPCTim Burks
 
Swaggerでのapi開発よもやま話
Swaggerでのapi開発よもやま話Swaggerでのapi開発よもやま話
Swaggerでのapi開発よもやま話KEISUKE KONISHI
 
メルカリアッテの実務で使えた、GAE/Goの開発を効率的にする方法
メルカリアッテの実務で使えた、GAE/Goの開発を効率的にする方法メルカリアッテの実務で使えた、GAE/Goの開発を効率的にする方法
メルカリアッテの実務で使えた、GAE/Goの開発を効率的にする方法Takuya Ueda
 
So You Wanna Go Fast?
So You Wanna Go Fast?So You Wanna Go Fast?
So You Wanna Go Fast?Tyler Treat
 
Solving anything in VCL
Solving anything in VCLSolving anything in VCL
Solving anything in VCLFastly
 

Viewers also liked (20)

What’s New in Amazon Aurora
What’s New in Amazon AuroraWhat’s New in Amazon Aurora
What’s New in Amazon Aurora
 
Spiderストレージエンジンの使い方と利用事例 他ストレージエンジンの紹介
Spiderストレージエンジンの使い方と利用事例 他ストレージエンジンの紹介Spiderストレージエンジンの使い方と利用事例 他ストレージエンジンの紹介
Spiderストレージエンジンの使い方と利用事例 他ストレージエンジンの紹介
 
AndApp開発における全て #denatechcon
AndApp開発における全て #denatechconAndApp開発における全て #denatechcon
AndApp開発における全て #denatechcon
 
神に近づくx/net/context (Finding God with x/net/context)
神に近づくx/net/context (Finding God with x/net/context)神に近づくx/net/context (Finding God with x/net/context)
神に近づくx/net/context (Finding God with x/net/context)
 
golang.tokyo #6 (in Japanese)
golang.tokyo #6 (in Japanese)golang.tokyo #6 (in Japanese)
golang.tokyo #6 (in Japanese)
 
MongoDBの可能性の話
MongoDBの可能性の話MongoDBの可能性の話
MongoDBの可能性の話
 
Operations: Production Readiness Review – How to stop bad things from Happening
Operations: Production Readiness Review – How to stop bad things from HappeningOperations: Production Readiness Review – How to stop bad things from Happening
Operations: Production Readiness Review – How to stop bad things from Happening
 
SLOのすすめ
SLOのすすめSLOのすすめ
SLOのすすめ
 
ScalaからGoへ
ScalaからGoへScalaからGoへ
ScalaからGoへ
 
Blockchain on Go
Blockchain on GoBlockchain on Go
Blockchain on Go
 
Apache Spark Streaming + Kafka 0.10 with Joan Viladrosariera
Apache Spark Streaming + Kafka 0.10 with Joan ViladrosarieraApache Spark Streaming + Kafka 0.10 with Joan Viladrosariera
Apache Spark Streaming + Kafka 0.10 with Joan Viladrosariera
 
An introduction and future of Ruby coverage library
An introduction and future of Ruby coverage libraryAn introduction and future of Ruby coverage library
An introduction and future of Ruby coverage library
 
AWS X-Rayによるアプリケーションの分析とデバッグ
AWS X-Rayによるアプリケーションの分析とデバッグAWS X-Rayによるアプリケーションの分析とデバッグ
AWS X-Rayによるアプリケーションの分析とデバッグ
 
Microservices at Mercari
Microservices at MercariMicroservices at Mercari
Microservices at Mercari
 
Fast and Reliable Swift APIs with gRPC
Fast and Reliable Swift APIs with gRPCFast and Reliable Swift APIs with gRPC
Fast and Reliable Swift APIs with gRPC
 
Swaggerでのapi開発よもやま話
Swaggerでのapi開発よもやま話Swaggerでのapi開発よもやま話
Swaggerでのapi開発よもやま話
 
メルカリアッテの実務で使えた、GAE/Goの開発を効率的にする方法
メルカリアッテの実務で使えた、GAE/Goの開発を効率的にする方法メルカリアッテの実務で使えた、GAE/Goの開発を効率的にする方法
メルカリアッテの実務で使えた、GAE/Goの開発を効率的にする方法
 
So You Wanna Go Fast?
So You Wanna Go Fast?So You Wanna Go Fast?
So You Wanna Go Fast?
 
Solving anything in VCL
Solving anything in VCLSolving anything in VCL
Solving anything in VCL
 
Google Home and Google Assistant Workshop: Build your own serverless Action o...
Google Home and Google Assistant Workshop: Build your own serverless Action o...Google Home and Google Assistant Workshop: Build your own serverless Action o...
Google Home and Google Assistant Workshop: Build your own serverless Action o...
 

Similar to Streaming Data Analytics with Amazon Redshift and Kinesis Firehose

(BDT320) New! Streaming Data Flows with Amazon Kinesis Firehose
(BDT320) New! Streaming Data Flows with Amazon Kinesis Firehose(BDT320) New! Streaming Data Flows with Amazon Kinesis Firehose
(BDT320) New! Streaming Data Flows with Amazon Kinesis FirehoseAmazon Web Services
 
AWS October Webinar Series - Introducing Amazon Kinesis Firehose
AWS October Webinar Series - Introducing Amazon Kinesis FirehoseAWS October Webinar Series - Introducing Amazon Kinesis Firehose
AWS October Webinar Series - Introducing Amazon Kinesis FirehoseAmazon Web Services
 
Amazon Kinesis Firehose - Pop-up Loft TLV 2017
Amazon Kinesis Firehose - Pop-up Loft TLV 2017Amazon Kinesis Firehose - Pop-up Loft TLV 2017
Amazon Kinesis Firehose - Pop-up Loft TLV 2017Amazon Web Services
 
Getting Started with Amazon Kinesis | AWS Public Sector Summit 2016
Getting Started with Amazon Kinesis | AWS Public Sector Summit 2016Getting Started with Amazon Kinesis | AWS Public Sector Summit 2016
Getting Started with Amazon Kinesis | AWS Public Sector Summit 2016Amazon Web Services
 
Log Analytics with Amazon Elasticsearch Service and Amazon Kinesis - March 20...
Log Analytics with Amazon Elasticsearch Service and Amazon Kinesis - March 20...Log Analytics with Amazon Elasticsearch Service and Amazon Kinesis - March 20...
Log Analytics with Amazon Elasticsearch Service and Amazon Kinesis - March 20...Amazon Web Services
 
Getting Started with Amazon Kinesis
Getting Started with Amazon KinesisGetting Started with Amazon Kinesis
Getting Started with Amazon KinesisAmazon Web Services
 
Em tempo real: Ingestão, processamento e analise de dados
Em tempo real: Ingestão, processamento e analise de dadosEm tempo real: Ingestão, processamento e analise de dados
Em tempo real: Ingestão, processamento e analise de dadosAmazon Web Services LATAM
 
Deep dive and best practices on real time streaming applications nyc-loft_oct...
Deep dive and best practices on real time streaming applications nyc-loft_oct...Deep dive and best practices on real time streaming applications nyc-loft_oct...
Deep dive and best practices on real time streaming applications nyc-loft_oct...Amazon Web Services
 
AWS April 2016 Webinar Series - Getting Started with Real-Time Data Analytics...
AWS April 2016 Webinar Series - Getting Started with Real-Time Data Analytics...AWS April 2016 Webinar Series - Getting Started with Real-Time Data Analytics...
AWS April 2016 Webinar Series - Getting Started with Real-Time Data Analytics...Amazon Web Services
 
Streaming Data Analytics with Amazon Redshift Firehose
Streaming Data Analytics with Amazon Redshift FirehoseStreaming Data Analytics with Amazon Redshift Firehose
Streaming Data Analytics with Amazon Redshift FirehoseAmazon Web Services
 
Easy Analytics with AWS - AWS Summit Bahrain 2017
Easy Analytics with AWS - AWS Summit Bahrain 2017Easy Analytics with AWS - AWS Summit Bahrain 2017
Easy Analytics with AWS - AWS Summit Bahrain 2017Amazon Web Services
 
Streaming Data Analytics with Amazon Redshift and Kinesis Firehose
Streaming Data Analytics with Amazon Redshift and Kinesis FirehoseStreaming Data Analytics with Amazon Redshift and Kinesis Firehose
Streaming Data Analytics with Amazon Redshift and Kinesis FirehoseAmazon Web Services
 
Deep Dive and Best Practices for Real Time Streaming Applications
Deep Dive and Best Practices for Real Time Streaming ApplicationsDeep Dive and Best Practices for Real Time Streaming Applications
Deep Dive and Best Practices for Real Time Streaming ApplicationsAmazon Web Services
 
Cloud Data Migration Strategies - AWS May 2016 Webinar Series
Cloud Data Migration Strategies - AWS May 2016 Webinar SeriesCloud Data Migration Strategies - AWS May 2016 Webinar Series
Cloud Data Migration Strategies - AWS May 2016 Webinar SeriesAmazon Web Services
 
AWS Data Transfer Services: Data Ingest Strategies Into the AWS Cloud
AWS Data Transfer Services: Data Ingest Strategies Into the AWS CloudAWS Data Transfer Services: Data Ingest Strategies Into the AWS Cloud
AWS Data Transfer Services: Data Ingest Strategies Into the AWS CloudAmazon Web Services
 
Getting started with Amazon Kinesis
Getting started with Amazon KinesisGetting started with Amazon Kinesis
Getting started with Amazon KinesisAmazon Web Services
 
Getting started with amazon kinesis
Getting started with amazon kinesisGetting started with amazon kinesis
Getting started with amazon kinesisJampp
 
BDA308 Serverless Analytics with Amazon Athena and Amazon QuickSight, featuri...
BDA308 Serverless Analytics with Amazon Athena and Amazon QuickSight, featuri...BDA308 Serverless Analytics with Amazon Athena and Amazon QuickSight, featuri...
BDA308 Serverless Analytics with Amazon Athena and Amazon QuickSight, featuri...Amazon Web Services
 
(BDT322) How Redfin & Twitter Leverage Amazon S3 For Big Data
(BDT322) How Redfin & Twitter Leverage Amazon S3 For Big Data(BDT322) How Redfin & Twitter Leverage Amazon S3 For Big Data
(BDT322) How Redfin & Twitter Leverage Amazon S3 For Big DataAmazon Web Services
 

Similar to Streaming Data Analytics with Amazon Redshift and Kinesis Firehose (20)

(BDT320) New! Streaming Data Flows with Amazon Kinesis Firehose
(BDT320) New! Streaming Data Flows with Amazon Kinesis Firehose(BDT320) New! Streaming Data Flows with Amazon Kinesis Firehose
(BDT320) New! Streaming Data Flows with Amazon Kinesis Firehose
 
AWS October Webinar Series - Introducing Amazon Kinesis Firehose
AWS October Webinar Series - Introducing Amazon Kinesis FirehoseAWS October Webinar Series - Introducing Amazon Kinesis Firehose
AWS October Webinar Series - Introducing Amazon Kinesis Firehose
 
Amazon Kinesis Firehose - Pop-up Loft TLV 2017
Amazon Kinesis Firehose - Pop-up Loft TLV 2017Amazon Kinesis Firehose - Pop-up Loft TLV 2017
Amazon Kinesis Firehose - Pop-up Loft TLV 2017
 
Getting Started with Amazon Kinesis | AWS Public Sector Summit 2016
Getting Started with Amazon Kinesis | AWS Public Sector Summit 2016Getting Started with Amazon Kinesis | AWS Public Sector Summit 2016
Getting Started with Amazon Kinesis | AWS Public Sector Summit 2016
 
Log Analytics with Amazon Elasticsearch Service and Amazon Kinesis - March 20...
Log Analytics with Amazon Elasticsearch Service and Amazon Kinesis - March 20...Log Analytics with Amazon Elasticsearch Service and Amazon Kinesis - March 20...
Log Analytics with Amazon Elasticsearch Service and Amazon Kinesis - March 20...
 
Serverless Real Time Analytics
Serverless Real Time AnalyticsServerless Real Time Analytics
Serverless Real Time Analytics
 
Getting Started with Amazon Kinesis
Getting Started with Amazon KinesisGetting Started with Amazon Kinesis
Getting Started with Amazon Kinesis
 
Em tempo real: Ingestão, processamento e analise de dados
Em tempo real: Ingestão, processamento e analise de dadosEm tempo real: Ingestão, processamento e analise de dados
Em tempo real: Ingestão, processamento e analise de dados
 
Deep dive and best practices on real time streaming applications nyc-loft_oct...
Deep dive and best practices on real time streaming applications nyc-loft_oct...Deep dive and best practices on real time streaming applications nyc-loft_oct...
Deep dive and best practices on real time streaming applications nyc-loft_oct...
 
AWS April 2016 Webinar Series - Getting Started with Real-Time Data Analytics...
AWS April 2016 Webinar Series - Getting Started with Real-Time Data Analytics...AWS April 2016 Webinar Series - Getting Started with Real-Time Data Analytics...
AWS April 2016 Webinar Series - Getting Started with Real-Time Data Analytics...
 
Streaming Data Analytics with Amazon Redshift Firehose
Streaming Data Analytics with Amazon Redshift FirehoseStreaming Data Analytics with Amazon Redshift Firehose
Streaming Data Analytics with Amazon Redshift Firehose
 
Easy Analytics with AWS - AWS Summit Bahrain 2017
Easy Analytics with AWS - AWS Summit Bahrain 2017Easy Analytics with AWS - AWS Summit Bahrain 2017
Easy Analytics with AWS - AWS Summit Bahrain 2017
 
Streaming Data Analytics with Amazon Redshift and Kinesis Firehose
Streaming Data Analytics with Amazon Redshift and Kinesis FirehoseStreaming Data Analytics with Amazon Redshift and Kinesis Firehose
Streaming Data Analytics with Amazon Redshift and Kinesis Firehose
 
Deep Dive and Best Practices for Real Time Streaming Applications
Deep Dive and Best Practices for Real Time Streaming ApplicationsDeep Dive and Best Practices for Real Time Streaming Applications
Deep Dive and Best Practices for Real Time Streaming Applications
 
Cloud Data Migration Strategies - AWS May 2016 Webinar Series
Cloud Data Migration Strategies - AWS May 2016 Webinar SeriesCloud Data Migration Strategies - AWS May 2016 Webinar Series
Cloud Data Migration Strategies - AWS May 2016 Webinar Series
 
AWS Data Transfer Services: Data Ingest Strategies Into the AWS Cloud
AWS Data Transfer Services: Data Ingest Strategies Into the AWS CloudAWS Data Transfer Services: Data Ingest Strategies Into the AWS Cloud
AWS Data Transfer Services: Data Ingest Strategies Into the AWS Cloud
 
Getting started with Amazon Kinesis
Getting started with Amazon KinesisGetting started with Amazon Kinesis
Getting started with Amazon Kinesis
 
Getting started with amazon kinesis
Getting started with amazon kinesisGetting started with amazon kinesis
Getting started with amazon kinesis
 
BDA308 Serverless Analytics with Amazon Athena and Amazon QuickSight, featuri...
BDA308 Serverless Analytics with Amazon Athena and Amazon QuickSight, featuri...BDA308 Serverless Analytics with Amazon Athena and Amazon QuickSight, featuri...
BDA308 Serverless Analytics with Amazon Athena and Amazon QuickSight, featuri...
 
(BDT322) How Redfin & Twitter Leverage Amazon S3 For Big Data
(BDT322) How Redfin & Twitter Leverage Amazon S3 For Big Data(BDT322) How Redfin & Twitter Leverage Amazon S3 For Big Data
(BDT322) How Redfin & Twitter Leverage Amazon S3 For Big Data
 

More from Amazon Web Services

Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...
Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...
Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...Amazon Web Services
 
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...Big Data per le Startup: come creare applicazioni Big Data in modalità Server...
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...Amazon Web Services
 
Esegui pod serverless con Amazon EKS e AWS Fargate
Esegui pod serverless con Amazon EKS e AWS FargateEsegui pod serverless con Amazon EKS e AWS Fargate
Esegui pod serverless con Amazon EKS e AWS FargateAmazon Web Services
 
Costruire Applicazioni Moderne con AWS
Costruire Applicazioni Moderne con AWSCostruire Applicazioni Moderne con AWS
Costruire Applicazioni Moderne con AWSAmazon Web Services
 
Come spendere fino al 90% in meno con i container e le istanze spot
Come spendere fino al 90% in meno con i container e le istanze spot Come spendere fino al 90% in meno con i container e le istanze spot
Come spendere fino al 90% in meno con i container e le istanze spot Amazon Web Services
 
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...Amazon Web Services
 
OpsWorks Configuration Management: automatizza la gestione e i deployment del...
OpsWorks Configuration Management: automatizza la gestione e i deployment del...OpsWorks Configuration Management: automatizza la gestione e i deployment del...
OpsWorks Configuration Management: automatizza la gestione e i deployment del...Amazon Web Services
 
Microsoft Active Directory su AWS per supportare i tuoi Windows Workloads
Microsoft Active Directory su AWS per supportare i tuoi Windows WorkloadsMicrosoft Active Directory su AWS per supportare i tuoi Windows Workloads
Microsoft Active Directory su AWS per supportare i tuoi Windows WorkloadsAmazon Web Services
 
Database Oracle e VMware Cloud on AWS i miti da sfatare
Database Oracle e VMware Cloud on AWS i miti da sfatareDatabase Oracle e VMware Cloud on AWS i miti da sfatare
Database Oracle e VMware Cloud on AWS i miti da sfatareAmazon Web Services
 
Crea la tua prima serverless ledger-based app con QLDB e NodeJS
Crea la tua prima serverless ledger-based app con QLDB e NodeJSCrea la tua prima serverless ledger-based app con QLDB e NodeJS
Crea la tua prima serverless ledger-based app con QLDB e NodeJSAmazon Web Services
 
API moderne real-time per applicazioni mobili e web
API moderne real-time per applicazioni mobili e webAPI moderne real-time per applicazioni mobili e web
API moderne real-time per applicazioni mobili e webAmazon Web Services
 
Database Oracle e VMware Cloud™ on AWS: i miti da sfatare
Database Oracle e VMware Cloud™ on AWS: i miti da sfatareDatabase Oracle e VMware Cloud™ on AWS: i miti da sfatare
Database Oracle e VMware Cloud™ on AWS: i miti da sfatareAmazon Web Services
 
Tools for building your MVP on AWS
Tools for building your MVP on AWSTools for building your MVP on AWS
Tools for building your MVP on AWSAmazon Web Services
 
How to Build a Winning Pitch Deck
How to Build a Winning Pitch DeckHow to Build a Winning Pitch Deck
How to Build a Winning Pitch DeckAmazon Web Services
 
Building a web application without servers
Building a web application without serversBuilding a web application without servers
Building a web application without serversAmazon Web Services
 
AWS_HK_StartupDay_Building Interactive websites while automating for efficien...
AWS_HK_StartupDay_Building Interactive websites while automating for efficien...AWS_HK_StartupDay_Building Interactive websites while automating for efficien...
AWS_HK_StartupDay_Building Interactive websites while automating for efficien...Amazon Web Services
 
Introduzione a Amazon Elastic Container Service
Introduzione a Amazon Elastic Container ServiceIntroduzione a Amazon Elastic Container Service
Introduzione a Amazon Elastic Container ServiceAmazon Web Services
 

More from Amazon Web Services (20)

Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...
Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...
Come costruire servizi di Forecasting sfruttando algoritmi di ML e deep learn...
 
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...Big Data per le Startup: come creare applicazioni Big Data in modalità Server...
Big Data per le Startup: come creare applicazioni Big Data in modalità Server...
 
Esegui pod serverless con Amazon EKS e AWS Fargate
Esegui pod serverless con Amazon EKS e AWS FargateEsegui pod serverless con Amazon EKS e AWS Fargate
Esegui pod serverless con Amazon EKS e AWS Fargate
 
Costruire Applicazioni Moderne con AWS
Costruire Applicazioni Moderne con AWSCostruire Applicazioni Moderne con AWS
Costruire Applicazioni Moderne con AWS
 
Come spendere fino al 90% in meno con i container e le istanze spot
Come spendere fino al 90% in meno con i container e le istanze spot Come spendere fino al 90% in meno con i container e le istanze spot
Come spendere fino al 90% in meno con i container e le istanze spot
 
Open banking as a service
Open banking as a serviceOpen banking as a service
Open banking as a service
 
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...
Rendi unica l’offerta della tua startup sul mercato con i servizi Machine Lea...
 
OpsWorks Configuration Management: automatizza la gestione e i deployment del...
OpsWorks Configuration Management: automatizza la gestione e i deployment del...OpsWorks Configuration Management: automatizza la gestione e i deployment del...
OpsWorks Configuration Management: automatizza la gestione e i deployment del...
 
Microsoft Active Directory su AWS per supportare i tuoi Windows Workloads
Microsoft Active Directory su AWS per supportare i tuoi Windows WorkloadsMicrosoft Active Directory su AWS per supportare i tuoi Windows Workloads
Microsoft Active Directory su AWS per supportare i tuoi Windows Workloads
 
Computer Vision con AWS
Computer Vision con AWSComputer Vision con AWS
Computer Vision con AWS
 
Database Oracle e VMware Cloud on AWS i miti da sfatare
Database Oracle e VMware Cloud on AWS i miti da sfatareDatabase Oracle e VMware Cloud on AWS i miti da sfatare
Database Oracle e VMware Cloud on AWS i miti da sfatare
 
Crea la tua prima serverless ledger-based app con QLDB e NodeJS
Crea la tua prima serverless ledger-based app con QLDB e NodeJSCrea la tua prima serverless ledger-based app con QLDB e NodeJS
Crea la tua prima serverless ledger-based app con QLDB e NodeJS
 
API moderne real-time per applicazioni mobili e web
API moderne real-time per applicazioni mobili e webAPI moderne real-time per applicazioni mobili e web
API moderne real-time per applicazioni mobili e web
 
Database Oracle e VMware Cloud™ on AWS: i miti da sfatare
Database Oracle e VMware Cloud™ on AWS: i miti da sfatareDatabase Oracle e VMware Cloud™ on AWS: i miti da sfatare
Database Oracle e VMware Cloud™ on AWS: i miti da sfatare
 
Tools for building your MVP on AWS
Tools for building your MVP on AWSTools for building your MVP on AWS
Tools for building your MVP on AWS
 
How to Build a Winning Pitch Deck
How to Build a Winning Pitch DeckHow to Build a Winning Pitch Deck
How to Build a Winning Pitch Deck
 
Building a web application without servers
Building a web application without serversBuilding a web application without servers
Building a web application without servers
 
Fundraising Essentials
Fundraising EssentialsFundraising Essentials
Fundraising Essentials
 
AWS_HK_StartupDay_Building Interactive websites while automating for efficien...
AWS_HK_StartupDay_Building Interactive websites while automating for efficien...AWS_HK_StartupDay_Building Interactive websites while automating for efficien...
AWS_HK_StartupDay_Building Interactive websites while automating for efficien...
 
Introduzione a Amazon Elastic Container Service
Introduzione a Amazon Elastic Container ServiceIntroduzione a Amazon Elastic Container Service
Introduzione a Amazon Elastic Container Service
 

Streaming Data Analytics with Amazon Redshift and Kinesis Firehose

  • 1. Pop-up Loft Streaming Data Analytics with Amazon Kinesis Firehose and Redshift Joyjeet Banerjee Enterprise Solutions Architect
  • 2. What to Expect From This Session Amazon Kinesis streaming data on the AWS cloud • Amazon Kinesis Streams • Amazon Kinesis Firehose • Amazon Kinesis Analytics
  • 3. What to Expect From This Session Amazon Kinesis streaming data on the AWS cloud • Amazon Kinesis Streams • Amazon Kinesis Firehose (focus of this session) • Amazon Kinesis Analytics
  • 4. What to Expect From This Session Amazon Kinesis and Streaming data on the AWS cloud Amazon Kinesis Firehose • Key concepts • Experience for S3 and Redshift • Putting data using the Amazon Kinesis Agent and Key APIs • Pricing • Data delivery patterns • Understanding key metrics • Troubleshooting
  • 5. Amazon Kinesis: Streaming Data Done the AWS Way Makes it easy to capture, deliver, and process real-time data streams Pay as you go, no up-front costs Elastically scalable Right services for your specific use cases Real-time latencies Easy to provision, deploy, and manage
  • 6. Amazon Web Services AZ AZ AZ Durable, highly consistent storage replicates data across three data centers (availability zones) Aggregate and archive to S3 Millions of sources producing 100s of terabytes per hour Front End Authentication Authorization Ordered stream of events supports multiple readers Real-time dashboards and alarms Machine learning algorithms or sliding window analytics Aggregate analysis in Hadoop or a data warehouse Inexpensive: $0.028 per million puts Real-Time Streaming Data Ingestion Custom-built Streaming Applications (KCL) Inexpensive: $0.014 per 1,000,000 PUT Payload Units Amazon Kinesis Streams Fully managed service for real-time processing of streaming data
  • 7. Select Kinesis Customer Case Studies Ad Tech Gaming IoT
  • 8. We have listened to our customers…
  • 9. Amazon Kinesis Streams, features from customer… Kinesis Producer Library PutRecords API, 500 records or 5 MB payload Kinesis Client Library in Python, Node.JS, Ruby… Server-Side Timestamps Increased individual max record payload 50 KB to 1 MB Reduced end-to-end propagation delay Extended Stream Retention from 24 hours to 7 days
  • 10. Amazon Kinesis Streams Build your own data streaming applications Easy administration: Simply create a new stream, and set the desired level of capacity with shards. Scale to match your data throughput rate and volume. Build real-time applications: Perform continual processing on streaming big data using Kinesis Client Library (KCL), Apache Spark/Storm, AWS Lambda, and more. Low cost: Cost-efficient for workloads of any scale.
  • 11. Sushiro: Kaiten Sushi Restaurants 380 stores stream data from sushi plate sensors and stream to Kinesis
  • 12. Sushiro: Kaiten Sushi Restaurants 380 stores stream data from sushi plate sensors and stream to Kinesis
  • 13. Sushiro: Kaiten Sushi Restaurants 380 stores stream data from sushi plate sensors and stream to Kinesis
  • 14. Amazon Kinesis Firehose Load massive volumes of streaming data into Amazon S3 and Amazon Redshift Zero administration: Capture and deliver streaming data into S3, Redshift, and other destinations without writing an application or managing infrastructure. Direct-to-data store integration: Batch, compress, and encrypt streaming data for delivery into data destinations in as little as 60 secs using simple configurations. Seamless elasticity: Seamlessly scales to match data throughput w/o intervention Capture and submit streaming data to Firehose Firehose loads streaming data continuously into S3 and Redshift Analyze streaming data using your favorite BI tools
  • 15. Amazon Kinesis Firehose Customer Experience
  • 16. 1. Delivery Stream: The underlying entity of Firehose. Use Firehose by creating a delivery stream to a specified destination and send data to it. • You do not have to create a stream or provision shards. • You do not have to specify partition keys. 2. Records: The data producer sends data blobs as large as 1,000 KB to a delivery stream. That data blob is called a record. 3. Data Producers: Producers send records to a Delivery Stream. For example, a web server sends log data to a delivery stream is a data producer. Amazon Kinesis Firehose 3 Simple Concepts
  • 17. Amazon Kinesis Firehose Console Experience Unified Console Experience for Firehose and Streams
  • 18. Amazon Kinesis Firehose Console Experience (S3) Create fully managed resources for delivery without building an app
  • 19. Amazon Kinesis Firehose Console Experience Configure data delivery options simply using the console
  • 20. Amazon Kinesis Firehose to Redshift
  • 21. Amazon Kinesis Firehose to Redshift A two-step process Use customer-provided S3 bucket as an intermediate destination • Still the most efficient way to do large-scale loads to Redshift. • Never lose data, always safe, and available in your S3 bucket.1
  • 22. Amazon Kinesis Firehose to Redshift A two-step process Use customer-provided S3 bucket as an intermediate destination • Still the most efficient way to do large scale loads to Redshift. • Never lose data, always safe, and available in your S3 bucket. Firehose issues customer-provided COPY command synchronously. It continuously issues a COPY command once the previous COPY command is finished and acknowledged back from Redshift. 1 2
  • 23. Amazon Kinesis Firehose Console Configure data delivery to Redshift simply using the console
  • 24. Amazon Kinesis Agent Software agent makes submitting data to Firehose easy • Monitors files and sends new data records to your delivery stream • Handles file rotation, checkpointing, and retry upon failures • Delivers all data in a reliable, timely, and simple manner • Emits AWS CloudWatch metrics to help you better monitor and troubleshoot the streaming process • Supported on Amazon Linux AMI with version 2015.09 or later, or Red Hat Enterprise Linux version 7 or later. Install on Linux-based server environments such as web servers, front ends, log servers, and more • Also enabled for Streams
  • 25. • CreateDeliveryStream: Create a DeliveryStream. You specify the S3 bucket information where your data will be delivered to. • DeleteDeliveryStream: Delete a DeliveryStream. • DescribeDeliveryStream: Describe a DeliveryStream and returns configuration information. • ListDeliveryStreams: Return a list of DeliveryStreams under this account. • UpdateDestination: Update the destination S3 bucket information for a DeliveryStream. • PutRecord: Put a single data record (up to 1000KB) to a DeliveryStream. • PutRecordBatch: Put multiple data records (up to 500 records OR 5MBs) to a DeliveryStream. Amazon Kinesis Firehose API Overview
  • 26. Amazon Kinesis Firehose Pricing Simple, pay-as-you-go and no up-front costs Dimension Value Per 1 GB of data ingested $0.035
  • 27. Amazon Kinesis Firehose or Amazon Kinesis Streams?
  • 28. Amazon Kinesis Streams is a service for workloads that requires custom processing, per incoming record, with sub-1 second processing latency, and a choice of stream processing frameworks. Amazon Kinesis Firehose is a service for workloads that require zero administration, ability to use existing analytics tools based on S3 or Redshift, and a data latency of 60 seconds or higher.
  • 30. Amazon Kinesis Firehose to S3 Data Delivery Specifics
  • 31. Data Delivery Methodology for Amazon S3 Controlling size and frequency of S3 objects • Single delivery stream delivers to single S3 bucket • Buffer size/interval values control size + frequency of data delivery • Buffer size – 1 MB to 128 MBs or • Buffer interval - 60 to 900 seconds • Firehose concatenates records into a single larger object • Condition satisfied first triggers data delivery • (Optional) compress data after data is buffered • GZIP, ZIP, SNAPPY • Delivered S3 objects might be smaller than total data Put into Firehose • For Amazon Redshift, GZIP is only supported format currently
  • 32. Data Delivery Methodology for Amazon S3 AWS IAM Roles and Encryption • Firehose needs IAM role to access to your S3 bucket • Delivery stream assumes specified role to gain access to target S3 bucket • (Optional) encrypt data with AWS Key Management Service • AWS KMS is a managed service that makes it easy for you to create and control the encryption keys used to encrypt your data • Uses Hardware Security Modules (HSMs) to protect keys • Firehose passes KMS-id to S3 which uses existing integration • S3 uses bucket and object name as encryption context
  • 33. Data Delivery Methodology for Amazon S3 S3 Object Name Format • Firehose adds a UTC time prefix in the format YYYY/MM/DD/HH before putting objects to Amazon S3. Translates to an Amazon S3 folder structure. Each ‘/’ is a sub-folder • Specify S3-prefix to get your own top-level folder • Modify this by adding your own top-level folder with ‘/’ to get myApp/YYYY/MM/DD/HH • Or simply pre-pend text without a ‘/’ to get myapp YYYY/MM/DD/HH • Expect following naming pattern “DeliveryStreamName- DeliveryStreamVersion-YYYY-MM-DD-HH-MM-SS-RandomString” • DeliveryStreamVersion == 1 if delivery stream config is not updated • DeliveryStreamVersion increases by 1 for every config change
  • 34. Data Delivery Methodology for Amazon S3 If Firehose cannot reach your S3 bucket • If your Amazon S3 bucket is not reachable by Amazon Kinesis Firehose, Firehose tries to access the S3 bucket every 5 seconds until the issue is resolved. • Firehose durably stores your data for 24 hours • It resumes delivery (per configurations) as soon as your S3 bucket becomes available and catches up • Data can be lost if bucket is not accessible for > 24 hours
  • 35. Amazon Kinesis Firehose to Redshift Data Delivery Specifics
  • 36. Data Delivery Methodology for Amazon Redshift 2-step process executed on your behalf • Use customer-provided S3 bucket as an intermediate destination • Still the most efficient way to do large scale loads to Redshift • Never lose data, always safe, and available in your S3 bucket • Firehose issues customer-provided Redshift COPY command synchronously • Loads data from intermediate-S3 bucket to Redshift cluster • Single delivery stream loads into a single Redshift cluster, database, and table
  • 37. Data Delivery Methodology for Amazon Redshift Frequency of loads into Redshift Cluster • Based on delivery to S3 – buffer size and interval • Redshift COPY command frequency • Continuously issues the Redshift COPY command once the previous one is acknowledged from Redshift. • Frequency of COPYs from Amazon S3 to Redshift is determined by how fast your Amazon Redshift cluster can finish the COPY command • For efficiency, Firehose uses manifest copy • Intermediate S3 bucket contains ‘manifests’ folder – that holds a manifest of files that are to be copied into Redshift
  • 38. Data Delivery Methodology for Amazon Redshift Redshift COPY command mechanics • Firehose issues the Redshift COPY command with zero error tolerance • If a single record within a file fails, the whole batch of files within the COPY command fails • Firehose does not perform partial file or batch copies to your Amazon Redshift table • Skipped objects' information is delivered to S3 bucket as manifest file in the errors folder
  • 39. Data Delivery Methodology for Amazon Redshift If Redshift cluster is not accessible • If Firehose cannot reach cluster it tries every 5 minutes for up to a total of 60 minutes • If cluster is still not reachable after 60 minutes, it skips the current batch of S3 objects and moves on to next • Like before, skipped objects' information is delivered to S3 bucket as manifest file in the errors folder • Use info for manual backfill as appropriate
  • 40. Amazon Kinesis Firehose Metrics and Troubleshooting
  • 41. Amazon Kinesis Firehose Monitoring Load to S3 specific metrics Metrics Description Incoming. Bytes The number of bytes ingested into the Firehose delivery stream. Incoming. Records The number of records ingested into the Firehose delivery stream. DeliveryToS3. Bytes The number of bytes delivered to Amazon S3 DeliveryToS3. DataFreshness Age of the oldest record in Firehose. Any record older than this age has been delivered to the S3 bucket. DeliveryToS3. Records The number of records delivered to Amazon S3 DeliveryToS3. Success Sum of successful Amazon S3 put commands over sum of all Amazon S3 put commands.
  • 42. Amazon Kinesis Firehose Monitoring Load to Redshift specific metrics Metrics Description DeliveryToRedshift. Bytes The number of bytes copied to Amazon Redshift DeliveryToRedshift. Records The number of records copied to Amazon Redshift DeliveryToRedshift. Success Sum of successful Amazon Redshift COPY commands over the sum of all Amazon Redshift COPY commands.
  • 43. Amazon Kinesis Firehose Troubleshooting Is data flowing end-to-end? • Key question: Is data flowing through Firehose from producers to the S3 destination end-to-end? • Key metrics: Incoming.Bytes, and Incoming.Records metrics along with DeliveryToS3.Success and DeliveryToRedshift.Success can be used to confirm that data is flowing end-to-end • Additionally, check the Put* APIs Bytes, Latency, and Request metrics for Firehose and or S3
  • 44. Amazon Kinesis Firehose Troubleshooting What is being captured, and what is being delivered? • Key question: How much data is being captured and delivered to the destination? • Key metrics: Put* Records, Bytes, and Requests to determine the data volume captured by Firehose. Next check DeliveryToS3.Bytes/ Records and DeliverytoRedshift.Bytes/Records • Additionally, check the S3 bucket related metrics above confirm the data volume delivered to S3. Verify with Incoming.Records and Incoming.Bytes
  • 45. Amazon Kinesis Firehose Troubleshooting Is my data showing ‘on time’? • Key question: Is Firehose delivering data in a timely way? • Key metrics: DeliveryToS3.DataFreshness in seconds helps determine freshness of data. It shows the maximum age in seconds of oldest undelivered record in Firehose. • Any record that is older than this age has been delivered into S3.
  • 46. Amazon Kinesis Firehose Troubleshooting Something wrong with my Firehose or bucket or cluster? • Key question: Is something wrong with Firehose or with the customers’ own configuration, say S3 bucket? • Key metrics: The DeliveryToS3.Success metric computes the sum of successful S3 Puts over sum of all S3 puts managed by Firehose. Similarly DeliveryToRedshift.Success metric computes sum of successful COPY commands over sum of all COPY commands. • If this trends below “1” it suggests potential issues with the S3 bucket configuration such that Firehose puts to S3 are failing.
  • 47. • Check Incoming.Bytes and Incoming.Records metrics to ensure that data is Put successfully. • Check DeliveryToS3.Success metric to ensure that Firehose is putting data to your S3 bucket. • Ensure that the S3 bucket specified in your delivery stream still exists. • Ensure that the IAM role specified in your delivery stream has access to your S3 bucket. Amazon Kinesis Firehose Troubleshooting Something key steps if data isn’t showing up in S3
  • 48. • Check Incoming.Bytes and Incoming.Records metrics to ensure that data is Put successfully, and the DeliveryToS3.Success metric to ensure that data is being staged in the S3 bucket. • Check DeliveryToRedshift.Success metric to ensure that Firehose has copied data from your S3 bucket to the cluster. • Check Redshift STL_LOAD_ERRORS table to verify the reason of the COPY failure. • Ensure that Redshift config in Firehose is accurate Amazon Kinesis Firehose Troubleshooting Something key steps if data isn’t showing up in Redshift
  • 50. Amazon Kinesis Analytics Analyze data streams continuously with standard SQL Apply SQL on streams: Easily connect to data streams and apply existing SQL skills. Build real-time applications: Perform continual processing on streaming big data with sub-second processing latencies Scale elastically: Elastically scales to match data throughput without any operator intervention. Announcement Only! Amazon Confidential Connect to Kinesis streams, Firehose delivery streams Run standard SQL queries against data streams Kinesis Analytics can send processed data to analytics tools so you can create alerts and respond in real-time
  • 52. Amazon Kinesis Streams Build your own custom applications that process or analyze streaming data Amazon Kinesis Firehose Easily load massive volumes of streaming data into Amazon S3 and Redshift Amazon Kinesis Analytics Easily analyze data streams using standard SQL queries Amazon Kinesis: Streaming data made easy Services make it easy to capture, deliver, and process streams on AWS Amazon Confidential
  • 53. Q&A