Copyright | (c) 2013-2015 Brendan Hay |
---|---|
License | Mozilla Public License, v. 2.0. |
Maintainer | Brendan Hay <brendan.g.hay@gmail.com> |
Stability | auto-generated |
Portability | non-portable (GHC extensions) |
Safe Haskell | None |
Language | Haskell2010 |
Writes a single data record into an Amazon Kinesis Firehose delivery stream. To write multiple data records into a delivery stream, use PutRecordBatch. Applications using these operations are referred to as producers.
By default, each delivery stream can take in up to 2,000 transactions per second, 5,000 records per second, or 5 MB per second. Note that if you use PutRecord and PutRecordBatch, the limits are an aggregate across these two operations for each delivery stream. For more information about limits and how to request an increase, see Amazon Kinesis Firehose Limits.
You must specify the name of the delivery stream and the data record when using PutRecord. The data record consists of a data blob that can be up to 1,000 KB in size, and any kind of data, for example, a segment from a log file, geographic location data, web site clickstream data, etc.
Amazon Kinesis Firehose buffers records before delivering them to the destination. To disambiguate the data blobs at the destination, a common solution is to use delimiters in the data, such as a newline ('\\n') or some other character unique within the data. This allows the consumer application(s) to parse individual data items when reading the data from the destination.
Amazon Kinesis Firehose does not maintain data record ordering. If the destination data needs to be re-ordered by the consumer application, the producer should include some form of sequence number in each data record.
The PutRecord operation returns a RecordId
, which is a unique string
assigned to each record. Producer applications can use this ID for
purposes such as auditability and investigation.
If the PutRecord operation throws a ServiceUnavailableException
, back
off and retry. If the exception persists, it is possible that the
throughput limits have been exceeded for the delivery stream.
Data records sent to Amazon Kinesis Firehose are stored for 24 hours from the time they are added to a delivery stream as it attempts to send the records to the destination. If the destination is unreachable for more than 24 hours, the data is no longer available.
See: AWS API Reference for PutRecord.
- putRecord :: Text -> Record -> PutRecord
- data PutRecord
- prDeliveryStreamName :: Lens' PutRecord Text
- prRecord :: Lens' PutRecord Record
- putRecordResponse :: Int -> Text -> PutRecordResponse
- data PutRecordResponse
- prrsResponseStatus :: Lens' PutRecordResponse Int
- prrsRecordId :: Lens' PutRecordResponse Text
Creating a Request
Creates a value of PutRecord
with the minimum fields required to make a request.
Use one of the following lenses to modify other fields as desired:
Contains the parameters for PutRecord.
See: putRecord
smart constructor.
Request Lenses
prDeliveryStreamName :: Lens' PutRecord Text Source
The name of the delivery stream.
Destructuring the Response
Creates a value of PutRecordResponse
with the minimum fields required to make a request.
Use one of the following lenses to modify other fields as desired:
data PutRecordResponse Source
Contains the output of PutRecord.
See: putRecordResponse
smart constructor.
Response Lenses
prrsResponseStatus :: Lens' PutRecordResponse Int Source
The response status code.
prrsRecordId :: Lens' PutRecordResponse Text Source
The ID of the record.