{-# LANGUAGE DeriveGeneric #-} {-# LANGUAGE DuplicateRecordFields #-} {-# LANGUAGE NamedFieldPuns #-} {-# LANGUAGE OverloadedStrings #-} {-# LANGUAGE RecordWildCards #-} {-# LANGUAGE StrictData #-} {-# LANGUAGE TypeFamilies #-} {-# LANGUAGE NoImplicitPrelude #-} {-# OPTIONS_GHC -fno-warn-unused-binds #-} {-# OPTIONS_GHC -fno-warn-unused-imports #-} {-# OPTIONS_GHC -fno-warn-unused-matches #-} -- Derived from AWS service descriptions, licensed under Apache 2.0. -- | -- Module : Amazonka.Kinesis.PutRecord -- Copyright : (c) 2013-2023 Brendan Hay -- License : Mozilla Public License, v. 2.0. -- Maintainer : Brendan Hay -- Stability : auto-generated -- Portability : non-portable (GHC extensions) -- -- Writes a single data record into an Amazon Kinesis data stream. Call -- @PutRecord@ to send data into the stream for real-time ingestion and -- subsequent processing, one record at a time. Each shard can support -- writes up to 1,000 records per second, up to a maximum data write total -- of 1 MiB per second. -- -- When invoking this API, it is recommended you use the @StreamARN@ input -- parameter rather than the @StreamName@ input parameter. -- -- You must specify the name of the stream that captures, stores, and -- transports the data; a partition key; and the data blob itself. -- -- The data blob can be any type of data; for example, a segment from a log -- file, geographic\/location data, website clickstream data, and so on. -- -- The partition key is used by Kinesis Data Streams to distribute data -- across shards. Kinesis Data Streams segregates the data records that -- belong to a stream into multiple shards, using the partition key -- associated with each data record to determine the shard to which a given -- data record belongs. -- -- Partition keys are Unicode strings, with a maximum length limit of 256 -- characters for each key. An MD5 hash function is used to map partition -- keys to 128-bit integer values and to map associated data records to -- shards using the hash key ranges of the shards. You can override hashing -- the partition key to determine the shard by explicitly specifying a hash -- value using the @ExplicitHashKey@ parameter. For more information, see -- -- in the /Amazon Kinesis Data Streams Developer Guide/. -- -- @PutRecord@ returns the shard ID of where the data record was placed and -- the sequence number that was assigned to the data record. -- -- Sequence numbers increase over time and are specific to a shard within a -- stream, not across all shards within a stream. To guarantee strictly -- increasing ordering, write serially to a shard and use the -- @SequenceNumberForOrdering@ parameter. For more information, see -- -- in the /Amazon Kinesis Data Streams Developer Guide/. -- -- After you write a record to a stream, you cannot modify that record or -- its order within the stream. -- -- If a @PutRecord@ request cannot be processed because of insufficient -- provisioned throughput on the shard involved in the request, @PutRecord@ -- throws @ProvisionedThroughputExceededException@. -- -- By default, data records are accessible for 24 hours from the time that -- they are added to a stream. You can use IncreaseStreamRetentionPeriod or -- DecreaseStreamRetentionPeriod to modify this retention period. module Amazonka.Kinesis.PutRecord ( -- * Creating a Request PutRecord (..), newPutRecord, -- * Request Lenses putRecord_explicitHashKey, putRecord_sequenceNumberForOrdering, putRecord_streamARN, putRecord_streamName, putRecord_data, putRecord_partitionKey, -- * Destructuring the Response PutRecordResponse (..), newPutRecordResponse, -- * Response Lenses putRecordResponse_encryptionType, putRecordResponse_httpStatus, putRecordResponse_shardId, putRecordResponse_sequenceNumber, ) where import qualified Amazonka.Core as Core import qualified Amazonka.Core.Lens.Internal as Lens import qualified Amazonka.Data as Data import Amazonka.Kinesis.Types import qualified Amazonka.Prelude as Prelude import qualified Amazonka.Request as Request import qualified Amazonka.Response as Response -- | Represents the input for @PutRecord@. -- -- /See:/ 'newPutRecord' smart constructor. data PutRecord = PutRecord' { -- | The hash value used to explicitly determine the shard the data record is -- assigned to by overriding the partition key hash. explicitHashKey :: Prelude.Maybe Prelude.Text, -- | Guarantees strictly increasing sequence numbers, for puts from the same -- client and to the same partition key. Usage: set the -- @SequenceNumberForOrdering@ of record /n/ to the sequence number of -- record /n-1/ (as returned in the result when putting record /n-1/). If -- this parameter is not set, records are coarsely ordered based on arrival -- time. sequenceNumberForOrdering :: Prelude.Maybe Prelude.Text, -- | The ARN of the stream. streamARN :: Prelude.Maybe Prelude.Text, -- | The name of the stream to put the data record into. streamName :: Prelude.Maybe Prelude.Text, -- | The data blob to put into the record, which is base64-encoded when the -- blob is serialized. When the data blob (the payload before -- base64-encoding) is added to the partition key size, the total size must -- not exceed the maximum record size (1 MiB). data' :: Data.Base64, -- | Determines which shard in the stream the data record is assigned to. -- Partition keys are Unicode strings with a maximum length limit of 256 -- characters for each key. Amazon Kinesis Data Streams uses the partition -- key as input to a hash function that maps the partition key and -- associated data to a specific shard. Specifically, an MD5 hash function -- is used to map partition keys to 128-bit integer values and to map -- associated data records to shards. As a result of this hashing -- mechanism, all data records with the same partition key map to the same -- shard within the stream. partitionKey :: Prelude.Text } deriving (Prelude.Eq, Prelude.Read, Prelude.Show, Prelude.Generic) -- | -- Create a value of 'PutRecord' with all optional fields omitted. -- -- Use or to modify other optional fields. -- -- The following record fields are available, with the corresponding lenses provided -- for backwards compatibility: -- -- 'explicitHashKey', 'putRecord_explicitHashKey' - The hash value used to explicitly determine the shard the data record is -- assigned to by overriding the partition key hash. -- -- 'sequenceNumberForOrdering', 'putRecord_sequenceNumberForOrdering' - Guarantees strictly increasing sequence numbers, for puts from the same -- client and to the same partition key. Usage: set the -- @SequenceNumberForOrdering@ of record /n/ to the sequence number of -- record /n-1/ (as returned in the result when putting record /n-1/). If -- this parameter is not set, records are coarsely ordered based on arrival -- time. -- -- 'streamARN', 'putRecord_streamARN' - The ARN of the stream. -- -- 'streamName', 'putRecord_streamName' - The name of the stream to put the data record into. -- -- 'data'', 'putRecord_data' - The data blob to put into the record, which is base64-encoded when the -- blob is serialized. When the data blob (the payload before -- base64-encoding) is added to the partition key size, the total size must -- not exceed the maximum record size (1 MiB).-- -- -- /Note:/ This 'Lens' automatically encodes and decodes Base64 data. -- -- The underlying isomorphism will encode to Base64 representation during -- -- serialisation, and decode from Base64 representation during deserialisation. -- -- This 'Lens' accepts and returns only raw unencoded data. -- -- 'partitionKey', 'putRecord_partitionKey' - Determines which shard in the stream the data record is assigned to. -- Partition keys are Unicode strings with a maximum length limit of 256 -- characters for each key. Amazon Kinesis Data Streams uses the partition -- key as input to a hash function that maps the partition key and -- associated data to a specific shard. Specifically, an MD5 hash function -- is used to map partition keys to 128-bit integer values and to map -- associated data records to shards. As a result of this hashing -- mechanism, all data records with the same partition key map to the same -- shard within the stream. newPutRecord :: -- | 'data'' Prelude.ByteString -> -- | 'partitionKey' Prelude.Text -> PutRecord newPutRecord pData_ pPartitionKey_ = PutRecord' { explicitHashKey = Prelude.Nothing, sequenceNumberForOrdering = Prelude.Nothing, streamARN = Prelude.Nothing, streamName = Prelude.Nothing, data' = Data._Base64 Lens.# pData_, partitionKey = pPartitionKey_ } -- | The hash value used to explicitly determine the shard the data record is -- assigned to by overriding the partition key hash. putRecord_explicitHashKey :: Lens.Lens' PutRecord (Prelude.Maybe Prelude.Text) putRecord_explicitHashKey = Lens.lens (\PutRecord' {explicitHashKey} -> explicitHashKey) (\s@PutRecord' {} a -> s {explicitHashKey = a} :: PutRecord) -- | Guarantees strictly increasing sequence numbers, for puts from the same -- client and to the same partition key. Usage: set the -- @SequenceNumberForOrdering@ of record /n/ to the sequence number of -- record /n-1/ (as returned in the result when putting record /n-1/). If -- this parameter is not set, records are coarsely ordered based on arrival -- time. putRecord_sequenceNumberForOrdering :: Lens.Lens' PutRecord (Prelude.Maybe Prelude.Text) putRecord_sequenceNumberForOrdering = Lens.lens (\PutRecord' {sequenceNumberForOrdering} -> sequenceNumberForOrdering) (\s@PutRecord' {} a -> s {sequenceNumberForOrdering = a} :: PutRecord) -- | The ARN of the stream. putRecord_streamARN :: Lens.Lens' PutRecord (Prelude.Maybe Prelude.Text) putRecord_streamARN = Lens.lens (\PutRecord' {streamARN} -> streamARN) (\s@PutRecord' {} a -> s {streamARN = a} :: PutRecord) -- | The name of the stream to put the data record into. putRecord_streamName :: Lens.Lens' PutRecord (Prelude.Maybe Prelude.Text) putRecord_streamName = Lens.lens (\PutRecord' {streamName} -> streamName) (\s@PutRecord' {} a -> s {streamName = a} :: PutRecord) -- | The data blob to put into the record, which is base64-encoded when the -- blob is serialized. When the data blob (the payload before -- base64-encoding) is added to the partition key size, the total size must -- not exceed the maximum record size (1 MiB).-- -- -- /Note:/ This 'Lens' automatically encodes and decodes Base64 data. -- -- The underlying isomorphism will encode to Base64 representation during -- -- serialisation, and decode from Base64 representation during deserialisation. -- -- This 'Lens' accepts and returns only raw unencoded data. putRecord_data :: Lens.Lens' PutRecord Prelude.ByteString putRecord_data = Lens.lens (\PutRecord' {data'} -> data') (\s@PutRecord' {} a -> s {data' = a} :: PutRecord) Prelude.. Data._Base64 -- | Determines which shard in the stream the data record is assigned to. -- Partition keys are Unicode strings with a maximum length limit of 256 -- characters for each key. Amazon Kinesis Data Streams uses the partition -- key as input to a hash function that maps the partition key and -- associated data to a specific shard. Specifically, an MD5 hash function -- is used to map partition keys to 128-bit integer values and to map -- associated data records to shards. As a result of this hashing -- mechanism, all data records with the same partition key map to the same -- shard within the stream. putRecord_partitionKey :: Lens.Lens' PutRecord Prelude.Text putRecord_partitionKey = Lens.lens (\PutRecord' {partitionKey} -> partitionKey) (\s@PutRecord' {} a -> s {partitionKey = a} :: PutRecord) instance Core.AWSRequest PutRecord where type AWSResponse PutRecord = PutRecordResponse request overrides = Request.postJSON (overrides defaultService) response = Response.receiveJSON ( \s h x -> PutRecordResponse' Prelude.<$> (x Data..?> "EncryptionType") Prelude.<*> (Prelude.pure (Prelude.fromEnum s)) Prelude.<*> (x Data..:> "ShardId") Prelude.<*> (x Data..:> "SequenceNumber") ) instance Prelude.Hashable PutRecord where hashWithSalt _salt PutRecord' {..} = _salt `Prelude.hashWithSalt` explicitHashKey `Prelude.hashWithSalt` sequenceNumberForOrdering `Prelude.hashWithSalt` streamARN `Prelude.hashWithSalt` streamName `Prelude.hashWithSalt` data' `Prelude.hashWithSalt` partitionKey instance Prelude.NFData PutRecord where rnf PutRecord' {..} = Prelude.rnf explicitHashKey `Prelude.seq` Prelude.rnf sequenceNumberForOrdering `Prelude.seq` Prelude.rnf streamARN `Prelude.seq` Prelude.rnf streamName `Prelude.seq` Prelude.rnf data' `Prelude.seq` Prelude.rnf partitionKey instance Data.ToHeaders PutRecord where toHeaders = Prelude.const ( Prelude.mconcat [ "X-Amz-Target" Data.=# ("Kinesis_20131202.PutRecord" :: Prelude.ByteString), "Content-Type" Data.=# ( "application/x-amz-json-1.1" :: Prelude.ByteString ) ] ) instance Data.ToJSON PutRecord where toJSON PutRecord' {..} = Data.object ( Prelude.catMaybes [ ("ExplicitHashKey" Data..=) Prelude.<$> explicitHashKey, ("SequenceNumberForOrdering" Data..=) Prelude.<$> sequenceNumberForOrdering, ("StreamARN" Data..=) Prelude.<$> streamARN, ("StreamName" Data..=) Prelude.<$> streamName, Prelude.Just ("Data" Data..= data'), Prelude.Just ("PartitionKey" Data..= partitionKey) ] ) instance Data.ToPath PutRecord where toPath = Prelude.const "/" instance Data.ToQuery PutRecord where toQuery = Prelude.const Prelude.mempty -- | Represents the output for @PutRecord@. -- -- /See:/ 'newPutRecordResponse' smart constructor. data PutRecordResponse = PutRecordResponse' { -- | The encryption type to use on the record. This parameter can be one of -- the following values: -- -- - @NONE@: Do not encrypt the records in the stream. -- -- - @KMS@: Use server-side encryption on the records in the stream using -- a customer-managed Amazon Web Services KMS key. encryptionType :: Prelude.Maybe EncryptionType, -- | The response's http status code. httpStatus :: Prelude.Int, -- | The shard ID of the shard where the data record was placed. shardId :: Prelude.Text, -- | The sequence number identifier that was assigned to the put data record. -- The sequence number for the record is unique across all records in the -- stream. A sequence number is the identifier associated with every record -- put into the stream. sequenceNumber :: Prelude.Text } deriving (Prelude.Eq, Prelude.Read, Prelude.Show, Prelude.Generic) -- | -- Create a value of 'PutRecordResponse' with all optional fields omitted. -- -- Use or to modify other optional fields. -- -- The following record fields are available, with the corresponding lenses provided -- for backwards compatibility: -- -- 'encryptionType', 'putRecordResponse_encryptionType' - The encryption type to use on the record. This parameter can be one of -- the following values: -- -- - @NONE@: Do not encrypt the records in the stream. -- -- - @KMS@: Use server-side encryption on the records in the stream using -- a customer-managed Amazon Web Services KMS key. -- -- 'httpStatus', 'putRecordResponse_httpStatus' - The response's http status code. -- -- 'shardId', 'putRecordResponse_shardId' - The shard ID of the shard where the data record was placed. -- -- 'sequenceNumber', 'putRecordResponse_sequenceNumber' - The sequence number identifier that was assigned to the put data record. -- The sequence number for the record is unique across all records in the -- stream. A sequence number is the identifier associated with every record -- put into the stream. newPutRecordResponse :: -- | 'httpStatus' Prelude.Int -> -- | 'shardId' Prelude.Text -> -- | 'sequenceNumber' Prelude.Text -> PutRecordResponse newPutRecordResponse pHttpStatus_ pShardId_ pSequenceNumber_ = PutRecordResponse' { encryptionType = Prelude.Nothing, httpStatus = pHttpStatus_, shardId = pShardId_, sequenceNumber = pSequenceNumber_ } -- | The encryption type to use on the record. This parameter can be one of -- the following values: -- -- - @NONE@: Do not encrypt the records in the stream. -- -- - @KMS@: Use server-side encryption on the records in the stream using -- a customer-managed Amazon Web Services KMS key. putRecordResponse_encryptionType :: Lens.Lens' PutRecordResponse (Prelude.Maybe EncryptionType) putRecordResponse_encryptionType = Lens.lens (\PutRecordResponse' {encryptionType} -> encryptionType) (\s@PutRecordResponse' {} a -> s {encryptionType = a} :: PutRecordResponse) -- | The response's http status code. putRecordResponse_httpStatus :: Lens.Lens' PutRecordResponse Prelude.Int putRecordResponse_httpStatus = Lens.lens (\PutRecordResponse' {httpStatus} -> httpStatus) (\s@PutRecordResponse' {} a -> s {httpStatus = a} :: PutRecordResponse) -- | The shard ID of the shard where the data record was placed. putRecordResponse_shardId :: Lens.Lens' PutRecordResponse Prelude.Text putRecordResponse_shardId = Lens.lens (\PutRecordResponse' {shardId} -> shardId) (\s@PutRecordResponse' {} a -> s {shardId = a} :: PutRecordResponse) -- | The sequence number identifier that was assigned to the put data record. -- The sequence number for the record is unique across all records in the -- stream. A sequence number is the identifier associated with every record -- put into the stream. putRecordResponse_sequenceNumber :: Lens.Lens' PutRecordResponse Prelude.Text putRecordResponse_sequenceNumber = Lens.lens (\PutRecordResponse' {sequenceNumber} -> sequenceNumber) (\s@PutRecordResponse' {} a -> s {sequenceNumber = a} :: PutRecordResponse) instance Prelude.NFData PutRecordResponse where rnf PutRecordResponse' {..} = Prelude.rnf encryptionType `Prelude.seq` Prelude.rnf httpStatus `Prelude.seq` Prelude.rnf shardId `Prelude.seq` Prelude.rnf sequenceNumber