h& #L$      !"#$%&'()*+,-./0123456789:;<=>?@ABCDEFGHIJKLMNOPQRSTUVWXYZ[\]^_`abcdefghijklmnopqrstuvwxyz{|}~                                                                                                               !!!!!!!!!!!!!!!!"""""""""""""""""#############$$$$$$$$$$$$$$%%%%%%%%%%%%%%&&&&&&&&&&&&&&&&&&&&&&&&'''''''''''''''''((((((((((((((((())))))))))))))))))))))))))))******************++++++++++++++++++,,,,,,,,,,,,,,,,,,------------------..................//////////////00000000000000000011111111111111111111111111111111112222222222222222222222222222233333333333333333333444444444444444444555555555555555555556666666666666666667777777777777778888888888888888888888888888888888999999999999999999999999999::::::::::::: : : : ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; < < < < < < < < < < < < < < = = = = = = = = = = = = = = = = = = > > > > > > > > > > > > > > > > > > > > > > > > ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ A A A A A A A A A A A A A A A A A A B B B B B B B B B B B B B B C C C C C C C C C C C C C C D D D D D D D D D D D D D D E E E E E E E E E E E E E E F F F F F F F F F F F F F F F F F G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G H H H H H H H H H H H H H H H H H H H H H H H I I I I I I I I I I I I I I J J J J J J J J J J J J J J J J J J J J J J K K K K K K K K K K K K K K K K K K L L L L L L L L L L L L L L L L L L M M M M M M M M M M M M M M M M M M M M M M M M M M N N N N N N N N N N N N N N O O O O O O O O O O O O O O O O P P P P P P P P P P P P P P P P P P P P P P P P Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q Q R R R R R R R R R R R R R R R R R R R R R R R S S S S S S S S S S S S S S S S S S S S S T T T T T T T T T T T T T T T T T T T T T T T T T T T T U U U U U U U U U U U U U U U U U U U U V V V V V V V V V V V V V V W W W W W W W W W W W W W W W X X X X X X X X X X X XXXXXXXYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYZZZZZZZZZZZZZZZZZZZZZZZZ[[[[[[[[[[[[[[\\\\\\\\\\\\\\]]]]]]]]]]]]]]]]]]^^^^^^^^^^^^^^^____________________________________````````````````````````````````````aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaabbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbcccccccccccccccccccccccccccccccccccccccccddddddddddddddddddddddddddddddddddddddeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeefffffffffffffffffffffffffffffffffffffffffffffgggggggggggggggggggggggggggggggggggghhhhhhhhhhhhhhhhhiiiiiiiiiiiiiiiiiiiiiiiiiiiijjjjjjjjjjjjjjjjjjjjjjjjjjkkkkkkkkkkkkkkkkkkkkkkkkkkllllllllllllllllllllllllllllllmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnooooooooooooooooooooooooooooooppppppppppppppppppppppppppppppqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrssssssssssssssssssssssssssssssssssssssssssssssssttttttttttttttttttttttttttttttttttttttttttttttttuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{||||||||||||||||||||||||||||||||}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~                                                                                                                                !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""################################################################################################################################$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?bW(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';gamazonka-comprehendDescribes an error that occurred while processing a document in a batch. The operation returns on BatchItemError3 object for each document that contained an error.See:  smart constructor.amazonka-comprehend$The numeric error code of the error.amazonka-comprehend A text description of the error.amazonka-comprehend7The zero-based index of the document in the input list.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, ' - The numeric error code of the error., # - A text description of the error.,  : - The zero-based index of the document in the input list.amazonka-comprehend$The numeric error code of the error.amazonka-comprehend A text description of the error. amazonka-comprehend7The zero-based index of the document in the input list.   (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?he(,+)*(,+)*,+(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';qp @amazonka-comprehendThe bounding box around the detected page or around an element on a document page. The left (x-coordinate) and top (y-coordinate) are coordinates that represent the top and left sides of the bounding box. Note that the upper-left corner of the image is the origin (0,0).!For additional information, see  https://docs.aws.amazon.com/textract/latest/dg/API_BoundingBox.html BoundingBox' in the Amazon Textract API reference.See: F smart constructor.Bamazonka-comprehendThe height of the bounding box as a ratio of the overall document page height.Camazonka-comprehendThe left coordinate of the bounding box as a ratio of overall document page width.Damazonka-comprehendThe top coordinate of the bounding box as a ratio of overall document page height.Eamazonka-comprehendThe width of the bounding box as a ratio of the overall document page width.Famazonka-comprehendCreate a value of @" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:B, G - The height of the bounding box as a ratio of the overall document page height.C, H - The left coordinate of the bounding box as a ratio of overall document page width.D, I - The top coordinate of the bounding box as a ratio of overall document page height.E, J - The width of the bounding box as a ratio of the overall document page width.Gamazonka-comprehendThe height of the bounding box as a ratio of the overall document page height.Hamazonka-comprehendThe left coordinate of the bounding box as a ratio of overall document page width.Iamazonka-comprehendThe top coordinate of the bounding box as a ratio of overall document page height.Jamazonka-comprehendThe width of the bounding box as a ratio of the overall document page width. @DEBCAFGHIJ @DEBCAFGHIJ(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';vRamazonka-comprehend&Nested block contained within a block.See: W smart constructor.Tamazonka-comprehend?Offset of the start of the child block within its parent block.Uamazonka-comprehend&Unique identifier for the child block.Vamazonka-comprehend=Offset of the end of the child block within its parent block.Wamazonka-comprehendCreate a value of R" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:T, X - Offset of the start of the child block within its parent block.U, Y) - Unique identifier for the child block.V, Z - Offset of the end of the child block within its parent block.Xamazonka-comprehend?Offset of the start of the child block within its parent block.Yamazonka-comprehend&Unique identifier for the child block.Zamazonka-comprehend=Offset of the end of the child block within its parent block. RVUTSWXYZ RVUTSWXYZ(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';| bamazonka-comprehendA reference to a block.See: h smart constructor.damazonka-comprehend9Offset of the start of the block within its parent block.eamazonka-comprehend Unique identifier for the block.famazonka-comprehend'List of child blocks within this block.gamazonka-comprehend7Offset of the end of the block within its parent block.hamazonka-comprehendCreate a value of b" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:b, i< - Offset of the start of the block within its parent block.e, j# - Unique identifier for the block.f, k* - List of child blocks within this block.b, l: - Offset of the end of the block within its parent block.iamazonka-comprehend9Offset of the start of the block within its parent block.jamazonka-comprehend Unique identifier for the block.kamazonka-comprehend'List of child blocks within this block.lamazonka-comprehend7Offset of the end of the block within its parent block. bfegdchijkl bfegdchijkl(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';tamazonka-comprehendDescribes the result metrics for the test data associated with an documentation classifier.See: ~ smart constructor.vamazonka-comprehendThe fraction of the labels that were correct recognized. It is computed by dividing the number of labels in the test documents that were correctly recognized by the total number of labels in the test documents.wamazonka-comprehendA measure of how accurate the classifier results are for the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. The highest score is 1, and the worst score is 0.xamazonka-comprehendIndicates the fraction of labels that are incorrectly predicted. Also seen as the fraction of wrong labels compared to the total number of labels. Scores closer to zero are better.yamazonka-comprehendA measure of how accurate the classifier results are for the test data. It is a combination of the Micro Precision and  Micro Recall values. The  Micro F1Score is the harmonic mean of the two scores. The highest score is 1, and the worst score is 0.zamazonka-comprehendA measure of the usefulness of the recognizer results in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones. Unlike the Precision metric which comes from averaging the precision of all available labels, this is based on the overall score of all precision scores added together.{amazonka-comprehendA measure of how complete the classifier results are for the test data. High recall means that the classifier returned most of the relevant results. Specifically, this indicates how many of the correct categories in the text that the model can predict. It is a percentage of correct categories in the text that can found. Instead of averaging the recall scores of all labels (as with Recall), micro Recall is based on the overall score of all recall scores added together.|amazonka-comprehendA measure of the usefulness of the classifier results in the test data. High precision means that the classifier returned substantially more relevant results than irrelevant ones.}amazonka-comprehendA measure of how complete the classifier results are for the test data. High recall means that the classifier returned most of the relevant results.~amazonka-comprehendCreate a value of t" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:v,  - The fraction of the labels that were correct recognized. It is computed by dividing the number of labels in the test documents that were correctly recognized by the total number of labels in the test documents.w,  - A measure of how accurate the classifier results are for the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. The highest score is 1, and the worst score is 0.x,  - Indicates the fraction of labels that are incorrectly predicted. Also seen as the fraction of wrong labels compared to the total number of labels. Scores closer to zero are better.y,  - A measure of how accurate the classifier results are for the test data. It is a combination of the Micro Precision and  Micro Recall values. The  Micro F1Score is the harmonic mean of the two scores. The highest score is 1, and the worst score is 0.z,  - A measure of the usefulness of the recognizer results in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones. Unlike the Precision metric which comes from averaging the precision of all available labels, this is based on the overall score of all precision scores added together.{,  - A measure of how complete the classifier results are for the test data. High recall means that the classifier returned most of the relevant results. Specifically, this indicates how many of the correct categories in the text that the model can predict. It is a percentage of correct categories in the text that can found. Instead of averaging the recall scores of all labels (as with Recall), micro Recall is based on the overall score of all recall scores added together.|,  - A measure of the usefulness of the classifier results in the test data. High precision means that the classifier returned substantially more relevant results than irrelevant ones.},  - A measure of how complete the classifier results are for the test data. High recall means that the classifier returned most of the relevant results.amazonka-comprehendThe fraction of the labels that were correct recognized. It is computed by dividing the number of labels in the test documents that were correctly recognized by the total number of labels in the test documents.amazonka-comprehendA measure of how accurate the classifier results are for the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. The highest score is 1, and the worst score is 0.amazonka-comprehendIndicates the fraction of labels that are incorrectly predicted. Also seen as the fraction of wrong labels compared to the total number of labels. Scores closer to zero are better.amazonka-comprehendA measure of how accurate the classifier results are for the test data. It is a combination of the Micro Precision and  Micro Recall values. The  Micro F1Score is the harmonic mean of the two scores. The highest score is 1, and the worst score is 0.amazonka-comprehendA measure of the usefulness of the recognizer results in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones. Unlike the Precision metric which comes from averaging the precision of all available labels, this is based on the overall score of all precision scores added together.amazonka-comprehendA measure of how complete the classifier results are for the test data. High recall means that the classifier returned most of the relevant results. Specifically, this indicates how many of the correct categories in the text that the model can predict. It is a percentage of correct categories in the text that can found. Instead of averaging the recall scores of all labels (as with Recall), micro Recall is based on the overall score of all recall scores added together.amazonka-comprehendA measure of the usefulness of the classifier results in the test data. High precision means that the classifier returned substantially more relevant results than irrelevant ones.amazonka-comprehendA measure of how complete the classifier results are for the test data. High recall means that the classifier returned most of the relevant results.t}|{zyxwvu~t}|{zyxwvu~(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';3 amazonka-comprehend1Provides information about a document classifier.See:  smart constructor.amazonka-comprehendDescribes the result metrics for the test data associated with an documentation classifier.amazonka-comprehend'The number of labels in the input data.amazonka-comprehendThe number of documents in the input data that were used to test the classifier. Typically this is 10 to 20 percent of the input documents, up to 10,000 documents.amazonka-comprehendThe number of documents in the input data that were used to train the classifier. Typically this is 80 to 90 percent of the input documents.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Describes the result metrics for the test data associated with an documentation classifier., * - The number of labels in the input data.,  - The number of documents in the input data that were used to test the classifier. Typically this is 10 to 20 percent of the input documents, up to 10,000 documents.,  - The number of documents in the input data that were used to train the classifier. Typically this is 80 to 90 percent of the input documents.amazonka-comprehendDescribes the result metrics for the test data associated with an documentation classifier.amazonka-comprehend'The number of labels in the input data.amazonka-comprehendThe number of documents in the input data that were used to test the classifier. Typically this is 10 to 20 percent of the input documents, up to 10,000 documents.amazonka-comprehendThe number of documents in the input data that were used to train the classifier. Typically this is 80 to 90 percent of the input documents.   (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-comprehendSpecifies the class that categorizes the document being analyzedSee:  smart constructor.amazonka-comprehendThe name of the class.amazonka-comprehendPage number in the input document. This field is present in the response only if your request includes the Byte parameter.amazonka-comprehendThe confidence score that Amazon Comprehend has this class correctly attributed.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The name of the class.,  - Page number in the input document. This field is present in the response only if your request includes the Byte parameter.,  - The confidence score that Amazon Comprehend has this class correctly attributed.amazonka-comprehendThe name of the class.amazonka-comprehendPage number in the input document. This field is present in the response only if your request includes the Byte parameter.amazonka-comprehendThe confidence score that Amazon Comprehend has this class correctly attributed.   (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?h (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?$ (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-comprehendProvides output results configuration parameters for custom classifier jobs.See:  smart constructor.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"KMS Key Alias: "alias/ExampleAlias"ARN of a KMS Key Alias: 7"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias"amazonka-comprehendWhen you use the OutputDataConfig object while creating a custom classifier, you specify the Amazon S3 location where you want to write the confusion matrix. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of this output file.When the custom classifier job is finished, the service creates the output file in a directory specific to the job. The S3Uri9 field contains the location of the output file, called  output.tar.gz. It is a compressed archive that contains the confusion matrix.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"KMS Key Alias: "alias/ExampleAlias"ARN of a KMS Key Alias: 7"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias",  - When you use the OutputDataConfig object while creating a custom classifier, you specify the Amazon S3 location where you want to write the confusion matrix. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of this output file.When the custom classifier job is finished, the service creates the output file in a directory specific to the job. The S3Uri9 field contains the location of the output file, called  output.tar.gz. It is a compressed archive that contains the confusion matrix.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"KMS Key Alias: "alias/ExampleAlias"ARN of a KMS Key Alias: 7"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias"amazonka-comprehendWhen you use the OutputDataConfig object while creating a custom classifier, you specify the Amazon S3 location where you want to write the confusion matrix. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of this output file.When the custom classifier job is finished, the service creates the output file in a directory specific to the job. The S3Uri9 field contains the location of the output file, called  output.tar.gz. It is a compressed archive that contains the confusion matrix. (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-comprehendSpecifies one of the label or labels that categorize the document being analyzed.See:  smart constructor.amazonka-comprehendThe name of the label.amazonka-comprehendPage number where the label occurs. This field is present in the response only if your request includes the Byte parameter.amazonka-comprehendThe confidence score that Amazon Comprehend has this label correctly attributed.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The name of the label.,  - Page number where the label occurs. This field is present in the response only if your request includes the Byte parameter.,  - The confidence score that Amazon Comprehend has this label correctly attributed.amazonka-comprehendThe name of the label.amazonka-comprehendPage number where the label occurs. This field is present in the response only if your request includes the Byte parameter.amazonka-comprehendThe confidence score that Amazon Comprehend has this label correctly attributed.  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?¼(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?$amazonka-comprehendSpecifies the type of Amazon Textract features to apply. If you chose TEXTRACT_ANALYZE_DOCUMENT as the read action, you must specify one or both of the following values:TABLES - Returns additional information about any tables that are detected in the input document.FORMS - Returns additional information about any forms that are detected in the input document.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';'amazonka-comprehendProvides configuration parameters to override the default actions for extracting text from PDF documents and image files.By default, Amazon Comprehend performs the following actions to extract text from files, based on the input file type: Word files. - Amazon Comprehend parser extracts the text.Digital PDF files. - Amazon Comprehend parser extracts the text.!Image files and scanned PDF files3 - Amazon Comprehend uses the Amazon Textract DetectDocumentText API to extract the text.DocumentReaderConfig2 does not apply to plain text files or Word files.For image files and PDF documents, you can override these default actions using the fields listed below. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/detecting-cer.html#detecting-cer-pdfSetting text extraction options.See:  smart constructor.amazonka-comprehendDetermines the text extraction actions for PDF files. Enter one of the following values:SERVICE_DEFAULT - use the Amazon Comprehend service defaults for PDF files.FORCE_DOCUMENT_READ_ACTION - Amazon Comprehend uses the Textract API specified by DocumentReadAction for all PDF files, including digital PDF files.amazonka-comprehendSpecifies the type of Amazon Textract features to apply. If you chose TEXTRACT_ANALYZE_DOCUMENT as the read action, you must specify one or both of the following values:TABLES - Returns information about any tables that are detected in the input document.FORMS - Returns information and the data from any forms that are detected in the input document.amazonka-comprehendThis field defines the Amazon Textract API operation that Amazon Comprehend uses to extract text from PDF files and image files. Enter one of the following values:TEXTRACT_DETECT_DOCUMENT_TEXT/ - The Amazon Comprehend service uses the DetectDocumentText API operation.TEXTRACT_ANALYZE_DOCUMENT/ - The Amazon Comprehend service uses the AnalyzeDocument API operation.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Determines the text extraction actions for PDF files. Enter one of the following values:SERVICE_DEFAULT - use the Amazon Comprehend service defaults for PDF files.FORCE_DOCUMENT_READ_ACTION - Amazon Comprehend uses the Textract API specified by DocumentReadAction for all PDF files, including digital PDF files.,  - Specifies the type of Amazon Textract features to apply. If you chose TEXTRACT_ANALYZE_DOCUMENT as the read action, you must specify one or both of the following values:TABLES - Returns information about any tables that are detected in the input document.FORMS - Returns information and the data from any forms that are detected in the input document.,  - This field defines the Amazon Textract API operation that Amazon Comprehend uses to extract text from PDF files and image files. Enter one of the following values:TEXTRACT_DETECT_DOCUMENT_TEXT/ - The Amazon Comprehend service uses the DetectDocumentText API operation.TEXTRACT_ANALYZE_DOCUMENT/ - The Amazon Comprehend service uses the AnalyzeDocument API operation.amazonka-comprehendDetermines the text extraction actions for PDF files. Enter one of the following values:SERVICE_DEFAULT - use the Amazon Comprehend service defaults for PDF files.FORCE_DOCUMENT_READ_ACTION - Amazon Comprehend uses the Textract API specified by DocumentReadAction for all PDF files, including digital PDF files.amazonka-comprehendSpecifies the type of Amazon Textract features to apply. If you chose TEXTRACT_ANALYZE_DOCUMENT as the read action, you must specify one or both of the following values:TABLES - Returns information about any tables that are detected in the input document.FORMS - Returns information and the data from any forms that are detected in the input document.amazonka-comprehendThis field defines the Amazon Textract API operation that Amazon Comprehend uses to extract text from PDF files and image files. Enter one of the following values:TEXTRACT_DETECT_DOCUMENT_TEXT/ - The Amazon Comprehend service uses the DetectDocumentText API operation.TEXTRACT_ANALYZE_DOCUMENT/ - The Amazon Comprehend service uses the AnalyzeDocument API operation.amazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";? (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';ޒamazonka-comprehend,Document type for each page in the document.See:  smart constructor.amazonka-comprehend Page number.amazonka-comprehendDocument type.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Page number.,  - Document type.amazonka-comprehend Page number.amazonka-comprehendDocument type.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-comprehendReturns the code for the dominant language in the input text and the level of confidence that Amazon Comprehend has in the accuracy of the detection.See:  smart constructor.amazonka-comprehendThe RFC 5646 language code for the dominant language. For more information about RFC 5646, see  #https://tools.ietf.org/html/rfc5646Tags for Identifying Languages on the  IETF Tools web site.amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of the detection.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The RFC 5646 language code for the dominant language. For more information about RFC 5646, see  #https://tools.ietf.org/html/rfc5646Tags for Identifying Languages on the  IETF Tools web site.,  - The level of confidence that Amazon Comprehend has in the accuracy of the detection.amazonka-comprehendThe RFC 5646 language code for the dominant language. For more information about RFC 5646, see  #https://tools.ietf.org/html/rfc5646Tags for Identifying Languages on the  IETF Tools web site.amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of the detection.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-comprehendThe result of calling the operation. The operation returns one object for each document that is successfully processed by the operation.See:  smart constructor.amazonka-comprehend7The zero-based index of the document in the input list.amazonka-comprehendOne or more DominantLanguage objects describing the dominant languages in the document.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, : - The zero-based index of the document in the input list.,  - One or more DominantLanguage objects describing the dominant languages in the document.amazonka-comprehend7The zero-based index of the document in the input list.amazonka-comprehendOne or more DominantLanguage objects describing the dominant languages in the document.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";? (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';(amazonka-comprehendSpecifies information about the specified endpoint. For information about endpoints, see  https://docs.aws.amazon.com/comprehend/latest/dg/manage-endpoints.htmlManaging endpoints.See:  smart constructor.amazonka-comprehend+The creation date and time of the endpoint.amazonka-comprehendThe number of inference units currently used by the model using this endpoint.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).amazonka-comprehendData access role ARN to use in case the new model is encrypted with a customer KMS key.amazonka-comprehendThe desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.amazonka-comprehendARN of the new model to use for updating an existing endpoint. This ARN is going to be different from the model ARN when the update is in progressamazonka-comprehend1The Amazon Resource Number (ARN) of the endpoint.amazonka-comprehend6The date and time that the endpoint was last modified.amazonka-comprehend+Specifies a reason for failure in cases of Failed status.amazonka-comprehendThe Amazon Resource Number (ARN) of the model to which the endpoint is attached.amazonka-comprehendSpecifies the status of the endpoint. Because the endpoint updates and creation are asynchronous, so customers will need to wait for the endpoint to be Ready) status before making inference requests.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, . - The creation date and time of the endpoint.,  - The number of inference units currently used by the model using this endpoint.,  - The Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).,  - Data access role ARN to use in case the new model is encrypted with a customer KMS key.,  - The desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.,  - ARN of the new model to use for updating an existing endpoint. This ARN is going to be different from the model ARN when the update is in progress, 4 - The Amazon Resource Number (ARN) of the endpoint., 9 - The date and time that the endpoint was last modified., . - Specifies a reason for failure in cases of Failed status.,  - The Amazon Resource Number (ARN) of the model to which the endpoint is attached.,  - Specifies the status of the endpoint. Because the endpoint updates and creation are asynchronous, so customers will need to wait for the endpoint to be Ready) status before making inference requests.amazonka-comprehend+The creation date and time of the endpoint.amazonka-comprehendThe number of inference units currently used by the model using this endpoint.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).amazonka-comprehendData access role ARN to use in case the new model is encrypted with a customer KMS key.amazonka-comprehendThe desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.amazonka-comprehendARN of the new model to use for updating an existing endpoint. This ARN is going to be different from the model ARN when the update is in progressamazonka-comprehend1The Amazon Resource Number (ARN) of the endpoint.amazonka-comprehend6The date and time that the endpoint was last modified.amazonka-comprehend+Specifies a reason for failure in cases of Failed status.amazonka-comprehendThe Amazon Resource Number (ARN) of the model to which the endpoint is attached.amazonka-comprehendSpecifies the status of the endpoint. Because the endpoint updates and creation are asynchronous, so customers will need to wait for the endpoint to be Ready) status before making inference requests.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';  amazonka-comprehendThe filter used to determine which endpoints are returned. You can filter jobs on their name, model, status, or the date and time that they were created. You can only set one filter at a time.See:  smart constructor.amazonka-comprehendSpecifies a date after which the returned endpoint or endpoints were created.amazonka-comprehendSpecifies a date before which the returned endpoint or endpoints were created.amazonka-comprehendThe Amazon Resource Number (ARN) of the model to which the endpoint is attached.amazonka-comprehendSpecifies the status of the endpoint being returned. Possible values are: Creating, Ready, Updating, Deleting, Failed.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies a date after which the returned endpoint or endpoints were created.,  - Specifies a date before which the returned endpoint or endpoints were created.,  - The Amazon Resource Number (ARN) of the model to which the endpoint is attached.,  - Specifies the status of the endpoint being returned. Possible values are: Creating, Ready, Updating, Deleting, Failed.amazonka-comprehendSpecifies a date after which the returned endpoint or endpoints were created.amazonka-comprehendSpecifies a date before which the returned endpoint or endpoints were created.amazonka-comprehendThe Amazon Resource Number (ARN) of the model to which the endpoint is attached.amazonka-comprehendSpecifies the status of the endpoint being returned. Possible values are: Creating, Ready, Updating, Deleting, Failed.  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-comprehend>Describes the annotations associated with a entity recognizer.See:  smart constructor.amazonka-comprehendSpecifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.amazonka-comprehendSpecifies the Amazon S3 location where the annotations for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.,  - Specifies the Amazon S3 location where the annotations for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.amazonka-comprehendSpecifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.amazonka-comprehendSpecifies the Amazon S3 location where the annotations for an entity recognizer are located. The URI must be in the same region as the API endpoint that you are calling.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?h(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';>amazonka-comprehendDescribes the entity recognizer submitted with an entity recognizer.See:  smart constructor.amazonka-comprehendSpecifies the Amazon S3 location where the entity list is located. The URI must be in the same region as the API endpoint that you are calling.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Specifies the Amazon S3 location where the entity list is located. The URI must be in the same region as the API endpoint that you are calling.amazonka-comprehendSpecifies the Amazon S3 location where the entity list is located. The URI must be in the same region as the API endpoint that you are calling.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';"?amazonka-comprehendDetailed information about the accuracy of an entity recognizer.See:  smart constructor.amazonka-comprehendA measure of how accurate the recognizer results are for the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. For plain text entity recognizer models, the range is 0 to 100, where 100 is the best score. For PDF/Word entity recognizer models, the range is 0 to 1, where 1 is the best score.amazonka-comprehendA measure of the usefulness of the recognizer results in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones.amazonka-comprehendA measure of how complete the recognizer results are for the test data. High recall means that the recognizer returned most of the relevant results.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A measure of how accurate the recognizer results are for the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. For plain text entity recognizer models, the range is 0 to 100, where 100 is the best score. For PDF/Word entity recognizer models, the range is 0 to 1, where 1 is the best score.,  - A measure of the usefulness of the recognizer results in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones.,  - A measure of how complete the recognizer results are for the test data. High recall means that the recognizer returned most of the relevant results.amazonka-comprehendA measure of how accurate the recognizer results are for the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. For plain text entity recognizer models, the range is 0 to 100, where 100 is the best score. For PDF/Word entity recognizer models, the range is 0 to 1, where 1 is the best score.amazonka-comprehendA measure of the usefulness of the recognizer results in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones.amazonka-comprehendA measure of how complete the recognizer results are for the test data. High recall means that the recognizer returned most of the relevant results.  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?# (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';0Famazonka-comprehend%Provides information about an entity.See:  smart constructor.amazonka-comprehendThe zero-based offset from the beginning of the source text to the first character in the entity.'This field is empty for non-text input.amazonka-comprehendA reference to each block for this entity. This field is empty for plain-text input.amazonka-comprehendThe zero-based offset from the beginning of the source text to the last character in the entity.'This field is empty for non-text input.amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of the detection.amazonka-comprehendThe text of the entity.amazonka-comprehendThe entity type. For entity detection using the built-in model, this field contains one of the standard entity types listed below.For custom entity detection, this field contains one of the entity types that you specified when you trained your custom model.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The zero-based offset from the beginning of the source text to the first character in the entity.'This field is empty for non-text input.,  - A reference to each block for this entity. This field is empty for plain-text input.,  - The zero-based offset from the beginning of the source text to the last character in the entity.'This field is empty for non-text input.,  - The level of confidence that Amazon Comprehend has in the accuracy of the detection.,  - The text of the entity.,  - The entity type. For entity detection using the built-in model, this field contains one of the standard entity types listed below.For custom entity detection, this field contains one of the entity types that you specified when you trained your custom model.amazonka-comprehendThe zero-based offset from the beginning of the source text to the first character in the entity.'This field is empty for non-text input.amazonka-comprehendA reference to each block for this entity. This field is empty for plain-text input.amazonka-comprehendThe zero-based offset from the beginning of the source text to the last character in the entity.'This field is empty for non-text input.amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of the detection.amazonka-comprehendThe text of the entity.amazonka-comprehendThe entity type. For entity detection using the built-in model, this field contains one of the standard entity types listed below.For custom entity detection, this field contains one of the entity types that you specified when you trained your custom model.(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';5amazonka-comprehendThe result of calling the operation. The operation returns one object for each document that is successfully processed by the operation.See:  smart constructor.amazonka-comprehendOne or more Entity objects, one for each entity detected in the document.amazonka-comprehend7The zero-based index of the document in the input list.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - One or more Entity objects, one for each entity detected in the document., : - The zero-based index of the document in the input list.amazonka-comprehendOne or more Entity objects, one for each entity detected in the document.amazonka-comprehend7The zero-based index of the document in the input list. (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';A amazonka-comprehendDetailed information about the accuracy of an entity recognizer for a specific entity type.See:  smart constructor.amazonka-comprehendA measure of how accurate the recognizer results are for a specific entity type in the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. The highest score is 1, and the worst score is 0.amazonka-comprehendA measure of the usefulness of the recognizer results for a specific entity type in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones.amazonka-comprehendA measure of how complete the recognizer results are for a specific entity type in the test data. High recall means that the recognizer returned most of the relevant results.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A measure of how accurate the recognizer results are for a specific entity type in the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. The highest score is 1, and the worst score is 0.,  - A measure of the usefulness of the recognizer results for a specific entity type in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones.,  - A measure of how complete the recognizer results are for a specific entity type in the test data. High recall means that the recognizer returned most of the relevant results.amazonka-comprehendA measure of how accurate the recognizer results are for a specific entity type in the test data. It is derived from the  Precision and Recall values. The F1Score is the harmonic average of the two scores. The highest score is 1, and the worst score is 0.amazonka-comprehendA measure of the usefulness of the recognizer results for a specific entity type in the test data. High precision means that the recognizer returned substantially more relevant results than irrelevant ones.amazonka-comprehendA measure of how complete the recognizer results are for a specific entity type in the test data. High recall means that the recognizer returned most of the relevant results.  !(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';H amazonka-comprehendIndividual item from the list of entity types in the metadata of an entity recognizer.See:  smart constructor.amazonka-comprehendDetailed information about the accuracy of the entity recognizer for a specific item on the list of entity types.amazonka-comprehendIndicates the number of times the given entity type was seen in the training data.amazonka-comprehendType of entity from the list of entity types in the metadata of an entity recognizer.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Detailed information about the accuracy of the entity recognizer for a specific item on the list of entity types.,  - Indicates the number of times the given entity type was seen in the training data.,  - Type of entity from the list of entity types in the metadata of an entity recognizer.amazonka-comprehendDetailed information about the accuracy of the entity recognizer for a specific item on the list of entity types.amazonka-comprehendIndicates the number of times the given entity type was seen in the training data.amazonka-comprehendType of entity from the list of entity types in the metadata of an entity recognizer.  "(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';P amazonka-comprehend0Detailed information about an entity recognizer.See:  smart constructor.amazonka-comprehend7Entity types from the metadata of an entity recognizer.amazonka-comprehendDetailed information about the accuracy of an entity recognizer.amazonka-comprehendThe number of documents in the input data that were used to test the entity recognizer. Typically this is 10 to 20 percent of the input documents.amazonka-comprehendThe number of documents in the input data that were used to train the entity recognizer. Typically this is 80 to 90 percent of the input documents.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, : - Entity types from the metadata of an entity recognizer.,  - Detailed information about the accuracy of an entity recognizer.,  - The number of documents in the input data that were used to test the entity recognizer. Typically this is 10 to 20 percent of the input documents.,  - The number of documents in the input data that were used to train the entity recognizer. Typically this is 80 to 90 percent of the input documents.amazonka-comprehend7Entity types from the metadata of an entity recognizer.amazonka-comprehendDetailed information about the accuracy of an entity recognizer.amazonka-comprehendThe number of documents in the input data that were used to test the entity recognizer. Typically this is 10 to 20 percent of the input documents.amazonka-comprehendThe number of documents in the input data that were used to train the entity recognizer. Typically this is 80 to 90 percent of the input documents.  #(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';X1amazonka-comprehendAn entity type within a labeled training dataset that Amazon Comprehend uses to train a custom entity recognizer.See:  smart constructor.amazonka-comprehendAn entity type within a labeled training dataset that Amazon Comprehend uses to train a custom entity recognizer.Entity types must not contain the following invalid characters: \n (line break), \\n (escaped line break, \r (carriage return), \\r (escaped carriage return), \t (tab), \\t (escaped tab), space, and , (comma).amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - An entity type within a labeled training dataset that Amazon Comprehend uses to train a custom entity recognizer.Entity types must not contain the following invalid characters: \n (line break), \\n (escaped line break, \r (carriage return), \\r (escaped carriage return), \t (tab), \\t (escaped tab), space, and , (comma).amazonka-comprehendAn entity type within a labeled training dataset that Amazon Comprehend uses to train a custom entity recognizer.Entity types must not contain the following invalid characters: \n (line break), \\n (escaped line break, \r (carriage return), \\r (escaped carriage return), \t (tab), \\t (escaped tab), space, and , (comma).amazonka-comprehend$(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';\!amazonka-comprehend;Array of the number of characters extracted from each page.See:  smart constructor.amazonka-comprehend.Number of characters extracted from each page.amazonka-comprehend Page number.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 1 - Number of characters extracted from each page.,  - Page number.amazonka-comprehend.Number of characters extracted from each page.amazonka-comprehend Page number.%(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';`amazonka-comprehendInformation about the document, discovered during text extraction.See:  smart constructor.amazonka-comprehendList of pages in the document, with the number of characters extracted from each page.amazonka-comprehend Number of pages in the document.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - List of pages in the document, with the number of characters extracted from each page., # - Number of pages in the document.amazonka-comprehendList of pages in the document, with the number of characters extracted from each page.amazonka-comprehend Number of pages in the document.&(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?a'(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';qKamazonka-comprehendThe input properties for an inference job. The document reader config field applies only to non-text inputs for custom analysis.See:  smart constructor.amazonka-comprehendProvides configuration parameters to override the default actions for extracting text from PDF documents and image files.amazonka-comprehendSee:  smart constructor.amazonka-comprehend1The name that you assigned the entity recognizer.amazonka-comprehend#The status of an entity recognizer.amazonka-comprehendFilters the list of entities based on the time that the list was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in ascending order, oldest to newest.amazonka-comprehendFilters the list of entities based on the time that the list was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in descending order, newest to oldest.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 4 - The name that you assigned the entity recognizer., & - The status of an entity recognizer.,  - Filters the list of entities based on the time that the list was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in ascending order, oldest to newest.,  - Filters the list of entities based on the time that the list was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in descending order, newest to oldest.amazonka-comprehend1The name that you assigned the entity recognizer.amazonka-comprehend#The status of an entity recognizer.amazonka-comprehendFilters the list of entities based on the time that the list was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in ascending order, oldest to newest.amazonka-comprehendFilters the list of entities based on the time that the list was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in descending order, newest to oldest.  5(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';V amazonka-comprehendDescribes information about a document classifier and its versions.See:  smart constructor.amazonka-comprehend3The name that you assigned the document classifier.amazonka-comprehendThe time that the latest document classifier version was submitted for processing.amazonka-comprehendThe version name you assigned to the latest document classifier version.amazonka-comprehend>Provides the status of the latest document classifier version.amazonka-comprehend#The number of versions you created.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 6 - The name that you assigned the document classifier.,  - The time that the latest document classifier version was submitted for processing.,  - The version name you assigned to the latest document classifier version.,  - Provides the status of the latest document classifier version., & - The number of versions you created.amazonka-comprehend3The name that you assigned the document classifier.amazonka-comprehendThe time that the latest document classifier version was submitted for processing.amazonka-comprehendThe version name you assigned to the latest document classifier version.amazonka-comprehend>Provides the status of the latest document classifier version.amazonka-comprehend#The number of versions you created.  6(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';E amazonka-comprehendProvides information for filtering a list of document classifiers. You can only specify one filtering parameter in a request. For more information, see the operation.See:  smart constructor.amazonka-comprehend5The name that you assigned to the document classifieramazonka-comprehend0Filters the list of classifiers based on status.amazonka-comprehendFilters the list of classifiers based on the time that the classifier was submitted for processing. Returns only classifiers submitted after the specified time. Classifiers are returned in descending order, newest to oldest.amazonka-comprehendFilters the list of classifiers based on the time that the classifier was submitted for processing. Returns only classifiers submitted before the specified time. Classifiers are returned in ascending order, oldest to newest.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 8 - The name that you assigned to the document classifier, 3 - Filters the list of classifiers based on status.,  - Filters the list of classifiers based on the time that the classifier was submitted for processing. Returns only classifiers submitted after the specified time. Classifiers are returned in descending order, newest to oldest.,  - Filters the list of classifiers based on the time that the classifier was submitted for processing. Returns only classifiers submitted before the specified time. Classifiers are returned in ascending order, oldest to newest.amazonka-comprehend5The name that you assigned to the document classifieramazonka-comprehend0Filters the list of classifiers based on status.amazonka-comprehendFilters the list of classifiers based on the time that the classifier was submitted for processing. Returns only classifiers submitted after the specified time. Classifiers are returned in descending order, newest to oldest.amazonka-comprehendFilters the list of classifiers based on the time that the classifier was submitted for processing. Returns only classifiers submitted before the specified time. Classifiers are returned in ascending order, oldest to newest.  7(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-comprehendProvides configuration parameters for the output of inference jobs.See:  smart constructor.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"KMS Key Alias: "alias/ExampleAlias"ARN of a KMS Key Alias: 7"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias"amazonka-comprehendWhen you use the OutputDataConfig object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of the output file.When the topic detection job is finished, the service creates an output file in a directory specific to the job. The S3Uri9 field contains the location of the output file, called  output.tar.gz. It is a compressed archive that contains the ouput of the operation.For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out appended at the end.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"KMS Key Alias: "alias/ExampleAlias"ARN of a KMS Key Alias: 7"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias",  - When you use the OutputDataConfig object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of the output file.When the topic detection job is finished, the service creates an output file in a directory specific to the job. The S3Uri9 field contains the location of the output file, called  output.tar.gz. It is a compressed archive that contains the ouput of the operation.For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out appended at the end.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. The KmsKeyId can be one of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"KMS Key Alias: "alias/ExampleAlias"ARN of a KMS Key Alias: 7"arn:aws:kms:us-west-2:111122223333:alias/ExampleAlias"amazonka-comprehendWhen you use the OutputDataConfig object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data. The URI must be in the same region as the API endpoint that you are calling. The location is used as the prefix for the actual location of the output file.When the topic detection job is finished, the service creates an output file in a directory specific to the job. The S3Uri9 field contains the location of the output file, called  output.tar.gz. It is a compressed archive that contains the ouput of the operation.For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out appended at the end.amazonka-comprehend8(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehend3Provides information about an events detection job.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identify and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehend1The time that the events detection job completed.amazonka-comprehendThe input data configuration that you supplied when you created the events detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::events-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend4The identifier assigned to the events detection job.amazonka-comprehend/The name you assigned the events detection job.amazonka-comprehend/The current status of the events detection job.amazonka-comprehend)The language code of the input documents.amazonka-comprehend8A description of the status of the events detection job.amazonka-comprehendThe output data configuration that you supplied when you created the events detection job.amazonka-comprehendThe time that the events detection job was submitted for processing.amazonka-comprehend1The types of events that are detected by the job.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the AWS Identify and Access Management (IAM) role that grants Amazon Comprehend read access to your input data., 4 - The time that the events detection job completed.,  - The input data configuration that you supplied when you created the events detection job.,  - The Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::events-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890ab, 7 - The identifier assigned to the events detection job., 2 - The name you assigned the events detection job., 2 - The current status of the events detection job., , - The language code of the input documents., ; - A description of the status of the events detection job.,  - The output data configuration that you supplied when you created the events detection job.,  - The time that the events detection job was submitted for processing., 4 - The types of events that are detected by the job.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identify and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehend1The time that the events detection job completed.amazonka-comprehendThe input data configuration that you supplied when you created the events detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::events-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend4The identifier assigned to the events detection job.amazonka-comprehend/The name you assigned the events detection job.amazonka-comprehend/The current status of the events detection job.amazonka-comprehend)The language code of the input documents.amazonka-comprehend8A description of the status of the events detection job.amazonka-comprehendThe output data configuration that you supplied when you created the events detection job.amazonka-comprehendThe time that the events detection job was submitted for processing.amazonka-comprehend1The types of events that are detected by the job.9(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?  :(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';samazonka-comprehendText extraction encountered one or more page-level errors in the input document.The  ErrorCode& contains one of the following values:TEXTRACT_BAD_PAGE - Amazon Textract cannot read the page. For more information about page limits in Amazon Textract, see  https://docs.aws.amazon.com/textract/latest/dg/limits-document.htmlPage Quotas in Amazon Textract.TEXTRACT_PROVISIONED_THROUGHPUT_EXCEEDED - The number of requests exceeded your throughput limit. For more information about throughput quotas in Amazon Textract, see  https://docs.aws.amazon.com/textract/latest/dg/limits-quotas-explained.html!Default quotas in Amazon Textract.PAGE_CHARACTERS_EXCEEDED - Too many text characters on the page (10,000 characters maximum).4PAGE_SIZE_EXCEEDED - The maximum page size is 10 MB.INTERNAL_SERVER_ERROR - The request encountered a service issue. Try the API request again.See:  smart constructor.amazonka-comprehend&Error code for the cause of the error.amazonka-comprehend1Text message explaining the reason for the error.amazonka-comprehend%Page number where the error occurred.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, ) - Error code for the cause of the error., 4 - Text message explaining the reason for the error., ( - Page number where the error occurred.amazonka-comprehend&Error code for the cause of the error.amazonka-comprehend1Text message explaining the reason for the error.amazonka-comprehend%Page number where the error occurred.  ;(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?; ' <(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehendIdentifies the part of speech represented by the token and gives the confidence that Amazon Comprehend has that the part of speech was correctly identified. For more information about the parts of speech that Amazon Comprehend can identify, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-syntax.htmlSyntax$ in the Comprehend Developer Guide.See:   smart constructor. amazonka-comprehendThe confidence that Amazon Comprehend has that the part of speech was correctly identified. amazonka-comprehend8Identifies the part of speech that the token represents. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The confidence that Amazon Comprehend has that the part of speech was correctly identified. ,  ; - Identifies the part of speech that the token represents. amazonka-comprehendThe confidence that Amazon Comprehend has that the part of speech was correctly identified. amazonka-comprehend8Identifies the part of speech that the token represents.  =(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';' amazonka-comprehendProvides information for filtering a list of PII entity detection jobs.See:   smart constructor. amazonka-comprehendFilters on the name of the job. amazonka-comprehendFilters the list of jobs based on job status. Returns only jobs with the specified status. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  " - Filters on the name of the job. ,   - Filters the list of jobs based on job status. Returns only jobs with the specified status. ,   - Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. ,   - Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. amazonka-comprehendFilters on the name of the job. amazonka-comprehendFilters the list of jobs based on job status. Returns only jobs with the specified status. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. >(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?(  ?(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?)i  @(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?*%( A(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';2J amazonka-comprehend(Provides information about a PII entity.See:   smart constructor. amazonka-comprehendThe zero-based offset from the beginning of the source text to the first character in the entity. amazonka-comprehendThe zero-based offset from the beginning of the source text to the last character in the entity. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of the detection. amazonka-comprehendThe entity's type. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The zero-based offset from the beginning of the source text to the first character in the entity. ,   - The zero-based offset from the beginning of the source text to the last character in the entity. ,   - The level of confidence that Amazon Comprehend has in the accuracy of the detection. ,   - The entity's type. amazonka-comprehendThe zero-based offset from the beginning of the source text to the first character in the entity. amazonka-comprehendThe zero-based offset from the beginning of the source text to the last character in the entity. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of the detection. amazonka-comprehendThe entity's type. B(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';7( amazonka-comprehendSpecifies one of the label or labels that categorize the personally identifiable information (PII) entity being analyzed.See:   smart constructor. amazonka-comprehendThe name of the label. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of the detection. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The name of the label. ,   - The level of confidence that Amazon Comprehend has in the accuracy of the detection. amazonka-comprehendThe name of the label. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of the detection.  C(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';@1 amazonka-comprehendProvides configuration parameters for the output of PII entity detection jobs.See:   smart constructor. amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. amazonka-comprehendWhen you use the PiiOutputDataConfig object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data.For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out appended at the end. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. ,   - When you use the PiiOutputDataConfig object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data.For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out appended at the end. amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt the output results from an analysis job. amazonka-comprehendWhen you use the PiiOutputDataConfig object with asynchronous operations, you specify the Amazon S3 location where you want to write the output data.For a PII entity detection job, the output file is plain text, not a compressed archive. The output file name is the same as the input file, with .out appended at the end. amazonka-comprehend   D(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';EQ amazonka-comprehend6The X and Y coordinates of a point on a document page.!For additional information, see  =https://docs.aws.amazon.com/textract/latest/dg/API_Point.htmlPoint' in the Amazon Textract API reference.See:   smart constructor. amazonka-comprehend6The value of the X coordinate for a point on a polygon amazonka-comprehend6The value of the Y coordinate for a point on a polygon amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  9 - The value of the X coordinate for a point on a polygon ,  9 - The value of the Y coordinate for a point on a polygon amazonka-comprehend6The value of the X coordinate for a point on a polygon amazonka-comprehend6The value of the Y coordinate for a point on a polygon  E(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';KL amazonka-comprehend;Information about the location of items on a document page.!For additional information, see  https://docs.aws.amazon.com/textract/latest/dg/API_Geometry.htmlGeometry' in the Amazon Textract API reference.See:   smart constructor. amazonka-comprehendAn axis-aligned coarse representation of the location of the recognized item on the document page. amazonka-comprehendWithin the bounding box, a fine-grained polygon around the recognized item. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - An axis-aligned coarse representation of the location of the recognized item on the document page. ,   - Within the bounding box, a fine-grained polygon around the recognized item. amazonka-comprehendAn axis-aligned coarse representation of the location of the recognized item on the document page. amazonka-comprehendWithin the bounding box, a fine-grained polygon around the recognized item.  F(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';Q amazonka-comprehend;Provides configuration parameters for PII entity redaction.See:   smart constructor. amazonka-comprehendA character that replaces each character in the redacted PII entity. amazonka-comprehendSpecifies whether the PII entity is redacted with the mask character or the entity type. amazonka-comprehendAn array of the types of PII entities that Amazon Comprehend detects in the input text for your request. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - A character that replaces each character in the redacted PII entity. ,   - Specifies whether the PII entity is redacted with the mask character or the entity type. ,   - An array of the types of PII entities that Amazon Comprehend detects in the input text for your request. amazonka-comprehendA character that replaces each character in the redacted PII entity. amazonka-comprehendSpecifies whether the PII entity is redacted with the mask character or the entity type. amazonka-comprehendAn array of the types of PII entities that Amazon Comprehend detects in the input text for your request. G(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';j amazonka-comprehend8Provides information about a PII entities detection job.See:   smart constructor. amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data. amazonka-comprehend7The time that the PII entities detection job completed. amazonka-comprehend6The input properties for a PII entities detection job. amazonka-comprehendThe Amazon Resource Name (ARN) of the PII entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::pii-entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890ab amazonka-comprehend:The identifier assigned to the PII entities detection job. amazonka-comprehend:The name that you assigned the PII entities detection job. amazonka-comprehendThe current status of the PII entities detection job. If the status is FAILED, the Message( field shows the reason for the failure. amazonka-comprehend(The language code of the input documents amazonka-comprehend%A description of the status of a job. amazonka-comprehendSpecifies whether the output provides the locations (offsets) of PII entities or a file in which PII entities are redacted. amazonka-comprehendThe output data configuration that you supplied when you created the PII entities detection job. amazonka-comprehend;Provides configuration parameters for PII entity redaction.*This parameter is required if you set the Mode parameter to ONLY_REDACTION#. In that case, you must provide a RedactionConfig definition that includes the PiiEntityTypes parameter. amazonka-comprehendThe time that the PII entities detection job was submitted for processing. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data. ,  : - The time that the PII entities detection job completed. ,  9 - The input properties for a PII entities detection job. ,   - The Amazon Resource Name (ARN) of the PII entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::pii-entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890ab ,  = - The identifier assigned to the PII entities detection job. ,  = - The name that you assigned the PII entities detection job. ,   - The current status of the PII entities detection job. If the status is FAILED, the Message( field shows the reason for the failure. ,  + - The language code of the input documents ,  ( - A description of the status of a job. ,   - Specifies whether the output provides the locations (offsets) of PII entities or a file in which PII entities are redacted. ,   - The output data configuration that you supplied when you created the PII entities detection job. ,  > - Provides configuration parameters for PII entity redaction.*This parameter is required if you set the Mode parameter to ONLY_REDACTION#. In that case, you must provide a RedactionConfig definition that includes the PiiEntityTypes parameter. ,   - The time that the PII entities detection job was submitted for processing. amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data. amazonka-comprehend7The time that the PII entities detection job completed. amazonka-comprehend6The input properties for a PII entities detection job. amazonka-comprehendThe Amazon Resource Name (ARN) of the PII entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::pii-entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890ab amazonka-comprehend:The identifier assigned to the PII entities detection job. amazonka-comprehend:The name that you assigned the PII entities detection job. amazonka-comprehendThe current status of the PII entities detection job. If the status is FAILED, the Message( field shows the reason for the failure. amazonka-comprehend(The language code of the input documents amazonka-comprehend%A description of the status of a job. amazonka-comprehendSpecifies whether the output provides the locations (offsets) of PII entities or a file in which PII entities are redacted. amazonka-comprehendThe output data configuration that you supplied when you created the PII entities detection job. amazonka-comprehend;Provides configuration parameters for PII entity redaction.*This parameter is required if you set the Mode parameter to ONLY_REDACTION#. In that case, you must provide a RedactionConfig definition that includes the PiiEntityTypes parameter. amazonka-comprehendThe time that the PII entities detection job was submitted for processing.  H(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?k  I(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';p amazonka-comprehend+List of child blocks for the current block.See:   smart constructor. amazonka-comprehendIdentifers of the child blocks. amazonka-comprehend4Only supported relationship is a child relationship. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  " - Identifers of the child blocks. ,  7 - Only supported relationship is a child relationship. amazonka-comprehendIdentifers of the child blocks. amazonka-comprehend4Only supported relationship is a child relationship.  J(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';| amazonka-comprehendInformation about each word or line of text in the input document.!For additional information, see  =https://docs.aws.amazon.com/textract/latest/dg/API_Block.htmlBlock' in the Amazon Textract API reference.See:   smart constructor. amazonka-comprehend8The block represents a line of text or one word of text.WORD - A word that's detected on a document page. A word is one or more ISO basic Latin script characters that aren't separated by spaces.LINE - A string of tab-delimited, contiguous words that are detected on a document page amazonka-comprehendCo-ordinates of the rectangle or polygon that contains the text. amazonka-comprehend Unique identifier for the block. amazonka-comprehend$Page number where the block appears. amazonka-comprehendA list of child blocks of the current block. For example, a LINE object has child blocks for each WORD block that's part of the line of text. amazonka-comprehend2The word or line of text extracted from the block. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  ; - The block represents a line of text or one word of text.WORD - A word that's detected on a document page. A word is one or more ISO basic Latin script characters that aren't separated by spaces.LINE - A string of tab-delimited, contiguous words that are detected on a document page ,   - Co-ordinates of the rectangle or polygon that contains the text. ,  # - Unique identifier for the block. ,  ' - Page number where the block appears. ,   - A list of child blocks of the current block. For example, a LINE object has child blocks for each WORD block that's part of the line of text. ,  5 - The word or line of text extracted from the block. amazonka-comprehend8The block represents a line of text or one word of text.WORD - A word that's detected on a document page. A word is one or more ISO basic Latin script characters that aren't separated by spaces.LINE - A string of tab-delimited, contiguous words that are detected on a document page amazonka-comprehendCo-ordinates of the rectangle or polygon that contains the text. amazonka-comprehend Unique identifier for the block. amazonka-comprehend$Page number where the block appears. amazonka-comprehendA list of child blocks of the current block. For example, a LINE object has child blocks for each WORD block that's part of the line of text. amazonka-comprehend2The word or line of text extracted from the block.  K(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehendProvides information for filtering a list of dominant language detection jobs. For more information, see the operation.See:   smart constructor. amazonka-comprehendFilters on the name of the job. amazonka-comprehendFilters the list of jobs based on job status. Returns only jobs with the specified status. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  " - Filters on the name of the job. ,   - Filters the list of jobs based on job status. Returns only jobs with the specified status. ,   - Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. ,   - Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. amazonka-comprehendFilters on the name of the job. amazonka-comprehendFilters the list of jobs based on job status. Returns only jobs with the specified status. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. L(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehendDescribes the level of confidence that Amazon Comprehend has in the accuracy of its detection of sentiments.See:   smart constructor. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its detection of the MIXED sentiment. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its detection of the NEGATIVE sentiment. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its detection of the NEUTRAL sentiment. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its detection of the POSITIVE sentiment. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The level of confidence that Amazon Comprehend has in the accuracy of its detection of the MIXED sentiment. ,   - The level of confidence that Amazon Comprehend has in the accuracy of its detection of the NEGATIVE sentiment. ,   - The level of confidence that Amazon Comprehend has in the accuracy of its detection of the NEUTRAL sentiment. ,   - The level of confidence that Amazon Comprehend has in the accuracy of its detection of the POSITIVE sentiment. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its detection of the MIXED sentiment. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its detection of the NEGATIVE sentiment. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its detection of the NEUTRAL sentiment. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its detection of the POSITIVE sentiment. M(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?x N(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehendContains the sentiment and sentiment score for one mention of an entity.4For more information about targeted sentiment, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.htmlTargeted sentiment.See:   smart constructor. amazonka-comprehendThe sentiment of the mention. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The sentiment of the mention. ,   - Undocumented member. amazonka-comprehendThe sentiment of the mention. amazonka-comprehendUndocumented member.  O(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehendThe result of calling the operation. The operation returns one object for each document that is successfully processed by the operation.See:   smart constructor. amazonka-comprehend7The zero-based index of the document in the input list. amazonka-comprehend'The sentiment detected in the document. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its sentiment detection. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  : - The zero-based index of the document in the input list. ,  * - The sentiment detected in the document. ,   - The level of confidence that Amazon Comprehend has in the accuracy of its sentiment detection. amazonka-comprehend7The zero-based index of the document in the input list. amazonka-comprehend'The sentiment detected in the document. amazonka-comprehendThe level of confidence that Amazon Comprehend has in the accuracy of its sentiment detection. P(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?  Q(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';_ amazonka-comprehendAn augmented manifest file that provides training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.See:   smart constructor. amazonka-comprehendThe S3 prefix to the annotation files that are referred in the augmented manifest file. amazonka-comprehendThe type of augmented manifest. PlainTextDocument or SemiStructuredDocument. If you don't specify, the default is PlainTextDocument.PLAIN_TEXT_DOCUMENT A document type that represents any unicode text that is encoded in UTF-8.SEMI_STRUCTURED_DOCUMENT A document type with positional and structural context, like a PDF. For training with Amazon Comprehend, only PDFs are supported. For inference, Amazon Comprehend support PDFs, DOCX and TXT. amazonka-comprehendThe S3 prefix to the source files (PDFs) that are referred to in the augmented manifest file. amazonka-comprehendThe purpose of the data you've provided in the augmented manifest. You can either train or test this data. If you don't specify, the default is train.TRAIN - all of the documents in the manifest will be used for training. If no test documents are provided, Amazon Comprehend will automatically reserve a portion of the training documents for testing.TEST - all of the documents in the manifest will be used for testing. amazonka-comprehend6The Amazon S3 location of the augmented manifest file. amazonka-comprehendThe JSON attribute that contains the annotations for your training documents. The number of attribute names that you specify depends on whether your augmented manifest file is the output of a single labeling job or a chained labeling job.If your file is the output of a single labeling job, specify the LabelAttributeName key that was used when the job was created in Ground Truth.If your file is the output of a chained labeling job, specify the LabelAttributeName key for one or more jobs in the chain. Each LabelAttributeName key provides the annotations from an individual job. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The S3 prefix to the annotation files that are referred in the augmented manifest file. ,   - The type of augmented manifest. PlainTextDocument or SemiStructuredDocument. If you don't specify, the default is PlainTextDocument.PLAIN_TEXT_DOCUMENT A document type that represents any unicode text that is encoded in UTF-8.SEMI_STRUCTURED_DOCUMENT A document type with positional and structural context, like a PDF. For training with Amazon Comprehend, only PDFs are supported. For inference, Amazon Comprehend support PDFs, DOCX and TXT. ,   - The S3 prefix to the source files (PDFs) that are referred to in the augmented manifest file. ,   - The purpose of the data you've provided in the augmented manifest. You can either train or test this data. If you don't specify, the default is train.TRAIN - all of the documents in the manifest will be used for training. If no test documents are provided, Amazon Comprehend will automatically reserve a portion of the training documents for testing.TEST - all of the documents in the manifest will be used for testing. ,  9 - The Amazon S3 location of the augmented manifest file. ,   - The JSON attribute that contains the annotations for your training documents. The number of attribute names that you specify depends on whether your augmented manifest file is the output of a single labeling job or a chained labeling job.If your file is the output of a single labeling job, specify the LabelAttributeName key that was used when the job was created in Ground Truth.If your file is the output of a chained labeling job, specify the LabelAttributeName key for one or more jobs in the chain. Each LabelAttributeName key provides the annotations from an individual job. amazonka-comprehendThe S3 prefix to the annotation files that are referred in the augmented manifest file. amazonka-comprehendThe type of augmented manifest. PlainTextDocument or SemiStructuredDocument. If you don't specify, the default is PlainTextDocument.PLAIN_TEXT_DOCUMENT A document type that represents any unicode text that is encoded in UTF-8.SEMI_STRUCTURED_DOCUMENT A document type with positional and structural context, like a PDF. For training with Amazon Comprehend, only PDFs are supported. For inference, Amazon Comprehend support PDFs, DOCX and TXT. amazonka-comprehendThe S3 prefix to the source files (PDFs) that are referred to in the augmented manifest file. amazonka-comprehendThe purpose of the data you've provided in the augmented manifest. You can either train or test this data. If you don't specify, the default is train.TRAIN - all of the documents in the manifest will be used for training. If no test documents are provided, Amazon Comprehend will automatically reserve a portion of the training documents for testing.TEST - all of the documents in the manifest will be used for testing. amazonka-comprehend6The Amazon S3 location of the augmented manifest file. amazonka-comprehendThe JSON attribute that contains the annotations for your training documents. The number of attribute names that you specify depends on whether your augmented manifest file is the output of a single labeling job or a chained labeling job.If your file is the output of a single labeling job, specify the LabelAttributeName key that was used when the job was created in Ground Truth.If your file is the output of a chained labeling job, specify the LabelAttributeName key for one or more jobs in the chain. Each LabelAttributeName key provides the annotations from an individual job. amazonka-comprehend   R(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehend4Specifies the format and location of the input data.See:   smart constructor. amazonka-comprehendThe S3 location of the CSV file that annotates your training documents. amazonka-comprehendA list of augmented manifest files that provide training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.&This parameter is required if you set  DataFormat to AUGMENTED_MANIFEST. amazonka-comprehend!The format of your training data:COMPREHEND_CSV: A CSV file that supplements your training documents. The CSV file contains information about the custom entities that your trained model will detect. The required format of the file depends on whether you are providing annotations or an entity list.If you use this value, you must provide your CSV file by using either the  Annotations or  EntityList parameters. You must provide your training documents by using the  Documents parameter.AUGMENTED_MANIFEST: A labeled dataset that is produced by Amazon SageMaker Ground Truth. This file is in JSON lines format. Each line is a complete JSON object that contains a training document and its labels. Each label annotates a named entity in the training document.,If you use this value, you must provide the AugmentedManifests parameter in your request.5If you don't specify a value, Amazon Comprehend uses COMPREHEND_CSV as the default. amazonka-comprehendThe S3 location of the folder that contains the training documents for your custom entity recognizer.&This parameter is required if you set  DataFormat to COMPREHEND_CSV. amazonka-comprehendThe S3 location of the CSV file that has the entity list for your custom entity recognizer. amazonka-comprehendThe entity types in the labeled training data that Amazon Comprehend uses to train the custom entity recognizer. Any entity types that you don't specify are ignored.A maximum of 25 entity types can be used at one time to train an entity recognizer. Entity types must not contain the following invalid characters: \n (line break), \\n (escaped line break), \r (carriage return), \\r (escaped carriage return), \t (tab), \\t (escaped tab), space, and , (comma). amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The S3 location of the CSV file that annotates your training documents. ,   - A list of augmented manifest files that provide training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.&This parameter is required if you set  DataFormat to AUGMENTED_MANIFEST. ,  $ - The format of your training data:COMPREHEND_CSV: A CSV file that supplements your training documents. The CSV file contains information about the custom entities that your trained model will detect. The required format of the file depends on whether you are providing annotations or an entity list.If you use this value, you must provide your CSV file by using either the  Annotations or  EntityList parameters. You must provide your training documents by using the  Documents parameter.AUGMENTED_MANIFEST: A labeled dataset that is produced by Amazon SageMaker Ground Truth. This file is in JSON lines format. Each line is a complete JSON object that contains a training document and its labels. Each label annotates a named entity in the training document.,If you use this value, you must provide the AugmentedManifests parameter in your request.5If you don't specify a value, Amazon Comprehend uses COMPREHEND_CSV as the default. ,   - The S3 location of the folder that contains the training documents for your custom entity recognizer.&This parameter is required if you set  DataFormat to COMPREHEND_CSV. ,   - The S3 location of the CSV file that has the entity list for your custom entity recognizer. ,   - The entity types in the labeled training data that Amazon Comprehend uses to train the custom entity recognizer. Any entity types that you don't specify are ignored.A maximum of 25 entity types can be used at one time to train an entity recognizer. Entity types must not contain the following invalid characters: \n (line break), \\n (escaped line break), \r (carriage return), \\r (escaped carriage return), \t (tab), \\t (escaped tab), space, and , (comma). amazonka-comprehendThe S3 location of the CSV file that annotates your training documents. amazonka-comprehendA list of augmented manifest files that provide training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.&This parameter is required if you set  DataFormat to AUGMENTED_MANIFEST. amazonka-comprehend!The format of your training data:COMPREHEND_CSV: A CSV file that supplements your training documents. The CSV file contains information about the custom entities that your trained model will detect. The required format of the file depends on whether you are providing annotations or an entity list.If you use this value, you must provide your CSV file by using either the  Annotations or  EntityList parameters. You must provide your training documents by using the  Documents parameter.AUGMENTED_MANIFEST: A labeled dataset that is produced by Amazon SageMaker Ground Truth. This file is in JSON lines format. Each line is a complete JSON object that contains a training document and its labels. Each label annotates a named entity in the training document.,If you use this value, you must provide the AugmentedManifests parameter in your request.5If you don't specify a value, Amazon Comprehend uses COMPREHEND_CSV as the default. amazonka-comprehendThe S3 location of the folder that contains the training documents for your custom entity recognizer.&This parameter is required if you set  DataFormat to COMPREHEND_CSV. amazonka-comprehendThe S3 location of the CSV file that has the entity list for your custom entity recognizer. amazonka-comprehendThe entity types in the labeled training data that Amazon Comprehend uses to train the custom entity recognizer. Any entity types that you don't specify are ignored.A maximum of 25 entity types can be used at one time to train an entity recognizer. Entity types must not contain the following invalid characters: \n (line break), \\n (escaped line break), \r (carriage return), \\r (escaped carriage return), \t (tab), \\t (escaped tab), space, and , (comma).  S(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehend8The input properties for training a document classifier.>For more information on how the input file is formatted, see  https://docs.aws.amazon.com/comprehend/latest/dg/prep-classifier-data.htmlPreparing training data$ in the Comprehend Developer Guide.See:   smart constructor. amazonka-comprehendA list of augmented manifest files that provide training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.&This parameter is required if you set  DataFormat to AUGMENTED_MANIFEST. amazonka-comprehend!The format of your training data:COMPREHEND_CSV: A two-column CSV file, where labels are provided in the first column, and documents are provided in the second. If you use this value, you must provide the S3Uri parameter in your request.AUGMENTED_MANIFEST: A labeled dataset that is produced by Amazon SageMaker Ground Truth. This file is in JSON lines format. Each line is a complete JSON object that contains a training document and its associated labels.,If you use this value, you must provide the AugmentedManifests parameter in your request.5If you don't specify a value, Amazon Comprehend uses COMPREHEND_CSV as the default. amazonka-comprehendIndicates the delimiter used to separate each label for training a multi-label classifier. The default delimiter between labels is a pipe (|). You can use a different character as a delimiter (if it's an allowed character) by specifying it under Delimiter for labels. If the training documents use a delimiter other than the default or the delimiter you specify, the labels on that line will be combined to make a single unique label, such as LABELLABELLABEL. amazonka-comprehendThe Amazon S3 URI for the input data. The S3 bucket must be in the same region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of input files. For example, if you use the URI S3://bucketName/prefix, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.&This parameter is required if you set  DataFormat to COMPREHEND_CSV. amazonka-comprehendThis specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same AWS Region as the API endpoint that you are calling. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - A list of augmented manifest files that provide training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.&This parameter is required if you set  DataFormat to AUGMENTED_MANIFEST. ,  $ - The format of your training data:COMPREHEND_CSV: A two-column CSV file, where labels are provided in the first column, and documents are provided in the second. If you use this value, you must provide the S3Uri parameter in your request.AUGMENTED_MANIFEST: A labeled dataset that is produced by Amazon SageMaker Ground Truth. This file is in JSON lines format. Each line is a complete JSON object that contains a training document and its associated labels.,If you use this value, you must provide the AugmentedManifests parameter in your request.5If you don't specify a value, Amazon Comprehend uses COMPREHEND_CSV as the default. ,   - Indicates the delimiter used to separate each label for training a multi-label classifier. The default delimiter between labels is a pipe (|). You can use a different character as a delimiter (if it's an allowed character) by specifying it under Delimiter for labels. If the training documents use a delimiter other than the default or the delimiter you specify, the labels on that line will be combined to make a single unique label, such as LABELLABELLABEL. ,   - The Amazon S3 URI for the input data. The S3 bucket must be in the same region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of input files. For example, if you use the URI S3://bucketName/prefix, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.&This parameter is required if you set  DataFormat to COMPREHEND_CSV. ,   - This specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same AWS Region as the API endpoint that you are calling. amazonka-comprehendA list of augmented manifest files that provide training data for your custom model. An augmented manifest file is a labeled dataset that is produced by Amazon SageMaker Ground Truth.&This parameter is required if you set  DataFormat to AUGMENTED_MANIFEST. amazonka-comprehend!The format of your training data:COMPREHEND_CSV: A two-column CSV file, where labels are provided in the first column, and documents are provided in the second. If you use this value, you must provide the S3Uri parameter in your request.AUGMENTED_MANIFEST: A labeled dataset that is produced by Amazon SageMaker Ground Truth. This file is in JSON lines format. Each line is a complete JSON object that contains a training document and its associated labels.,If you use this value, you must provide the AugmentedManifests parameter in your request.5If you don't specify a value, Amazon Comprehend uses COMPREHEND_CSV as the default. amazonka-comprehendIndicates the delimiter used to separate each label for training a multi-label classifier. The default delimiter between labels is a pipe (|). You can use a different character as a delimiter (if it's an allowed character) by specifying it under Delimiter for labels. If the training documents use a delimiter other than the default or the delimiter you specify, the labels on that line will be combined to make a single unique label, such as LABELLABELLABEL. amazonka-comprehendThe Amazon S3 URI for the input data. The S3 bucket must be in the same region as the API endpoint that you are calling. The URI can point to a single input file or it can provide the prefix for a collection of input files. For example, if you use the URI S3://bucketName/prefix, if the prefix is a single file, Amazon Comprehend uses that file as input. If more than one file begins with the prefix, Amazon Comprehend uses all of them as input.&This parameter is required if you set  DataFormat to COMPREHEND_CSV. amazonka-comprehendThis specifies the Amazon S3 location where the test annotations for an entity recognizer are located. The URI must be in the same AWS Region as the API endpoint that you are calling. T(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?  U(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehendRepresents a work in the input text that was recognized and assigned a part of speech. There is one syntax token record for each word in the source text.See:   smart constructor. amazonka-comprehendThe zero-based offset from the beginning of the source text to the first character in the word. amazonka-comprehendThe zero-based offset from the beginning of the source text to the last character in the word. amazonka-comprehendProvides the part of speech label and the confidence level that Amazon Comprehend has that the part of speech was correctly identified. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-syntax.htmlSyntax$ in the Comprehend Developer Guide. amazonka-comprehend0The word that was recognized in the source text. amazonka-comprehend A unique identifier for a token. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The zero-based offset from the beginning of the source text to the first character in the word. ,   - The zero-based offset from the beginning of the source text to the last character in the word. ,   - Provides the part of speech label and the confidence level that Amazon Comprehend has that the part of speech was correctly identified. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-syntax.htmlSyntax$ in the Comprehend Developer Guide. ,  3 - The word that was recognized in the source text. ,  # - A unique identifier for a token. amazonka-comprehendThe zero-based offset from the beginning of the source text to the first character in the word. amazonka-comprehendThe zero-based offset from the beginning of the source text to the last character in the word. amazonka-comprehendProvides the part of speech label and the confidence level that Amazon Comprehend has that the part of speech was correctly identified. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-syntax.htmlSyntax$ in the Comprehend Developer Guide. amazonka-comprehend0The word that was recognized in the source text. amazonka-comprehend A unique identifier for a token. V(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehendThe result of calling the operation. The operation returns one object that is successfully processed by the operation.See:   smart constructor. amazonka-comprehend7The zero-based index of the document in the input list. amazonka-comprehendThe syntax tokens for the words in the document, one token for each word. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  : - The zero-based index of the document in the input list. ,   - The syntax tokens for the words in the document, one token for each word. amazonka-comprehend7The zero-based index of the document in the input list. amazonka-comprehendThe syntax tokens for the words in the document, one token for each word.  W(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';* amazonka-comprehendA key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with the key-value pair @Department@:@Sales@ might be added to a resource to indicate its use by a particular department.See:   smart constructor. amazonka-comprehendThe second part of a key-value pair that forms a tag associated with a given resource. For instance, if you want to show which resources are used by which departments, you might use @Department@ as the initial (key) portion of the pair, with a value of @sales@ to indicate the sales department. amazonka-comprehendThe initial part of a key-value pair that forms a tag associated with a given resource. For instance, if you want to show which resources are used by which departments, you might use @Department@ as the key portion of the pair, with multiple possible values such as @sales,@ @legal,@ and @administration.@ amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The second part of a key-value pair that forms a tag associated with a given resource. For instance, if you want to show which resources are used by which departments, you might use @Department@ as the initial (key) portion of the pair, with a value of @sales@ to indicate the sales department. ,   - The initial part of a key-value pair that forms a tag associated with a given resource. For instance, if you want to show which resources are used by which departments, you might use @Department@ as the key portion of the pair, with multiple possible values such as @sales,@ @legal,@ and @administration.@ amazonka-comprehendThe second part of a key-value pair that forms a tag associated with a given resource. For instance, if you want to show which resources are used by which departments, you might use @Department@ as the initial (key) portion of the pair, with a value of @sales@ to indicate the sales department. amazonka-comprehendThe initial part of a key-value pair that forms a tag associated with a given resource. For instance, if you want to show which resources are used by which departments, you might use @Department@ as the key portion of the pair, with multiple possible values such as @sales,@ @legal,@ and @administration.@ amazonka-comprehend   X(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';e amazonka-comprehendProvides information for filtering a list of dominant language detection jobs. For more information, see the operation.See:   smart constructor. amazonka-comprehendFilters on the name of the job. amazonka-comprehendFilters the list of jobs based on job status. Returns only jobs with the specified status. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  " - Filters on the name of the job. ,   - Filters the list of jobs based on job status. Returns only jobs with the specified status. ,   - Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. ,   - Filters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. amazonka-comprehendFilters on the name of the job. amazonka-comprehendFilters the list of jobs based on job status. Returns only jobs with the specified status. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted after the specified time. Jobs are returned in descending order, newest to oldest. amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Returns only jobs submitted before the specified time. Jobs are returned in ascending order, oldest to newest. Y(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred";?5%Z(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';'amazonka-comprehendInformation about one mention of an entity. The mention information includes the location of the mention in the text and the sentiment of the mention.4For more information about targeted sentiment, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.htmlTargeted sentiment.See:  smart constructor.amazonka-comprehend;The offset into the document text where the mention begins.amazonka-comprehend9The offset into the document text where the mention ends.amazonka-comprehendThe confidence that all the entities mentioned in the group relate to the same entity.amazonka-comprehend;Contains the sentiment and sentiment score for the mention.amazonka-comprehendModel confidence that the entity is relevant. Value range is zero to one, where one is highest confidence.amazonka-comprehend4The text in the document that identifies the entity.amazonka-comprehendThe type of the entity. Amazon Comprehend supports a variety of  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.html#how-targeted-sentiment-entities entity types.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, > - The offset into the document text where the mention begins., < - The offset into the document text where the mention ends.,  - The confidence that all the entities mentioned in the group relate to the same entity., > - Contains the sentiment and sentiment score for the mention.,  - Model confidence that the entity is relevant. Value range is zero to one, where one is highest confidence., 7 - The text in the document that identifies the entity.,  - The type of the entity. Amazon Comprehend supports a variety of  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.html#how-targeted-sentiment-entities entity types.amazonka-comprehend;The offset into the document text where the mention begins.amazonka-comprehend9The offset into the document text where the mention ends.amazonka-comprehendThe confidence that all the entities mentioned in the group relate to the same entity.amazonka-comprehend;Contains the sentiment and sentiment score for the mention.amazonka-comprehendModel confidence that the entity is relevant. Value range is zero to one, where one is highest confidence.amazonka-comprehend4The text in the document that identifies the entity.amazonka-comprehendThe type of the entity. Amazon Comprehend supports a variety of  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.html#how-targeted-sentiment-entities entity types.[(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';/amazonka-comprehendInformation about one of the entities found by targeted sentiment analysis.4For more information about targeted sentiment, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.htmlTargeted sentiment.See:  smart constructor.amazonka-comprehendOne or more index into the Mentions array that provides the best name for the entity group.amazonka-comprehendAn array of mentions of the entity in the document. The array represents a co-reference group. See  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.html#how-targeted-sentiment-valuesCo-reference group for an example.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - One or more index into the Mentions array that provides the best name for the entity group.,  - An array of mentions of the entity in the document. The array represents a co-reference group. See  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.html#how-targeted-sentiment-valuesCo-reference group for an example.amazonka-comprehendOne or more index into the Mentions array that provides the best name for the entity group.amazonka-comprehendAn array of mentions of the entity in the document. The array represents a co-reference group. See  https://docs.aws.amazon.com/comprehend/latest/dg/how-targeted-sentiment.html#how-targeted-sentiment-valuesCo-reference group for an example.\(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';4'amazonka-comprehend7Analysis results for one of the documents in the batch.See:  smart constructor.amazonka-comprehend(An array of targeted sentiment entities.amazonka-comprehend6The zero-based index of this result in the input list.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, + - An array of targeted sentiment entities., 9 - The zero-based index of this result in the input list.amazonka-comprehend(An array of targeted sentiment entities.amazonka-comprehend6The zero-based index of this result in the input list.](c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';=amazonka-comprehendProvides information for filtering topic detection jobs. For more information, see .See:  smart constructor.amazonka-comprehendFilters the list of topic detection jobs based on job status. Returns only jobs with the specified status.amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Only returns jobs submitted after the specified time. Jobs are returned in ascending order, oldest to newest.amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Only returns jobs submitted before the specified time. Jobs are returned in descending order, newest to oldest.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  -,  - Filters the list of topic detection jobs based on job status. Returns only jobs with the specified status.,  - Filters the list of jobs based on the time that the job was submitted for processing. Only returns jobs submitted after the specified time. Jobs are returned in ascending order, oldest to newest.,  - Filters the list of jobs based on the time that the job was submitted for processing. Only returns jobs submitted before the specified time. Jobs are returned in descending order, newest to oldest.amazonka-comprehendFilters the list of topic detection jobs based on job status. Returns only jobs with the specified status.amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Only returns jobs submitted after the specified time. Jobs are returned in ascending order, oldest to newest.amazonka-comprehendFilters the list of jobs based on the time that the job was submitted for processing. Only returns jobs submitted before the specified time. Jobs are returned in descending order, newest to oldest.  ^(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';Lramazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for the job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.See:  smart constructor.amazonka-comprehendThe ID number for a security group on an instance of your private VPC. Security groups on your VPC function serve as a virtual firewall to control inbound and outbound traffic and provides security for the resources that you@ll be accessing on the VPC. This ID number is preceded by "sg-", for instance: "sg-03b388029b0a285ea". For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.htmlSecurity Groups for your VPC.amazonka-comprehendThe ID for each subnet being used in your private VPC. This subnet is a subset of the a range of IPv4 addresses used by the VPC and is specific to a given availability zone in the VPC@s region. This ID number is preceded by "subnet-", for instance: "subnet-04ccf456919e69055". For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/VPC_Subnets.htmlVPCs and Subnets.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The ID number for a security group on an instance of your private VPC. Security groups on your VPC function serve as a virtual firewall to control inbound and outbound traffic and provides security for the resources that you@ll be accessing on the VPC. This ID number is preceded by "sg-", for instance: "sg-03b388029b0a285ea". For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.htmlSecurity Groups for your VPC.,  - The ID for each subnet being used in your private VPC. This subnet is a subset of the a range of IPv4 addresses used by the VPC and is specific to a given availability zone in the VPC@s region. This ID number is preceded by "subnet-", for instance: "subnet-04ccf456919e69055". For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/VPC_Subnets.htmlVPCs and Subnets.amazonka-comprehendThe ID number for a security group on an instance of your private VPC. Security groups on your VPC function serve as a virtual firewall to control inbound and outbound traffic and provides security for the resources that you@ll be accessing on the VPC. This ID number is preceded by "sg-", for instance: "sg-03b388029b0a285ea". For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/VPC_SecurityGroups.htmlSecurity Groups for your VPC.amazonka-comprehendThe ID for each subnet being used in your private VPC. This subnet is a subset of the a range of IPv4 addresses used by the VPC and is specific to a given availability zone in the VPC@s region. This ID number is preceded by "subnet-", for instance: "subnet-04ccf456919e69055". For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/VPC_Subnets.htmlVPCs and Subnets.amazonka-comprehendamazonka-comprehend_(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';ivamazonka-comprehend1Provides information about a topic detection job.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your job data.amazonka-comprehend4The time that the topic detection job was completed.amazonka-comprehendThe input data configuration supplied when you created the topic detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::topics-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:topics-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend3The identifier assigned to the topic detection job.amazonka-comprehend$The name of the topic detection job.amazonka-comprehendThe current status of the topic detection job. If the status is Failed-, the reason for the failure is shown in the Message field.amazonka-comprehend&A description for the status of a job.amazonka-comprehendThe number of topics to detect supplied when you created the topic detection job. The default is 10.amazonka-comprehendThe output data configuration supplied when you created the topic detection job.amazonka-comprehendThe time that the topic detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your topic detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your job data., 7 - The time that the topic detection job was completed.,  - The input data configuration supplied when you created the topic detection job.,  - The Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::topics-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:topics-detection-job/1234abcd12ab34cd56ef1234567890ab, 6 - The identifier assigned to the topic detection job., ' - The name of the topic detection job.,  - The current status of the topic detection job. If the status is Failed-, the reason for the failure is shown in the Message field., ) - A description for the status of a job.,  - The number of topics to detect supplied when you created the topic detection job. The default is 10.,  - The output data configuration supplied when you created the topic detection job.,  - The time that the topic detection job was submitted for processing.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your topic detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your job data.amazonka-comprehend4The time that the topic detection job was completed.amazonka-comprehendThe input data configuration supplied when you created the topic detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::topics-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:topics-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend3The identifier assigned to the topic detection job.amazonka-comprehend$The name of the topic detection job.amazonka-comprehendThe current status of the topic detection job. If the status is Failed-, the reason for the failure is shown in the Message field.amazonka-comprehend&A description for the status of a job.amazonka-comprehendThe number of topics to detect supplied when you created the topic detection job. The default is 10.amazonka-comprehendThe output data configuration supplied when you created the topic detection job.amazonka-comprehendThe time that the topic detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your topic detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.`(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';\amazonka-comprehend>Provides information about a targeted sentiment detection job.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend9The time that the targeted sentiment detection job ended.amazonka-comprehendThe Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::targeted-sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier assigned to the targeted sentiment detection job.amazonka-comprehendThe name that you assigned to the targeted sentiment detection job.amazonka-comprehendThe current status of the targeted sentiment detection job. If the status is FAILED, the Messages) field shows the reason for the failure.amazonka-comprehend)The language code of the input documents.amazonka-comprehend%A description of the status of a job.amazonka-comprehendThe time that the targeted sentiment detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the targeted sentiment detection job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data., < - The time that the targeted sentiment detection job ended.,  - Undocumented member.,  - The Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::targeted-sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier assigned to the targeted sentiment detection job.,  - The name that you assigned to the targeted sentiment detection job.,  - The current status of the targeted sentiment detection job. If the status is FAILED, the Messages) field shows the reason for the failure., , - The language code of the input documents., ( - A description of the status of a job.,  - Undocumented member.,  - The time that the targeted sentiment detection job was submitted for processing.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the targeted sentiment detection job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Undocumented member.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend9The time that the targeted sentiment detection job ended.amazonka-comprehendUndocumented member.amazonka-comprehendThe Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::targeted-sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier assigned to the targeted sentiment detection job.amazonka-comprehendThe name that you assigned to the targeted sentiment detection job.amazonka-comprehendThe current status of the targeted sentiment detection job. If the status is FAILED, the Messages) field shows the reason for the failure.amazonka-comprehend)The language code of the input documents.amazonka-comprehend%A description of the status of a job.amazonka-comprehendUndocumented member.amazonka-comprehendThe time that the targeted sentiment detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the targeted sentiment detection job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendUndocumented member.a(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';8amazonka-comprehend5Provides information about a sentiment detection job.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend0The time that the sentiment detection job ended.amazonka-comprehendThe input data configuration that you supplied when you created the sentiment detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend7The identifier assigned to the sentiment detection job.amazonka-comprehend9The name that you assigned to the sentiment detection jobamazonka-comprehendThe current status of the sentiment detection job. If the status is FAILED, the Messages( field shows the reason for the failure.amazonka-comprehend)The language code of the input documents.amazonka-comprehend%A description of the status of a job.amazonka-comprehendThe output data configuration that you supplied when you created the sentiment detection job.amazonka-comprehendThe time that the sentiment detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your sentiment detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data., 3 - The time that the sentiment detection job ended.,  - The input data configuration that you supplied when you created the sentiment detection job.,  - The Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab, : - The identifier assigned to the sentiment detection job., < - The name that you assigned to the sentiment detection job,  - The current status of the sentiment detection job. If the status is FAILED, the Messages( field shows the reason for the failure., , - The language code of the input documents., ( - A description of the status of a job.,  - The output data configuration that you supplied when you created the sentiment detection job.,  - The time that the sentiment detection job was submitted for processing.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your sentiment detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend0The time that the sentiment detection job ended.amazonka-comprehendThe input data configuration that you supplied when you created the sentiment detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend7The identifier assigned to the sentiment detection job.amazonka-comprehend9The name that you assigned to the sentiment detection jobamazonka-comprehendThe current status of the sentiment detection job. If the status is FAILED, the Messages( field shows the reason for the failure.amazonka-comprehend)The language code of the input documents.amazonka-comprehend%A description of the status of a job.amazonka-comprehendThe output data configuration that you supplied when you created the sentiment detection job.amazonka-comprehendThe time that the sentiment detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your sentiment detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.b(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';^amazonka-comprehend7Provides information about a key phrases detection job.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend6The time that the key phrases detection job completed.amazonka-comprehendThe input data configuration that you supplied when you created the key phrases detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the key phrases detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::key-phrases-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend9The identifier assigned to the key phrases detection job.amazonka-comprehend9The name that you assigned the key phrases detection job.amazonka-comprehendThe current status of the key phrases detection job. If the status is FAILED, the Message( field shows the reason for the failure.amazonka-comprehend)The language code of the input documents.amazonka-comprehend%A description of the status of a job.amazonka-comprehendThe output data configuration that you supplied when you created the key phrases detection job.amazonka-comprehendThe time that the key phrases detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your key phrases detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data., 9 - The time that the key phrases detection job completed.,  - The input data configuration that you supplied when you created the key phrases detection job.,  - The Amazon Resource Name (ARN) of the key phrases detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::key-phrases-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890ab, < - The identifier assigned to the key phrases detection job., < - The name that you assigned the key phrases detection job.,  - The current status of the key phrases detection job. If the status is FAILED, the Message( field shows the reason for the failure., , - The language code of the input documents., ( - A description of the status of a job.,  - The output data configuration that you supplied when you created the key phrases detection job.,  - The time that the key phrases detection job was submitted for processing.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your key phrases detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend6The time that the key phrases detection job completed.amazonka-comprehendThe input data configuration that you supplied when you created the key phrases detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the key phrases detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::key-phrases-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend9The identifier assigned to the key phrases detection job.amazonka-comprehend9The name that you assigned the key phrases detection job.amazonka-comprehendThe current status of the key phrases detection job. If the status is FAILED, the Message( field shows the reason for the failure.amazonka-comprehend)The language code of the input documents.amazonka-comprehend%A description of the status of a job.amazonka-comprehendThe output data configuration that you supplied when you created the key phrases detection job.amazonka-comprehendThe time that the key phrases detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your key phrases detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.c(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';"amazonka-comprehend1Describes information about an entity recognizer.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehend0The time that the recognizer creation completed.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer.amazonka-comprehend2The input data properties of an entity recognizer.amazonka-comprehendThe language of the input documents. All documents must be in the same language. Only English ("en") is currently supported.amazonka-comprehend.A description of the status of the recognizer.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehend0Provides information about an entity recognizer.amazonka-comprehendThe Amazon Resource Name (ARN) of the source model. This model was imported from a different AWS account to create the entity recognizer model in your AWS account.amazonka-comprehend-Provides the status of the entity recognizer.amazonka-comprehend:The time that the recognizer was submitted for processing.amazonka-comprehend>The time that training of the entity recognizer was completed.amazonka-comprehend8The time that training of the entity recognizer started.amazonka-comprehend7The version name you assigned to the entity recognizer.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your custom entity recognizer. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data., 3 - The time that the recognizer creation completed.,  - The Amazon Resource Name (ARN) that identifies the entity recognizer., 5 - The input data properties of an entity recognizer.,  - The language of the input documents. All documents must be in the same language. Only English ("en") is currently supported., 1 - A description of the status of the recognizer.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab", 3 - Provides information about an entity recognizer.,  - The Amazon Resource Name (ARN) of the source model. This model was imported from a different AWS account to create the entity recognizer model in your AWS account., 0 - Provides the status of the entity recognizer., = - The time that the recognizer was submitted for processing.,  - The time that training of the entity recognizer was completed., ; - The time that training of the entity recognizer started., : - The version name you assigned to the entity recognizer.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your custom entity recognizer. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehend0The time that the recognizer creation completed.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer.amazonka-comprehend2The input data properties of an entity recognizer.amazonka-comprehendThe language of the input documents. All documents must be in the same language. Only English ("en") is currently supported.amazonka-comprehend.A description of the status of the recognizer.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehend0Provides information about an entity recognizer.amazonka-comprehendThe Amazon Resource Name (ARN) of the source model. This model was imported from a different AWS account to create the entity recognizer model in your AWS account.amazonka-comprehend-Provides the status of the entity recognizer.amazonka-comprehend:The time that the recognizer was submitted for processing.amazonka-comprehend>The time that training of the entity recognizer was completed.amazonka-comprehend8The time that training of the entity recognizer started.amazonka-comprehend7The version name you assigned to the entity recognizer.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your custom entity recognizer. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.##d(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';amazonka-comprehend5Provides information about an entities detection job.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend2The time that the entities detection job completedamazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer.amazonka-comprehendThe input data configuration that you supplied when you created the entities detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend6The identifier assigned to the entities detection job.amazonka-comprehend6The name that you assigned the entities detection job.amazonka-comprehendThe current status of the entities detection job. If the status is FAILED, the Message( field shows the reason for the failure.amazonka-comprehend)The language code of the input documents.amazonka-comprehend%A description of the status of a job.amazonka-comprehendThe output data configuration that you supplied when you created the entities detection job.amazonka-comprehendThe time that the entities detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your entity detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data., 5 - The time that the entities detection job completed,  - The Amazon Resource Name (ARN) that identifies the entity recognizer.,  - The input data configuration that you supplied when you created the entities detection job.,  - The Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890ab, 9 - The identifier assigned to the entities detection job., 9 - The name that you assigned the entities detection job.,  - The current status of the entities detection job. If the status is FAILED, the Message( field shows the reason for the failure., , - The language code of the input documents., ( - A description of the status of a job.,  - The output data configuration that you supplied when you created the entities detection job.,  - The time that the entities detection job was submitted for processing.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your entity detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend2The time that the entities detection job completedamazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer.amazonka-comprehendThe input data configuration that you supplied when you created the entities detection job.amazonka-comprehendThe Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend6The identifier assigned to the entities detection job.amazonka-comprehend6The name that you assigned the entities detection job.amazonka-comprehendThe current status of the entities detection job. If the status is FAILED, the Message( field shows the reason for the failure.amazonka-comprehend)The language code of the input documents.amazonka-comprehend%A description of the status of a job.amazonka-comprehendThe output data configuration that you supplied when you created the entities detection job.amazonka-comprehendThe time that the entities detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your entity detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.e(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&'; amazonka-comprehend=Provides information about a dominant language detection job.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend:comprehend:::dominant-language-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend?The identifier assigned to the dominant language detection job.amazonka-comprehendThe name that you assigned to the dominant language detection job.amazonka-comprehendThe current status of the dominant language detection job. If the status is FAILED, the Message( field shows the reason for the failure.amazonka-comprehend&A description for the status of a job.amazonka-comprehendThe output data configuration that you supplied when you created the dominant language detection job.amazonka-comprehendThe time that the dominant language detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your dominant language detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data., ? - The time that the dominant language detection job completed.,  - The input data configuration that you supplied when you created the dominant language detection job.,  - The Amazon Resource Name (ARN) of the dominant language detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::dominant-language-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier assigned to the dominant language detection job.,  - The name that you assigned to the dominant language detection job.,  - The current status of the dominant language detection job. If the status is FAILED, the Message( field shows the reason for the failure., ) - A description for the status of a job.,  - The output data configuration that you supplied when you created the dominant language detection job.,  - The time that the dominant language detection job was submitted for processing.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your dominant language detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) that gives Amazon Comprehend read access to your input data.amazonka-comprehend:comprehend:::dominant-language-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend?The identifier assigned to the dominant language detection job.amazonka-comprehendThe name that you assigned to the dominant language detection job.amazonka-comprehendThe current status of the dominant language detection job. If the status is FAILED, the Message( field shows the reason for the failure.amazonka-comprehend&A description for the status of a job.amazonka-comprehendThe output data configuration that you supplied when you created the dominant language detection job.amazonka-comprehendThe time that the dominant language detection job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your dominant language detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.f(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';C&&amazonka-comprehend1Provides information about a document classifier.See:  smart constructor.amazonka-comprehendInformation about the document classifier, including the number of documents used for training the classifier, the number of documents used for test the classifier, and an accuracy rating.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier.amazonka-comprehend9The time that training the document classifier completed.amazonka-comprehendThe input data configuration that you supplied when you created the document classifier for training.amazonka-comprehendThe language code for the language of the documents that the classifier was trained on.amazonka-comprehend:Additional information about the status of the classifier.amazonka-comprehendIndicates the mode in which the specific classifier was trained. This also indicates the format of input documents and the format of the confusion matrix. Each classifier can only be trained in one mode and this cannot be changed once the classifier is trained.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendProvides output results configuration parameters for custom classifier jobs.amazonka-comprehendThe Amazon Resource Name (ARN) of the source model. This model was imported from a different AWS account to create the document classifier model in your AWS account.amazonka-comprehend8The status of the document classifier. If the status is TRAINED3 the classifier is ready to use. If the status is FAILED you can see additional information about why the classifier wasn't trained in the Message field.amazonka-comprehendThe time that the document classifier was submitted for training.amazonka-comprehendThe time that training of the document classifier was completed. Indicates the time when the training completes on documentation classifiers. You are billed for the time interval between this time and the value of TrainingStartTime.amazonka-comprehendIndicates the time when the training starts on documentation classifiers. You are billed for the time interval between this time and the value of TrainingEndTime.amazonka-comprehend>The version name that you assigned to the document classifier.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your custom classifier. For more information, see  https://docs.aws.amazon.com/vppc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Information about the document classifier, including the number of documents used for training the classifier, the number of documents used for test the classifier, and an accuracy rating.,  - The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.,  - The Amazon Resource Name (ARN) that identifies the document classifier., < - The time that training the document classifier completed.,  - The input data configuration that you supplied when you created the document classifier for training.,  - The language code for the language of the documents that the classifier was trained on., = - Additional information about the status of the classifier.,  - Indicates the mode in which the specific classifier was trained. This also indicates the format of input documents and the format of the confusion matrix. Each classifier can only be trained in one mode and this cannot be changed once the classifier is trained.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Provides output results configuration parameters for custom classifier jobs.,  - The Amazon Resource Name (ARN) of the source model. This model was imported from a different AWS account to create the document classifier model in your AWS account., ; - The status of the document classifier. If the status is TRAINED3 the classifier is ready to use. If the status is FAILED you can see additional information about why the classifier wasn't trained in the Message field.,  - The time that the document classifier was submitted for training.,  - The time that training of the document classifier was completed. Indicates the time when the training completes on documentation classifiers. You are billed for the time interval between this time and the value of TrainingStartTime.,  - Indicates the time when the training starts on documentation classifiers. You are billed for the time interval between this time and the value of TrainingEndTime.,  - The version name that you assigned to the document classifier.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your custom classifier. For more information, see  https://docs.aws.amazon.com/vppc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendInformation about the document classifier, including the number of documents used for training the classifier, the number of documents used for test the classifier, and an accuracy rating.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier.amazonka-comprehend9The time that training the document classifier completed.amazonka-comprehendThe input data configuration that you supplied when you created the document classifier for training.amazonka-comprehendThe language code for the language of the documents that the classifier was trained on.amazonka-comprehend:Additional information about the status of the classifier.amazonka-comprehendIndicates the mode in which the specific classifier was trained. This also indicates the format of input documents and the format of the confusion matrix. Each classifier can only be trained in one mode and this cannot be changed once the classifier is trained.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendProvides output results configuration parameters for custom classifier jobs.amazonka-comprehendThe Amazon Resource Name (ARN) of the source model. This model was imported from a different AWS account to create the document classifier model in your AWS account.amazonka-comprehend8The status of the document classifier. If the status is TRAINED3 the classifier is ready to use. If the status is FAILED you can see additional information about why the classifier wasn't trained in the Message field.amazonka-comprehendThe time that the document classifier was submitted for training.amazonka-comprehendThe time that training of the document classifier was completed. Indicates the time when the training completes on documentation classifiers. You are billed for the time interval between this time and the value of TrainingStartTime.amazonka-comprehendIndicates the time when the training starts on documentation classifiers. You are billed for the time interval between this time and the value of TrainingEndTime.amazonka-comprehend>The version name that you assigned to the document classifier.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your custom classifier. For more information, see  https://docs.aws.amazon.com/vppc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.''g(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%&';aamazonka-comprehend9Provides information about a document classification job.See:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier.amazonka-comprehend8The time that the document classification job completed.amazonka-comprehendThe input data configuration that you supplied when you created the document classification job.amazonka-comprehendThe Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::document-classification-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend;The identifier assigned to the document classification job.amazonka-comprehend>The name that you assigned to the document classification job.amazonka-comprehendThe current status of the document classification job. If the status is FAILED, the Message( field shows the reason for the failure.amazonka-comprehend'A description of the status of the job.amazonka-comprehendThe output data configuration that you supplied when you created the document classification job.amazonka-comprehendThe time that the document classification job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your document classification job. For more information, see  https://docs.aws.amazon.com/vppc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.,  - The Amazon Resource Name (ARN) that identifies the document classifier., ; - The time that the document classification job completed.,  - The input data configuration that you supplied when you created the document classification job.,  - The Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::document-classification-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab, > - The identifier assigned to the document classification job.,  - The name that you assigned to the document classification job.,  - The current status of the document classification job. If the status is FAILED, the Message( field shows the reason for the failure., * - A description of the status of the job.,  - The output data configuration that you supplied when you created the document classification job.,  - The time that the document classification job was submitted for processing.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your document classification job. For more information, see  https://docs.aws.amazon.com/vppc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier.amazonka-comprehend8The time that the document classification job completed.amazonka-comprehendThe input data configuration that you supplied when you created the document classification job.amazonka-comprehendThe Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::document-classification-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend;The identifier assigned to the document classification job.amazonka-comprehend>The name that you assigned to the document classification job.amazonka-comprehendThe current status of the document classification job. If the status is FAILED, the Message( field shows the reason for the failure.amazonka-comprehend'A description of the status of the job.amazonka-comprehendThe output data configuration that you supplied when you created the document classification job.amazonka-comprehendThe time that the document classification job was submitted for processing.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for a private Virtual Private Cloud (VPC) containing the resources you are using for your document classification job. For more information, see  https://docs.aws.amazon.com/vppc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.h(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"%kamazonka-comprehend API version  2017-11-27, of the Amazon Comprehend SDK configuration.amazonka-comprehendThe number of documents in the request exceeds the limit of 25. Try your request again with fewer documents.amazonka-comprehendConcurrent modification of the tags associated with an Amazon Comprehend resource is not supported.amazonka-comprehend6An internal server error occurred. Retry your request.amazonka-comprehendThe filter specified for the operation is invalid. Specify a different filter.amazonka-comprehendThe request is invalid.amazonka-comprehendThe specified job was not found. Check the job ID and try again.amazonka-comprehendThe KMS customer managed key (CMK) entered cannot be validated. Verify the key and re-enter it.amazonka-comprehendThe specified resource name is already in use. Use a different name and try your request again.amazonka-comprehendThe maximum number of resources per account has been exceeded. Review the resources, and then try your request again.amazonka-comprehendThe specified resource ARN was not found. Check the ARN and try your request again.amazonka-comprehendThe specified resource is not available. Check the resource and try your request again.amazonka-comprehendThe size of the input text exceeds the limit. Use a smaller document.amazonka-comprehendThe number of requests exceeds the limit. Resubmit your request later.amazonka-comprehendThe request contains more tag keys than can be associated with a resource (50 tag keys per resource).amazonka-comprehendThe request contains more tags than can be associated with a resource (50 tags per resource). The maximum number of tags includes both existing tags and those included in your current request.amazonka-comprehendAmazon Comprehend can't process the language of the input text. For custom entity recognition APIs, only English, Spanish, French, Italian, German, or Portuguese are accepted. For a list of supported languages,  https://docs.aws.amazon.com/comprehend/latest/dg/supported-languages.htmlSupported languages$ in the Comprehend Developer Guide.  (,+)*@ACBEDFGHIJRSTUVWXYZbcdgefhijkltuvwxyz{|}~  (,+)*,+    bcdgefhijkl@ACBEDFGHIJRSTUVWXYZtuvwxyz{|}~       i(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the given Amazon Comprehend resource to which you want to associate the tags.amazonka-comprehendTags being associated with a specific Amazon Comprehend resource. There can be a maximum of 50 tags (both existing and pending) associated with a specific resource.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the given Amazon Comprehend resource to which you want to associate the tags.,  - Tags being associated with a specific Amazon Comprehend resource. There can be a maximum of 50 tags (both existing and pending) associated with a specific resource.amazonka-comprehendThe Amazon Resource Name (ARN) of the given Amazon Comprehend resource to which you want to associate the tags.amazonka-comprehendTags being associated with a specific Amazon Comprehend resource. There can be a maximum of 50 tags (both existing and pending) associated with a specific resource.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  j(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Famazonka-comprehendSee:  smart constructor.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer currently being trained.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) that identifies the entity recognizer currently being trained.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer currently being trained.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  k(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier currently being trained.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) that identifies the document classifier currently being trained.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier currently being trained.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, # - The response's http status code.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  l(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';U amazonka-comprehendSee:  smart constructor.amazonka-comprehend?The identifier of the targeted sentiment detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopSentimentDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend?The identifier of the targeted sentiment detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier of the targeted sentiment detection job to stop.amazonka-comprehend?The identifier of the targeted sentiment detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier of the targeted sentiment detection job to stop.,  - Either STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopSentimentDetectionJob operation., # - The response's http status code.amazonka-comprehend?The identifier of the targeted sentiment detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopSentimentDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendm(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend6The identifier of the sentiment detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopSentimentDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend6The identifier of the sentiment detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 9 - The identifier of the sentiment detection job to stop.amazonka-comprehend6The identifier of the sentiment detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 9 - The identifier of the sentiment detection job to stop.,  - Either STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopSentimentDetectionJob operation., # - The response's http status code.amazonka-comprehend6The identifier of the sentiment detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopSentimentDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendn(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend9The identifier of the PII entities detection job to stop.amazonka-comprehend-The status of the PII entities detection job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend9The identifier of the PII entities detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, < - The identifier of the PII entities detection job to stop.amazonka-comprehend9The identifier of the PII entities detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, < - The identifier of the PII entities detection job to stop., 0 - The status of the PII entities detection job., # - The response's http status code.amazonka-comprehend9The identifier of the PII entities detection job to stop.amazonka-comprehend-The status of the PII entities detection job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendo(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend8The identifier of the key phrases detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopKeyPhrasesDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend8The identifier of the key phrases detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, ; - The identifier of the key phrases detection job to stop.amazonka-comprehend8The identifier of the key phrases detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, ; - The identifier of the key phrases detection job to stop.,  - Either STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopKeyPhrasesDetectionJob operation., # - The response's http status code.amazonka-comprehend8The identifier of the key phrases detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopKeyPhrasesDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendp(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend3The identifier of the events detection job to stop.amazonka-comprehend'The status of the events detection job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend3The identifier of the events detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 6 - The identifier of the events detection job to stop.amazonka-comprehend3The identifier of the events detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 6 - The identifier of the events detection job to stop., * - The status of the events detection job., # - The response's http status code.amazonka-comprehend3The identifier of the events detection job to stop.amazonka-comprehend'The status of the events detection job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendq(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend5The identifier of the entities detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopEntitiesDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend5The identifier of the entities detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 8 - The identifier of the entities detection job to stop.amazonka-comprehend5The identifier of the entities detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 8 - The identifier of the entities detection job to stop.,  - Either STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopEntitiesDetectionJob operation., # - The response's http status code.amazonka-comprehend5The identifier of the entities detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED- if the job was previously stopped with the StopEntitiesDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendr(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Ԧ amazonka-comprehendSee:  smart constructor.amazonka-comprehend>The identifier of the dominant language detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED. if the job was previously stopped with the  StopDominantLanguageDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend>The identifier of the dominant language detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier of the dominant language detection job to stop.amazonka-comprehend>The identifier of the dominant language detection job to stop.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier of the dominant language detection job to stop.,  - Either STOP_REQUESTED% if the job is currently running, or STOPPED. if the job was previously stopped with the  StopDominantLanguageDetectionJob operation., # - The response's http status code.amazonka-comprehend>The identifier of the dominant language detection job to stop.amazonka-comprehendEither STOP_REQUESTED% if the job is currently running, or STOPPED. if the job was previously stopped with the  StopDominantLanguageDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehends(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::topics-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of the job, use this identifier with the DescribeTopicDetectionJob operation.amazonka-comprehendThe status of the job:SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the DescribeTopicDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendThe number of topics to detect.amazonka-comprehendTags to be associated with the topics detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your topic detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehendSpecifies where to send the output files. The output is a compressed archive with two files, topic-terms.csv7 that lists the terms associated with each topic, and doc-topics.csv5 that lists the documents associated with each topicamazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.,  - The identifier of the job., " - The number of topics to detect.,  - Tags to be associated with the topics detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your topic detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.,  - Specifies the format and location of the input data for the job.,  - Specifies where to send the output files. The output is a compressed archive with two files, topic-terms.csv7 that lists the terms associated with each topic, and doc-topics.csv5 that lists the documents associated with each topic,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendA unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendThe number of topics to detect.amazonka-comprehendTags to be associated with the topics detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your topic detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehendSpecifies where to send the output files. The output is a compressed archive with two files, topic-terms.csv7 that lists the terms associated with each topic, and doc-topics.csv5 that lists the documents associated with each topicamazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::topics-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier generated for the job. To get the status of the job, use this identifier with the DescribeTopicDetectionJob operation.,  - The status of the job:SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the DescribeTopicDetectionJob operation., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the topics detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::topics-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of the job, use this identifier with the DescribeTopicDetectionJob operation.amazonka-comprehendThe status of the job:SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the DescribeTopicDetectionJob operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehend  t(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::targeted-sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of a job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the targeted sentiment detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the KMS key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissionsRole-based permissions.amazonka-comprehendThe language of the input documents. Currently, English is the only supported language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.,  - The identifier of the job.,  - Tags to be associated with the targeted sentiment detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - ID for the KMS key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Undocumented member.,  - Undocumented member., , - Specifies where to send the output files.,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissionsRole-based permissions.,  - The language of the input documents. Currently, English is the only supported language.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the targeted sentiment detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the KMS key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendUndocumented member.amazonka-comprehendUndocumented member.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissionsRole-based permissions.amazonka-comprehendThe language of the input documents. Currently, English is the only supported language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::targeted-sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier generated for the job. To get the status of a job, use this identifier with the operation.,  - The status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the targeted sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::targeted-sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:targeted-sentiment-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of a job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehend  u(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Famazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of a job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the sentiment detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your sentiment detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.,  - The identifier of the job.,  - Tags to be associated with the sentiment detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your sentiment detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.,  - Specifies the format and location of the input data for the job., , - Specifies where to send the output files.,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.,  - The language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the sentiment detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your sentiment detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier generated for the job. To get the status of a job, use this identifier with the operation.,  - The status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the sentiment detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::sentiment-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:sentiment-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of a job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehend  v(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';camazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the PII entity detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::pii-entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend%The identifier generated for the job.amazonka-comprehendThe status of the job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehend;Provides configuration parameters for PII entity redaction.*This parameter is required if you set the Mode parameter to ONLY_REDACTION#. In that case, you must provide a RedactionConfig definition that includes the PiiEntityTypes parameter.amazonka-comprehendTags to be associated with the PII entities detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehend6The input properties for a PII entities detection job.amazonka-comprehendProvides conguration parameters for the output of PII entity detection jobs.amazonka-comprehendSpecifies whether the output provides the locations (offsets) of PII entities or a file in which PII entities are redacted.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendThe language of the input documents. Currently, English is the only valid language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.,  - The identifier of the job., > - Provides configuration parameters for PII entity redaction.*This parameter is required if you set the Mode parameter to ONLY_REDACTION#. In that case, you must provide a RedactionConfig definition that includes the PiiEntityTypes parameter.,  - Tags to be associated with the PII entities detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department., 9 - The input properties for a PII entities detection job.,  - Provides conguration parameters for the output of PII entity detection jobs.,  - Specifies whether the output provides the locations (offsets) of PII entities or a file in which PII entities are redacted.,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.,  - The language of the input documents. Currently, English is the only valid language.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehend;Provides configuration parameters for PII entity redaction.*This parameter is required if you set the Mode parameter to ONLY_REDACTION#. In that case, you must provide a RedactionConfig definition that includes the PiiEntityTypes parameter.amazonka-comprehendTags to be associated with the PII entities detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehend6The input properties for a PII entities detection job.amazonka-comprehendProvides conguration parameters for the output of PII entity detection jobs.amazonka-comprehendSpecifies whether the output provides the locations (offsets) of PII entities or a file in which PII entities are redacted.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendThe language of the input documents. Currently, English is the only valid language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the PII entity detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::pii-entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890ab, ( - The identifier generated for the job.,  - The status of the job., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the PII entity detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::pii-entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:pii-entities-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehend%The identifier generated for the job.amazonka-comprehendThe status of the job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehend  w(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the key phrase detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::key-phrases-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of a job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the key phrases detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your key phrases detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.,  - The identifier of the job.,  - Tags to be associated with the key phrases detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your key phrases detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.,  - Specifies the format and location of the input data for the job., , - Specifies where to send the output files.,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.,  - The language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the key phrases detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your key phrases detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the key phrase detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::key-phrases-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier generated for the job. To get the status of a job, use this identifier with the operation.,  - The status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the key phrase detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::key-phrases-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:key-phrases-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of a job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehend  x(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::events-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendAn unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehend'The status of the events detection job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendAn unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehend+The identifier of the events detection job.amazonka-comprehendTags to be associated with the events detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehend)The language code of the input documents.amazonka-comprehend5The types of events to detect in the input documents.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - An unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one., . - The identifier of the events detection job.,  - Tags to be associated with the events detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - Specifies the format and location of the input data for the job., , - Specifies where to send the output files.,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data., , - The language code of the input documents., 8 - The types of events to detect in the input documents.amazonka-comprehendAn unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehend+The identifier of the events detection job.amazonka-comprehendTags to be associated with the events detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehend)The language code of the input documents.amazonka-comprehend5The types of events to detect in the input documents.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::events-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890ab,  - An unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one., * - The status of the events detection job., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the events detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::events-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:events-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendAn unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehend'The status of the events detection job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendy(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';R amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.STOP_REQUESTED - Amazon Comprehend has received a stop request for the job and is processing the request.>STOPPED - The job was successfully stopped without completing.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the specific entity recognizer to be used by the StartEntitiesDetectionJob. This ARN is optional and is only used for a custom entity recognition job.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the entities detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your entity detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendThe language of the input documents. All documents must be in the same language. You can specify any of the languages supported by Amazon Comprehend. If custom entities recognition is used, this parameter is ignored and the language used for training the model is used instead.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.,  - The Amazon Resource Name (ARN) that identifies the specific entity recognizer to be used by the StartEntitiesDetectionJob. This ARN is optional and is only used for a custom entity recognition job.,  - The identifier of the job.,  - Tags to be associated with the entities detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your entity detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.,  - Specifies the format and location of the input data for the job., , - Specifies where to send the output files.,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.,  - The language of the input documents. All documents must be in the same language. You can specify any of the languages supported by Amazon Comprehend. If custom entities recognition is used, this parameter is ignored and the language used for training the model is used instead.amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the specific entity recognizer to be used by the StartEntitiesDetectionJob. This ARN is optional and is only used for a custom entity recognition job.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the entities detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your entity detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendThe language of the input documents. All documents must be in the same language. You can specify any of the languages supported by Amazon Comprehend. If custom entities recognition is used, this parameter is ignored and the language used for training the model is used instead.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier generated for the job. To get the status of job, use this identifier with the operation.,  - The status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.STOP_REQUESTED - Amazon Comprehend has received a stop request for the job and is processing the request.>STOPPED - The job was successfully stopped without completing., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the entities detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::entities-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:entities-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.STOP_REQUESTED - Amazon Comprehend has received a stop request for the job and is processing the request.>STOPPED - The job was successfully stopped without completing.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehend""z(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the dominant language detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::dominant-language-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of a job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.amazonka-comprehendAn identifier for the job.amazonka-comprehendTags to be associated with the dominant language detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your dominant language detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.,  - An identifier for the job.,  - Tags to be associated with the dominant language detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your dominant language detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.,  - Specifies the format and location of the input data for the job., , - Specifies where to send the output files.,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendA unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.amazonka-comprehendAn identifier for the job.amazonka-comprehendTags to be associated with the dominant language detection job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your dominant language detection job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data. For more information, see  https://docs.aws.amazon.com/comprehend/latest/dg/access-control-managing-permissions.html#auth-role-permissions.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the dominant language detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::dominant-language-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier generated for the job. To get the status of a job, use this identifier with the operation.,  - The status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the dominant language detection job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::dominant-language-detection-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:dominant-language-detection-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of a job, use this identifier with the operation.amazonka-comprehendThe status of the job.SUBMITTED - The job has been received and is queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. To get details, use the operation.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehend{(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Xamazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::document-classification-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of the job, use this identifier with the operation.amazonka-comprehendThe status of the job:SUBMITTED - The job has been received and queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. For details, use the operation.STOP_REQUESTED - Amazon Comprehend has received a stop request for the job and is processing the request.>STOPPED - The job was successfully stopped without completing.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the document classification job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your document classification job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) of the document classifier to use to process the job.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.,  - The identifier of the job.,  - Tags to be associated with the document classification job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - Configuration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your document classification job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.,  - The Amazon Resource Name (ARN) of the document classifier to use to process the job.,  - Specifies the format and location of the input data for the job., , - Specifies where to send the output files.,  - The Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendA unique identifier for the request. If you do not set the client request token, Amazon Comprehend generates one.amazonka-comprehendThe identifier of the job.amazonka-comprehendTags to be associated with the document classification job. A tag is a key-value pair that adds metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your document classification job. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.amazonka-comprehendThe Amazon Resource Name (ARN) of the document classifier to use to process the job.amazonka-comprehendSpecifies the format and location of the input data for the job.amazonka-comprehend)Specifies where to send the output files.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Access Management (IAM) role that grants Amazon Comprehend read access to your input data.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::document-classification-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890ab,  - The identifier generated for the job. To get the status of the job, use this identifier with the operation.,  - The status of the job:SUBMITTED - The job has been received and queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. For details, use the operation.STOP_REQUESTED - Amazon Comprehend has received a stop request for the job and is processing the request.>STOPPED - The job was successfully stopped without completing., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the document classification job. It is a unique, fully qualified identifier for the job. It includes the AWS account, Region, and the job ID. The format of the ARN is as follows: arn::comprehend:::document-classification-job/$The following is an example job ARN: arn:aws:comprehend:us-west-2:111122223333:document-classification-job/1234abcd12ab34cd56ef1234567890abamazonka-comprehendThe identifier generated for the job. To get the status of the job, use this identifier with the operation.amazonka-comprehendThe status of the job:SUBMITTED - The job has been received and queued for processing.6IN_PROGRESS - Amazon Comprehend is processing the job.COMPLETED - The job was successfully completed and the output is available.FAILED - The job did not complete. For details, use the operation.STOP_REQUESTED - Amazon Comprehend has received a stop request for the job and is processing the request.>STOPPED - The job was successfully stopped without completing.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehendamazonka-comprehend  |(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';0wamazonka-comprehendSee:  smart constructor.amazonka-comprehendThe revision ID of the policy. Each time you modify a policy, Amazon Comprehend assigns a new revision ID, and it deletes the prior version of the policy.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe revision ID that Amazon Comprehend assigned to the policy that you are updating. If you are creating a new policy that has no prior version, don't use this parameter. Amazon Comprehend creates the revision ID for you.amazonka-comprehendThe Amazon Resource Name (ARN) of the custom model to attach the policy to.amazonka-comprehendThe JSON resource-based policy to attach to your custom model. Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The revision ID that Amazon Comprehend assigned to the policy that you are updating. If you are creating a new policy that has no prior version, don't use this parameter. Amazon Comprehend creates the revision ID for you.,  - The Amazon Resource Name (ARN) of the custom model to attach the policy to.,  - The JSON resource-based policy to attach to your custom model. Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'amazonka-comprehendThe revision ID that Amazon Comprehend assigned to the policy that you are updating. If you are creating a new policy that has no prior version, don't use this parameter. Amazon Comprehend creates the revision ID for you.amazonka-comprehendThe Amazon Resource Name (ARN) of the custom model to attach the policy to.amazonka-comprehendThe JSON resource-based policy to attach to your custom model. Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The revision ID of the policy. Each time you modify a policy, Amazon Comprehend assigns a new revision ID, and it deletes the prior version of the policy., # - The response's http status code.amazonka-comprehendThe revision ID of the policy. Each time you modify a policy, Amazon Comprehend assigns a new revision ID, and it deletes the prior version of the policy.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehend}(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';;amazonka-comprehendSee:  smart constructor.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. Jobs can be filtered on their name, status, or the date and time that they were submitted. You can set only one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. Jobs can be filtered on their name, status, or the date and time that they were submitted. You can set only one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. Jobs can be filtered on their name, status, or the date and time that they were submitted. You can set only one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 1 - Identifies the next page of results to return.,  - A list containing the properties of each job that is returned., # - The response's http status code.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend The response's http status code.amazonka-comprehend~(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Famazonka-comprehendSee:  smart constructor.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 1 - Identifies the next page of results to return.,  - A list containing the properties of each job that is returned., # - The response's http status code.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';R amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the given Amazon Comprehend resource you are querying.amazonka-comprehendTags associated with the Amazon Comprehend resource being queried. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the given Amazon Comprehend resource you are querying.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the given Amazon Comprehend resource you are querying.amazonka-comprehendThe Amazon Resource Name (ARN) of the given Amazon Comprehend resource you are querying.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the given Amazon Comprehend resource you are querying.,  - Tags associated with the Amazon Comprehend resource being queried. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the given Amazon Comprehend resource you are querying.amazonka-comprehendTags associated with the Amazon Comprehend resource being queried. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';]amazonka-comprehendSee:  smart constructor.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 1 - Identifies the next page of results to return.,  - A list containing the properties of each job that is returned., # - The response's http status code.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';hamazonka-comprehendSee:  smart constructor.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehend5The maximum number of results to return in each page.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time., 8 - The maximum number of results to return in each page., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehend5The maximum number of results to return in each page.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 1 - Identifies the next page of results to return.,  - A list containing the properties of each job that is returned., # - The response's http status code.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';tamazonka-comprehendSee:  smart constructor.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A list containing the properties of each job that is returned., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehend5The maximum number of results to return in each page.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time., 8 - The maximum number of results to return in each page., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehend5The maximum number of results to return in each page.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A list containing the properties of each job that is returned., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehend/The list of properties of an entity recognizer.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend9Filters the list of entities returned. You can filter on Status, SubmitTimeBefore, or SubmitTimeAfter). You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return on each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, < - Filters the list of entities returned. You can filter on Status, SubmitTimeBefore, or SubmitTimeAfter). You can only set one filter at a time.,  - The maximum number of results to return on each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehend9Filters the list of entities returned. You can filter on Status, SubmitTimeBefore, or SubmitTimeAfter). You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return on each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 2 - The list of properties of an entity recognizer., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehend/The list of properties of an entity recognizer.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehend%The list entity recognizer summaries.amazonka-comprehend%The list entity recognizer summaries.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe maximum number of results to return on each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The maximum number of results to return on each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendThe maximum number of results to return on each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, ( - The list entity recognizer summaries., ( - The list entity recognizer summaries., # - The response's http status code.amazonka-comprehend%The list entity recognizer summaries.amazonka-comprehend%The list entity recognizer summaries.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A list containing the properties of each job that is returned., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehendDisplays a list of endpoint properties being retrieved by the service in response to the request.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the endpoints that are returned. You can filter endpoints on their name, model, status, or the date and time that they were created. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the endpoints that are returned. You can filter endpoints on their name, model, status, or the date and time that they were created. You can only set one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the endpoints that are returned. You can filter endpoints on their name, model, status, or the date and time that they were created. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Displays a list of endpoint properties being retrieved by the service in response to the request., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehendDisplays a list of endpoint properties being retrieved by the service in response to the request.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';&amazonka-comprehendSee:  smart constructor.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters that jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters that jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters that jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A list containing the properties of each job that is returned., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehend>A list containing the properties of each job that is returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Tamazonka-comprehendSee:  smart constructor.amazonka-comprehend6A list containing the properties of each job returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their name, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 9 - A list containing the properties of each job returned., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehend6A list containing the properties of each job returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';)amazonka-comprehendSee:  smart constructor.amazonka-comprehend.The list of summaries of document classifiers.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe maximum number of results to return on each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The maximum number of results to return on each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendThe maximum number of results to return on each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 1 - The list of summaries of document classifiers., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehend.The list of summaries of document classifiers.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Ramazonka-comprehendSee:  smart constructor.amazonka-comprehend6A list containing the properties of each job returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their names, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Filters the jobs that are returned. You can filter jobs on their names, status, or the date and time that they were submitted. You can only set one filter at a time.,  - The maximum number of results to return in each page. The default is 100., 1 - Identifies the next page of results to return.amazonka-comprehendFilters the jobs that are returned. You can filter jobs on their names, status, or the date and time that they were submitted. You can only set one filter at a time.amazonka-comprehendThe maximum number of results to return in each page. The default is 100.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 9 - A list containing the properties of each job returned., 1 - Identifies the next page of results to return., # - The response's http status code.amazonka-comprehend6A list containing the properties of each job returned.amazonka-comprehend.Identifies the next page of results to return.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the custom model being imported.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that allows Amazon Comprehend to use Amazon Key Management Service (KMS) to encrypt or decrypt the custom model.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendThe name to assign to the custom model that is created in Amazon Comprehend by this import.amazonka-comprehendTags to be associated with the custom model that is created by this import. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendThe version name given to the custom model that is created by this import. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/AWS Region.amazonka-comprehend=The Amazon Resource Name (ARN) of the custom model to import.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that allows Amazon Comprehend to use Amazon Key Management Service (KMS) to encrypt or decrypt the custom model.,  - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab",  - The name to assign to the custom model that is created in Amazon Comprehend by this import.,  - Tags to be associated with the custom model that is created by this import. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.,  - The version name given to the custom model that is created by this import. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/AWS Region.,  - The Amazon Resource Name (ARN) of the custom model to import.amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that allows Amazon Comprehend to use Amazon Key Management Service (KMS) to encrypt or decrypt the custom model.amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"amazonka-comprehendThe name to assign to the custom model that is created in Amazon Comprehend by this import.amazonka-comprehendTags to be associated with the custom model that is created by this import. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.amazonka-comprehendThe version name given to the custom model that is created by this import. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/AWS Region.amazonka-comprehend=The Amazon Resource Name (ARN) of the custom model to import.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the custom model being imported., # - The response's http status code.amazonka-comprehendThe Amazon Resource Name (ARN) of the custom model being imported.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehendTargeted sentiment analysis for each of the entities identified in the input text.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend7A UTF-8 text string. The maximum string length is 5 KB.amazonka-comprehendThe language of the input documents. Currently, English is the only supported language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, : - A UTF-8 text string. The maximum string length is 5 KB.,  - The language of the input documents. Currently, English is the only supported language.amazonka-comprehend7A UTF-8 text string. The maximum string length is 5 KB.amazonka-comprehendThe language of the input documents. Currently, English is the only supported language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Targeted sentiment analysis for each of the entities identified in the input text., # - The response's http status code.amazonka-comprehendTargeted sentiment analysis for each of the entities identified in the input text.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehendA collection of syntax tokens describing the text. For each token, the response provides the text, the token type, where the text begins and ends, and the level of confidence that Amazon Comprehend has that the token is correct. For a list of token types, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-syntax.htmlSyntax$ in the Comprehend Developer Guide.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend0A UTF-8 string. The maximum string size is 5 KB.amazonka-comprehendThe language code of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt").amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 3 - A UTF-8 string. The maximum string size is 5 KB.,  - The language code of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt").amazonka-comprehend0A UTF-8 string. The maximum string size is 5 KB.amazonka-comprehendThe language code of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt").amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A collection of syntax tokens describing the text. For each token, the response provides the text, the token type, where the text begins and ends, and the level of confidence that Amazon Comprehend has that the token is correct. For a list of token types, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-syntax.htmlSyntax$ in the Comprehend Developer Guide., # - The response's http status code.amazonka-comprehendA collection of syntax tokens describing the text. For each token, the response provides the text, the token type, where the text begins and ends, and the level of confidence that Amazon Comprehend has that the token is correct. For a list of token types, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-syntax.htmlSyntax$ in the Comprehend Developer Guide.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Zamazonka-comprehendSee:  smart constructor.amazonka-comprehendThe inferred sentiment that Amazon Comprehend has the highest level of confidence in.amazonka-comprehendAn object that lists the sentiments, and their corresponding confidence levels.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend5A UTF-8 text string. The maximum string size is 5 KB.Amazon Comprehend performs real-time sentiment analysis on the first 500 characters of the input text and ignores any additional text in the input.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 8 - A UTF-8 text string. The maximum string size is 5 KB.Amazon Comprehend performs real-time sentiment analysis on the first 500 characters of the input text and ignores any additional text in the input.,  - The language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehend5A UTF-8 text string. The maximum string size is 5 KB.Amazon Comprehend performs real-time sentiment analysis on the first 500 characters of the input text and ignores any additional text in the input.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The inferred sentiment that Amazon Comprehend has the highest level of confidence in.,  - An object that lists the sentiments, and their corresponding confidence levels., # - The response's http status code.amazonka-comprehendThe inferred sentiment that Amazon Comprehend has the highest level of confidence in.amazonka-comprehendAn object that lists the sentiments, and their corresponding confidence levels.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehendA collection of PII entities identified in the input text. For each entity, the response provides the entity type, where the entity text begins and ends, and the level of confidence that Amazon Comprehend has in the detection.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend7A UTF-8 text string. The maximum string size is 100 KB.amazonka-comprehendThe language of the input documents. Currently, English is the only valid language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, : - A UTF-8 text string. The maximum string size is 100 KB.,  - The language of the input documents. Currently, English is the only valid language.amazonka-comprehend7A UTF-8 text string. The maximum string size is 100 KB.amazonka-comprehendThe language of the input documents. Currently, English is the only valid language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A collection of PII entities identified in the input text. For each entity, the response provides the entity type, where the entity text begins and ends, and the level of confidence that Amazon Comprehend has in the detection., # - The response's http status code.amazonka-comprehendA collection of PII entities identified in the input text. For each entity, the response provides the entity type, where the entity text begins and ends, and the level of confidence that Amazon Comprehend has in the detection.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';&d amazonka-comprehendSee:  smart constructor.amazonka-comprehendA collection of key phrases that Amazon Comprehend identified in the input text. For each key phrase, the response provides the text of the key phrase, where the key phrase begins and ends, and the level of confidence that Amazon Comprehend has in the accuracy of the detection.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA UTF-8 text string. The string must contain less than 100 KB of UTF-8 encoded characters.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A UTF-8 text string. The string must contain less than 100 KB of UTF-8 encoded characters.,  - The language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendA UTF-8 text string. The string must contain less than 100 KB of UTF-8 encoded characters.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A collection of key phrases that Amazon Comprehend identified in the input text. For each key phrase, the response provides the text of the key phrase, where the key phrase begins and ends, and the level of confidence that Amazon Comprehend has in the accuracy of the detection., # - The response's http status code.amazonka-comprehendA collection of key phrases that Amazon Comprehend identified in the input text. For each key phrase, the response provides the text of the key phrase, where the key phrase begins and ends, and the level of confidence that Amazon Comprehend has in the accuracy of the detection.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehendamazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';[Qamazonka-comprehendSee:  smart constructor.amazonka-comprehendInformation about each block of text in the input document. Blocks are nested. A page block contains a block for each line of text, which contains a block for each word.The Block7 content for a Word input document does not include a Geometry field.The Block< field is not present in the response for plain-text inputs.amazonka-comprehendInformation about the document, discovered during text extraction. This field is present in the response only if your request used the Byte parameter.amazonka-comprehendThe document type for each page in the input document. This field is present in the response only if your request used the Byte parameter.amazonka-comprehendA collection of entities identified in the input text. For each entity, the response provides the entity text, entity type, where the entity text begins and ends, and the level of confidence that Amazon Comprehend has in the detection.If your request uses a custom entity recognition model, Amazon Comprehend detects the entities that the model is trained to recognize. Otherwise, it detects the default entity types. For a list of default entity types, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-entities.htmlEntities$ in the Comprehend Developer Guide.amazonka-comprehendPage-level errors that the system detected while processing the input document. The field is empty if the system encountered no errors.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThis field applies only when you use a custom entity recognition model that was trained with PDF annotations. For other cases, enter your text input in the Text field.Use the Bytes parameter to input a text, PDF, Word or image file. Using a plain-text file in the Bytes' parameter is equivelent to using the Text parameter (the Entities& field in the response is identical).You can also use the Bytes( parameter to input an Amazon Textract DetectDocumentText or AnalyzeDocument output file.Provide the input document as a sequence of base64-encoded bytes. If your code uses an Amazon Web Services SDK to detect entities, the SDK may encode the document file bytes for you.The maximum length of this field depends on the input document type. For details, see  https://docs.aws.amazon.com/comprehend/latest/dg/idp-inputs-sync.html$Inputs for real-time custom analysis$ in the Comprehend Developer Guide.If you use the Bytes parameter, do not use the Text parameter.amazonka-comprehendProvides configuration parameters to override the default actions for extracting text from PDF documents and image files.amazonka-comprehendThe Amazon Resource Name of an endpoint that is associated with a custom entity recognition model. Provide an endpoint if you want to detect entities by using your own custom model instead of the default model that is used by Amazon Comprehend.If you specify an endpoint, Amazon Comprehend uses the language of your custom model, and it ignores any language code that you provide in your request.&For information about endpoints, see  https://docs.aws.amazon.com/comprehend/latest/dg/manage-endpoints.htmlManaging endpoints.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. If your request includes the endpoint for a custom entity recognition model, Amazon Comprehend uses the language of your custom model, and it ignores any language code that you specify here.1All input documents must be in the same language.amazonka-comprehendA UTF-8 text string. The maximum string size is 100 KB. If you enter text using this parameter, do not use the Bytes parameter.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - This field applies only when you use a custom entity recognition model that was trained with PDF annotations. For other cases, enter your text input in the Text field.Use the Bytes parameter to input a text, PDF, Word or image file. Using a plain-text file in the Bytes' parameter is equivelent to using the Text parameter (the Entities& field in the response is identical).You can also use the Bytes( parameter to input an Amazon Textract DetectDocumentText or AnalyzeDocument output file.Provide the input document as a sequence of base64-encoded bytes. If your code uses an Amazon Web Services SDK to detect entities, the SDK may encode the document file bytes for you.The maximum length of this field depends on the input document type. For details, see  https://docs.aws.amazon.com/comprehend/latest/dg/idp-inputs-sync.html$Inputs for real-time custom analysis$ in the Comprehend Developer Guide.If you use the Bytes parameter, do not use the Text parameter.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.,  - Provides configuration parameters to override the default actions for extracting text from PDF documents and image files.,  - The Amazon Resource Name of an endpoint that is associated with a custom entity recognition model. Provide an endpoint if you want to detect entities by using your own custom model instead of the default model that is used by Amazon Comprehend.If you specify an endpoint, Amazon Comprehend uses the language of your custom model, and it ignores any language code that you provide in your request.&For information about endpoints, see  https://docs.aws.amazon.com/comprehend/latest/dg/manage-endpoints.htmlManaging endpoints.,  - The language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. If your request includes the endpoint for a custom entity recognition model, Amazon Comprehend uses the language of your custom model, and it ignores any language code that you specify here.1All input documents must be in the same language.,  - A UTF-8 text string. The maximum string size is 100 KB. If you enter text using this parameter, do not use the Bytes parameter.amazonka-comprehendThis field applies only when you use a custom entity recognition model that was trained with PDF annotations. For other cases, enter your text input in the Text field.Use the Bytes parameter to input a text, PDF, Word or image file. Using a plain-text file in the Bytes' parameter is equivelent to using the Text parameter (the Entities& field in the response is identical).You can also use the Bytes( parameter to input an Amazon Textract DetectDocumentText or AnalyzeDocument output file.Provide the input document as a sequence of base64-encoded bytes. If your code uses an Amazon Web Services SDK to detect entities, the SDK may encode the document file bytes for you.The maximum length of this field depends on the input document type. For details, see  https://docs.aws.amazon.com/comprehend/latest/dg/idp-inputs-sync.html$Inputs for real-time custom analysis$ in the Comprehend Developer Guide.If you use the Bytes parameter, do not use the Text parameter.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.amazonka-comprehendProvides configuration parameters to override the default actions for extracting text from PDF documents and image files.amazonka-comprehendThe Amazon Resource Name of an endpoint that is associated with a custom entity recognition model. Provide an endpoint if you want to detect entities by using your own custom model instead of the default model that is used by Amazon Comprehend.If you specify an endpoint, Amazon Comprehend uses the language of your custom model, and it ignores any language code that you provide in your request.&For information about endpoints, see  https://docs.aws.amazon.com/comprehend/latest/dg/manage-endpoints.htmlManaging endpoints.amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. If your request includes the endpoint for a custom entity recognition model, Amazon Comprehend uses the language of your custom model, and it ignores any language code that you specify here.1All input documents must be in the same language.amazonka-comprehendA UTF-8 text string. The maximum string size is 100 KB. If you enter text using this parameter, do not use the Bytes parameter.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Information about each block of text in the input document. Blocks are nested. A page block contains a block for each line of text, which contains a block for each word.The Block7 content for a Word input document does not include a Geometry field.The Block< field is not present in the response for plain-text inputs.,  - Information about the document, discovered during text extraction. This field is present in the response only if your request used the Byte parameter.,  - The document type for each page in the input document. This field is present in the response only if your request used the Byte parameter.,  - A collection of entities identified in the input text. For each entity, the response provides the entity text, entity type, where the entity text begins and ends, and the level of confidence that Amazon Comprehend has in the detection.If your request uses a custom entity recognition model, Amazon Comprehend detects the entities that the model is trained to recognize. Otherwise, it detects the default entity types. For a list of default entity types, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-entities.htmlEntities$ in the Comprehend Developer Guide.,  - Page-level errors that the system detected while processing the input document. The field is empty if the system encountered no errors., # - The response's http status code.amazonka-comprehendInformation about each block of text in the input document. Blocks are nested. A page block contains a block for each line of text, which contains a block for each word.The Block7 content for a Word input document does not include a Geometry field.The Block< field is not present in the response for plain-text inputs.amazonka-comprehendInformation about the document, discovered during text extraction. This field is present in the response only if your request used the Byte parameter.amazonka-comprehendThe document type for each page in the input document. This field is present in the response only if your request used the Byte parameter.amazonka-comprehendA collection of entities identified in the input text. For each entity, the response provides the entity text, entity type, where the entity text begins and ends, and the level of confidence that Amazon Comprehend has in the detection.If your request uses a custom entity recognition model, Amazon Comprehend detects the entities that the model is trained to recognize. Otherwise, it detects the default entity types. For a list of default entity types, see  https://docs.aws.amazon.com/comprehend/latest/dg/how-entities.htmlEntities$ in the Comprehend Developer Guide.amazonka-comprehendPage-level errors that the system detected while processing the input document. The field is empty if the system encountered no errors.amazonka-comprehend The response's http status code.amazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';f amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe languages that Amazon Comprehend detected in the input text. For each language, the response returns the RFC 5646 language code and the level of confidence that Amazon Comprehend has in the accuracy of its inference. For more information about RFC 5646, see  #https://tools.ietf.org/html/rfc5646Tags for Identifying Languages on the  IETF Tools web site.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendA UTF-8 text string. The string must contain at least 20 characters. The maximum string size is 100 KB.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - A UTF-8 text string. The string must contain at least 20 characters. The maximum string size is 100 KB.amazonka-comprehendA UTF-8 text string. The string must contain at least 20 characters. The maximum string size is 100 KB.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The languages that Amazon Comprehend detected in the input text. For each language, the response returns the RFC 5646 language code and the level of confidence that Amazon Comprehend has in the accuracy of its inference. For more information about RFC 5646, see  #https://tools.ietf.org/html/rfc5646Tags for Identifying Languages on the  IETF Tools web site., # - The response's http status code.amazonka-comprehendThe languages that Amazon Comprehend detected in the input text. For each language, the response returns the RFC 5646 language code and the level of confidence that Amazon Comprehend has in the accuracy of its inference. For more information about RFC 5646, see  #https://tools.ietf.org/html/rfc5646Tags for Identifying Languages on the  IETF Tools web site.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';m amazonka-comprehendSee:  smart constructor.amazonka-comprehend-The list of properties for the requested job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend9The identifier assigned by the user to the detection job.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, < - The identifier assigned by the user to the detection job.amazonka-comprehend9The identifier assigned by the user to the detection job.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, 0 - The list of properties for the requested job., # - The response's http status code.amazonka-comprehend-The list of properties for the requested job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';v+ amazonka-comprehendSee:  smart constructor.amazonka-comprehendAn object that contains the properties associated with a targeted sentiment detection job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - An object that contains the properties associated with a targeted sentiment detection job., # - The response's http status code.amazonka-comprehendAn object that contains the properties associated with a targeted sentiment detection job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';~H amazonka-comprehendSee:  smart constructor.amazonka-comprehendAn object that contains the properties associated with a sentiment detection job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - An object that contains the properties associated with a sentiment detection job., # - The response's http status code.amazonka-comprehendAn object that contains the properties associated with a sentiment detection job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';_amazonka-comprehendSee:  smart constructor.amazonka-comprehend)The time at which the policy was created.amazonka-comprehend/The time at which the policy was last modified.amazonka-comprehendThe revision ID of the policy. Each time you modify a policy, Amazon Comprehend assigns a new revision ID, and it deletes the prior version of the policy.amazonka-comprehend+The JSON body of the resource-based policy.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) of the custom model version that has the resource policy.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) of the custom model version that has the resource policy.amazonka-comprehendThe Amazon Resource Name (ARN) of the custom model version that has the resource policy.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, , - The time at which the policy was created., 2 - The time at which the policy was last modified.,  - The revision ID of the policy. Each time you modify a policy, Amazon Comprehend assigns a new revision ID, and it deletes the prior version of the policy., . - The JSON body of the resource-based policy., # - The response's http status code.amazonka-comprehend)The time at which the policy was created.amazonka-comprehend/The time at which the policy was last modified.amazonka-comprehendThe revision ID of the policy. Each time you modify a policy, Amazon Comprehend assigns a new revision ID, and it deletes the prior version of the policy.amazonka-comprehend+The JSON body of the resource-based policy.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - Undocumented member., # - The response's http status code.amazonka-comprehendUndocumented member.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehendAn object that contains the properties associated with a key phrases detection job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - An object that contains the properties associated with a key phrases detection job., # - The response's http status code.amazonka-comprehendAn object that contains the properties associated with a key phrases detection job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehendAn object that contains the properties associated with an event detection job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehend+The identifier of the events detection job.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, . - The identifier of the events detection job.amazonka-comprehend+The identifier of the events detection job.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - An object that contains the properties associated with an event detection job., # - The response's http status code.amazonka-comprehendAn object that contains the properties associated with an event detection job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';I amazonka-comprehendSee:  smart constructor.amazonka-comprehend;Describes information associated with an entity recognizer.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The Amazon Resource Name (ARN) that identifies the entity recognizer.amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:, > - Describes information associated with an entity recognizer., # - The response's http status code.amazonka-comprehend;Describes information associated with an entity recognizer.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';f amazonka-comprehendSee:  smart constructor.amazonka-comprehendAn object that contains the properties associated with an entities detection job.amazonka-comprehend The response's http status code.amazonka-comprehendSee:  smart constructor.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response.amazonka-comprehendCreate a value of " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:,  - An object that contains the properties associated with an entities detection job., # - The response's http status code.amazonka-comprehendAn object that contains the properties associated with an entities detection job.amazonka-comprehend The response's http status code.amazonka-comprehendamazonka-comprehend  (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:  smart constructor.amazonka-comprehend amazonka-comprehendSee:   smart constructor. amazonka-comprehendAn object that describes the properties associated with the document classification job. amazonka-comprehend The response's http status code. amazonka-comprehendSee:   smart constructor. amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response. amazonka-comprehendThe identifier that Amazon Comprehend generated for the job. The operation returns this identifier in its response. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - An object that describes the properties associated with the document classification job. ,  # - The response's http status code. amazonka-comprehendAn object that describes the properties associated with the document classification job. amazonka-comprehend The response's http status code. amazonka-comprehend amazonka-comprehend (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';֌ amazonka-comprehendSee:   smart constructor. amazonka-comprehend The response's http status code. amazonka-comprehendSee:   smart constructor. amazonka-comprehend(The revision ID of the policy to delete. amazonka-comprehendThe Amazon Resource Name (ARN) of the custom model version that has the policy to delete. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  + - The revision ID of the policy to delete. ,   - The Amazon Resource Name (ARN) of the custom model version that has the policy to delete. amazonka-comprehend(The revision ID of the policy to delete. amazonka-comprehendThe Amazon Resource Name (ARN) of the custom model version that has the policy to delete. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  # - The response's http status code. amazonka-comprehend The response's http status code. amazonka-comprehend amazonka-comprehend (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:   smart constructor. amazonka-comprehend The response's http status code. amazonka-comprehendSee:   smart constructor. amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Name (ARN) that identifies the entity recognizer. amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  # - The response's http status code. amazonka-comprehend The response's http status code. amazonka-comprehend amazonka-comprehend (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; amazonka-comprehendSee:   smart constructor. amazonka-comprehend The response's http status code. amazonka-comprehendSee:   smart constructor. amazonka-comprehend?The Amazon Resource Number (ARN) of the endpoint being deleted. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,   - The Amazon Resource Number (ARN) of the endpoint being deleted. amazonka-comprehend?The Amazon Resource Number (ARN) of the endpoint being deleted. amazonka-comprehendCreate a value of  " with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility: ,  # - The response's http status code. amazonka-comprehend The response's http status code. amazonka-comprehend amazonka-comprehend (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';O!amazonka-comprehendSee: ! smart constructor.!amazonka-comprehend The response's http status code.!amazonka-comprehendSee: ! smart constructor.!amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier.!amazonka-comprehendCreate a value of !" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:!, ! - The Amazon Resource Name (ARN) that identifies the document classifier.!amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier.!amazonka-comprehendCreate a value of !" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:!, !# - The response's http status code.!amazonka-comprehend The response's http status code.!amazonka-comprehend!!amazonka-comprehend! !!!!!!!!!! !!!!!!!!!!(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';!amazonka-comprehendSee: ! smart constructor.!amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer.!amazonka-comprehend The response's http status code.!amazonka-comprehendSee: ! smart constructor.!amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.!amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"!amazonka-comprehendThe JSON resource-based policy to attach to your custom entity recognizer model. You can use this policy to allow another AWS account to import your custom model.Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'!amazonka-comprehendTags to be associated with the entity recognizer being created. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.!amazonka-comprehendThe version name given to the newly created recognizer. Version names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same recognizer name in the account/ AWS Region.!amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"!amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your custom entity recognizer. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.!amazonka-comprehendThe name given to the newly created recognizer. Recognizer names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The name must be unique in the account/region.!amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.!amazonka-comprehendSpecifies the format and location of the input data. The S3 bucket containing the input data must be located in the same region as the entity recognizer being created.!amazonka-comprehendYou can specify any of the following languages: English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), German ("de"), or Portuguese ("pt"). If you plan to use this entity recognizer with PDF, Word, or image input files, you must specify English as the language. All training documents must be in the same language.!amazonka-comprehendCreate a value of !" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:!, ! - A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.!, ! - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"!, ! - The JSON resource-based policy to attach to your custom entity recognizer model. You can use this policy to allow another AWS account to import your custom model.Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'!, ! - Tags to be associated with the entity recognizer being created. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.!, ! - The version name given to the newly created recognizer. Version names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same recognizer name in the account/ AWS Region.!, ! - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"!, ! - Configuration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your custom entity recognizer. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.!, ! - The name given to the newly created recognizer. Recognizer names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The name must be unique in the account/region.!, ! - The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.!, ! - Specifies the format and location of the input data. The S3 bucket containing the input data must be located in the same region as the entity recognizer being created.!, ! - You can specify any of the following languages: English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), German ("de"), or Portuguese ("pt"). If you plan to use this entity recognizer with PDF, Word, or image input files, you must specify English as the language. All training documents must be in the same language.!amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.!amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"!amazonka-comprehendThe JSON resource-based policy to attach to your custom entity recognizer model. You can use this policy to allow another AWS account to import your custom model.Provide your JSON as a UTF-8 encoded string without line breaks. To provide valid JSON for your policy, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'!amazonka-comprehendTags to be associated with the entity recognizer being created. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.!amazonka-comprehendThe version name given to the newly created recognizer. Version names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same recognizer name in the account/ AWS Region.!amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"!amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your custom entity recognizer. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.!amazonka-comprehendThe name given to the newly created recognizer. Recognizer names can be a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The name must be unique in the account/region.!amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.!amazonka-comprehendSpecifies the format and location of the input data. The S3 bucket containing the input data must be located in the same region as the entity recognizer being created.!amazonka-comprehendYou can specify any of the following languages: English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), German ("de"), or Portuguese ("pt"). If you plan to use this entity recognizer with PDF, Word, or image input files, you must specify English as the language. All training documents must be in the same language.!amazonka-comprehendCreate a value of !" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:!, ! - The Amazon Resource Name (ARN) that identifies the entity recognizer.!, !# - The response's http status code.!amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the entity recognizer.!amazonka-comprehend The response's http status code.!amazonka-comprehend!amazonka-comprehend!amazonka-comprehend!amazonka-comprehend!!amazonka-comprehend! !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!! !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';.!amazonka-comprehendSee: ! smart constructor.!amazonka-comprehend?The Amazon Resource Number (ARN) of the endpoint being created.!amazonka-comprehend The response's http status code.!amazonka-comprehendSee: ! smart constructor.!amazonka-comprehendAn idempotency token provided by the customer. If this token matches a previous endpoint creation request, Amazon Comprehend will not return a ResourceInUseException.!amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).!amazonka-comprehendTags associated with the endpoint being created. A tag is a key-value pair that adds metadata to the endpoint. For example, a tag with "Sales" as the key might be added to an endpoint to indicate its use by the sales department.!amazonka-comprehend8This is the descriptive suffix that becomes part of the  EndpointArn4 used for all subsequent requests to this resource.!amazonka-comprehendThe Amazon Resource Number (ARN) of the model to which the endpoint will be attached.!amazonka-comprehendThe desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.!amazonka-comprehendCreate a value of !" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:!, ! - An idempotency token provided by the customer. If this token matches a previous endpoint creation request, Amazon Comprehend will not return a ResourceInUseException.!, ! - The Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).!, ! - Tags associated with the endpoint being created. A tag is a key-value pair that adds metadata to the endpoint. For example, a tag with "Sales" as the key might be added to an endpoint to indicate its use by the sales department.!, !; - This is the descriptive suffix that becomes part of the  EndpointArn4 used for all subsequent requests to this resource.!, ! - The Amazon Resource Number (ARN) of the model to which the endpoint will be attached.!, ! - The desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.!amazonka-comprehendAn idempotency token provided by the customer. If this token matches a previous endpoint creation request, Amazon Comprehend will not return a ResourceInUseException.!amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS identity and Access Management (IAM) role that grants Amazon Comprehend read access to trained custom models encrypted with a customer managed key (ModelKmsKeyId).!amazonka-comprehendTags associated with the endpoint being created. A tag is a key-value pair that adds metadata to the endpoint. For example, a tag with "Sales" as the key might be added to an endpoint to indicate its use by the sales department.!amazonka-comprehend8This is the descriptive suffix that becomes part of the  EndpointArn4 used for all subsequent requests to this resource.!amazonka-comprehendThe Amazon Resource Number (ARN) of the model to which the endpoint will be attached.!amazonka-comprehendThe desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.!amazonka-comprehendCreate a value of !" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:!, ! - The Amazon Resource Number (ARN) of the endpoint being created.!, !# - The response's http status code.!amazonka-comprehend?The Amazon Resource Number (ARN) of the endpoint being created.!amazonka-comprehend The response's http status code.!amazonka-comprehend!amazonka-comprehend!amazonka-comprehend!!amazonka-comprehend!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';a "!amazonka-comprehendSee: " smart constructor.!amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier."amazonka-comprehend The response's http status code."amazonka-comprehendSee: " smart constructor."amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one."amazonka-comprehendIndicates the mode in which the classifier will be trained. The classifier can be trained in multi-class mode, which identifies one and only one class for each document, or multi-label mode, which identifies one or more labels for each document. In multi-label mode, multiple labels for an individual document are separated by a delimiter. The default delimiter between labels is a pipe (|)."amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab""amazonka-comprehendThe resource-based policy to attach to your custom document classifier model. You can use this policy to allow another AWS account to import your custom model.Provide your policy as a JSON body that you enter as a UTF-8 encoded string without line breaks. To provide valid JSON, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'"amazonka-comprehendEnables the addition of output results configuration parameters for custom classifier jobs."amazonka-comprehendTags to be associated with the document classifier being created. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department."amazonka-comprehendThe version name given to the newly created classifier. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/AWS Region."amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab""amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your custom classifier. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC."amazonka-comprehend$The name of the document classifier."amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data."amazonka-comprehendSpecifies the format and location of the input data for the job."amazonka-comprehendThe language of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt"). All documents must be in the same language."amazonka-comprehendCreate a value of "" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:", " - A unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one.", " - Indicates the mode in which the classifier will be trained. The classifier can be trained in multi-class mode, which identifies one and only one class for each document, or multi-label mode, which identifies one or more labels for each document. In multi-label mode, multiple labels for an individual document are separated by a delimiter. The default delimiter between labels is a pipe (|).", " - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"", " - The resource-based policy to attach to your custom document classifier model. You can use this policy to allow another AWS account to import your custom model.Provide your policy as a JSON body that you enter as a UTF-8 encoded string without line breaks. To provide valid JSON, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'", " - Enables the addition of output results configuration parameters for custom classifier jobs.", " - Tags to be associated with the document classifier being created. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department.", " - The version name given to the newly created classifier. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/AWS Region.", " - ID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab"", " - Configuration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your custom classifier. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC.", "' - The name of the document classifier.", " - The Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data.", " - Specifies the format and location of the input data for the job.", " - The language of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt"). All documents must be in the same language."amazonka-comprehendA unique identifier for the request. If you don't set the client request token, Amazon Comprehend generates one."amazonka-comprehendIndicates the mode in which the classifier will be trained. The classifier can be trained in multi-class mode, which identifies one and only one class for each document, or multi-label mode, which identifies one or more labels for each document. In multi-label mode, multiple labels for an individual document are separated by a delimiter. The default delimiter between labels is a pipe (|)."amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt trained custom models. The ModelKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab""amazonka-comprehendThe resource-based policy to attach to your custom document classifier model. You can use this policy to allow another AWS account to import your custom model.Provide your policy as a JSON body that you enter as a UTF-8 encoded string without line breaks. To provide valid JSON, enclose the attribute names and values in double quotes. If the JSON body is also enclosed in double quotes, then you must escape the double quotes that are inside the policy: 8"{\"attribute\": \"value\", \"attribute\": [\"value\"]}"To avoid escaping quotes, you can use single quotes to enclose the policy and double quotes to enclose the JSON names and values: 0'{"attribute": "value", "attribute": ["value"]}'"amazonka-comprehendEnables the addition of output results configuration parameters for custom classifier jobs."amazonka-comprehendTags to be associated with the document classifier being created. A tag is a key-value pair that adds as a metadata to a resource used by Amazon Comprehend. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department."amazonka-comprehendThe version name given to the newly created classifier. Version names can have a maximum of 256 characters. Alphanumeric characters, hyphens (-) and underscores (_) are allowed. The version name must be unique among all models with the same classifier name in the account/AWS Region."amazonka-comprehendID for the AWS Key Management Service (KMS) key that Amazon Comprehend uses to encrypt data on the storage volume attached to the ML compute instance(s) that process the analysis job. The VolumeKmsKeyId can be either of the following formats: KMS Key ID: &"1234abcd-12ab-34cd-56ef-1234567890ab".Amazon Resource Name (ARN) of a KMS Key: "arn:aws:kms:us-west-2:111122223333:key/1234abcd-12ab-34cd-56ef-1234567890ab""amazonka-comprehendConfiguration parameters for an optional private Virtual Private Cloud (VPC) containing the resources you are using for your custom classifier. For more information, see  https://docs.aws.amazon.com/vpc/latest/userguide/what-is-amazon-vpc.html Amazon VPC."amazonka-comprehend$The name of the document classifier."amazonka-comprehendThe Amazon Resource Name (ARN) of the AWS Identity and Management (IAM) role that grants Amazon Comprehend read access to your input data."amazonka-comprehendSpecifies the format and location of the input data for the job."amazonka-comprehendThe language of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt"). All documents must be in the same language."amazonka-comprehendCreate a value of !" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:!, " - The Amazon Resource Name (ARN) that identifies the document classifier.", "# - The response's http status code."amazonka-comprehendThe Amazon Resource Name (ARN) that identifies the document classifier."amazonka-comprehend The response's http status code."amazonka-comprehend"amazonka-comprehend"amazonka-comprehend"amazonka-comprehend""amazonka-comprehend"$!!!"""""""""""""""""""""""""""""""""$"""""""""""""""""""""""""""""!!!""""(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';j "amazonka-comprehendSee: " smart constructor."amazonka-comprehendThe labels used in the document being analyzed. Individual labels represent personally identifiable information (PII) entity types."amazonka-comprehend The response's http status code."amazonka-comprehendSee: " smart constructor."amazonka-comprehend7A UTF-8 text string. The maximum string size is 100 KB."amazonka-comprehendThe language of the input documents. Currently, English is the only valid language."amazonka-comprehendCreate a value of "" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:", ": - A UTF-8 text string. The maximum string size is 100 KB.", " - The language of the input documents. Currently, English is the only valid language."amazonka-comprehend7A UTF-8 text string. The maximum string size is 100 KB."amazonka-comprehendThe language of the input documents. Currently, English is the only valid language."amazonka-comprehendCreate a value of "" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:", " - The labels used in the document being analyzed. Individual labels represent personally identifiable information (PII) entity types.", "# - The response's http status code."amazonka-comprehendThe labels used in the document being analyzed. Individual labels represent personally identifiable information (PII) entity types."amazonka-comprehend The response's http status code."amazonka-comprehend"amazonka-comprehend""amazonka-comprehend"""""""""""""""""""""""""""""(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';;"amazonka-comprehendSee: " smart constructor."amazonka-comprehendThe classes used by the document being analyzed. These are used for multi-class trained models. Individual classes are mutually exclusive and each document is expected to have only a single class assigned to it. For example, an animal can be a dog or a cat, but not both at the same time."amazonka-comprehendExtraction information about the document. This field is present in the response only if your request includes the Byte parameter."amazonka-comprehendThe document type for each page in the input document. This field is present in the response only if your request includes the Byte parameter."amazonka-comprehendPage-level errors that the system detected while processing the input document. The field is empty if the system encountered no errors."amazonka-comprehendThe labels used the document being analyzed. These are used for multi-label trained models. Individual labels represent different categories that are related in some manner and are not mutually exclusive. For example, a movie can be just an action movie, or it can be an action movie, a science fiction movie, and a comedy, all at the same time."amazonka-comprehend The response's http status code."amazonka-comprehendSee: " smart constructor."amazonka-comprehendUse the Bytes parameter to input a text, PDF, Word or image file. You can also use the Bytes( parameter to input an Amazon Textract DetectDocumentText or AnalyzeDocument output file.Provide the input document as a sequence of base64-encoded bytes. If your code uses an Amazon Web Services SDK to classify documents, the SDK may encode the document file bytes for you.The maximum length of this field depends on the input document type. For details, see  https://docs.aws.amazon.com/comprehend/latest/dg/idp-inputs-sync.html$Inputs for real-time custom analysis$ in the Comprehend Developer Guide.If you use the Bytes parameter, do not use the Text parameter."amazonka-comprehendProvides configuration parameters to override the default actions for extracting text from PDF documents and image files."amazonka-comprehendThe document text to be analyzed. If you enter text using this parameter, do not use the Bytes parameter."amazonka-comprehendThe Amazon Resource Number (ARN) of the endpoint. For information about endpoints, see  https://docs.aws.amazon.com/comprehend/latest/dg/manage-endpoints.htmlManaging endpoints."amazonka-comprehendCreate a value of "" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:", " - Use the Bytes parameter to input a text, PDF, Word or image file. You can also use the Bytes( parameter to input an Amazon Textract DetectDocumentText or AnalyzeDocument output file.Provide the input document as a sequence of base64-encoded bytes. If your code uses an Amazon Web Services SDK to classify documents, the SDK may encode the document file bytes for you.The maximum length of this field depends on the input document type. For details, see  https://docs.aws.amazon.com/comprehend/latest/dg/idp-inputs-sync.html$Inputs for real-time custom analysis$ in the Comprehend Developer Guide.If you use the Bytes parameter, do not use the Text parameter.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data.", " - Provides configuration parameters to override the default actions for extracting text from PDF documents and image files.", " - The document text to be analyzed. If you enter text using this parameter, do not use the Bytes parameter.", " - The Amazon Resource Number (ARN) of the endpoint. For information about endpoints, see  https://docs.aws.amazon.com/comprehend/latest/dg/manage-endpoints.htmlManaging endpoints."amazonka-comprehendUse the Bytes parameter to input a text, PDF, Word or image file. You can also use the Bytes( parameter to input an Amazon Textract DetectDocumentText or AnalyzeDocument output file.Provide the input document as a sequence of base64-encoded bytes. If your code uses an Amazon Web Services SDK to classify documents, the SDK may encode the document file bytes for you.The maximum length of this field depends on the input document type. For details, see  https://docs.aws.amazon.com/comprehend/latest/dg/idp-inputs-sync.html$Inputs for real-time custom analysis$ in the Comprehend Developer Guide.If you use the Bytes parameter, do not use the Text parameter.-- -- Note: This Lens automatically encodes and decodes Base64 data. -- The underlying isomorphism will encode to Base64 representation during -- serialisation, and decode from Base64 representation during deserialisation. -- This Lens- accepts and returns only raw unencoded data."amazonka-comprehendProvides configuration parameters to override the default actions for extracting text from PDF documents and image files."amazonka-comprehendThe document text to be analyzed. If you enter text using this parameter, do not use the Bytes parameter."amazonka-comprehendThe Amazon Resource Number (ARN) of the endpoint. For information about endpoints, see  https://docs.aws.amazon.com/comprehend/latest/dg/manage-endpoints.htmlManaging endpoints."amazonka-comprehendCreate a value of "" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:", " - The classes used by the document being analyzed. These are used for multi-class trained models. Individual classes are mutually exclusive and each document is expected to have only a single class assigned to it. For example, an animal can be a dog or a cat, but not both at the same time.", " - Extraction information about the document. This field is present in the response only if your request includes the Byte parameter.", " - The document type for each page in the input document. This field is present in the response only if your request includes the Byte parameter.", " - Page-level errors that the system detected while processing the input document. The field is empty if the system encountered no errors.", " - The labels used the document being analyzed. These are used for multi-label trained models. Individual labels represent different categories that are related in some manner and are not mutually exclusive. For example, a movie can be just an action movie, or it can be an action movie, a science fiction movie, and a comedy, all at the same time.", "# - The response's http status code."amazonka-comprehendThe classes used by the document being analyzed. These are used for multi-class trained models. Individual classes are mutually exclusive and each document is expected to have only a single class assigned to it. For example, an animal can be a dog or a cat, but not both at the same time."amazonka-comprehendExtraction information about the document. This field is present in the response only if your request includes the Byte parameter."amazonka-comprehendThe document type for each page in the input document. This field is present in the response only if your request includes the Byte parameter."amazonka-comprehendPage-level errors that the system detected while processing the input document. The field is empty if the system encountered no errors."amazonka-comprehendThe labels used the document being analyzed. These are used for multi-label trained models. Individual labels represent different categories that are related in some manner and are not mutually exclusive. For example, a movie can be just an action movie, or it can be an action movie, a science fiction movie, and a comedy, all at the same time."amazonka-comprehend The response's http status code."amazonka-comprehend""amazonka-comprehend"""""""""""""""""""""""""""""""""""""""""""""""""""""(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';g"amazonka-comprehendSee: # smart constructor."amazonka-comprehend The response's http status code."amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty."amazonka-comprehend-List of errors that the operation can return."amazonka-comprehendSee: # smart constructor."amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB."amazonka-comprehendThe language of the input documents. Currently, English is the only supported language.#amazonka-comprehendCreate a value of "" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:", # - A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.", # - The language of the input documents. Currently, English is the only supported language.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.#amazonka-comprehendThe language of the input documents. Currently, English is the only supported language.#amazonka-comprehendCreate a value of "" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:", ## - The response's http status code.", # - A list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.", #0 - List of errors that the operation can return.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehend-List of errors that the operation can return.#amazonka-comprehend"amazonka-comprehend"#amazonka-comprehend""""""""""#######""""###"""""####(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';#amazonka-comprehendSee: # smart constructor.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehendSee: # smart constructor.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size for each document is 5 KB.#amazonka-comprehendThe language of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt"). All documents must be in the same language.#amazonka-comprehendCreate a value of #" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:#, # - A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size for each document is 5 KB.#, # - The language of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt"). All documents must be in the same language.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size for each document is 5 KB.#amazonka-comprehendThe language of the input documents. You can specify any of the following languages supported by Amazon Comprehend: German ("de"), English ("en"), Spanish ("es"), French ("fr"), Italian ("it"), or Portuguese ("pt"). All documents must be in the same language.#amazonka-comprehendCreate a value of #" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:#, ## - The response's http status code.#, # - A list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#, # - A list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehend#amazonka-comprehend##amazonka-comprehend#################################(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';Y#amazonka-comprehendSee: # smart constructor.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehendSee: # smart constructor.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.Amazon Comprehend performs real-time sentiment analysis on the first 500 characters of the input text and ignores any additional text in the input.#amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendCreate a value of #" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:#, # - A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.Amazon Comprehend performs real-time sentiment analysis on the first 500 characters of the input text and ignores any additional text in the input.#, # - The language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.Amazon Comprehend performs real-time sentiment analysis on the first 500 characters of the input text and ignores any additional text in the input.#amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendCreate a value of #" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:#, ## - The response's http status code.#, # - A list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#, # - A list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehend#amazonka-comprehend##amazonka-comprehend#################################(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';#amazonka-comprehendSee: # smart constructor.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehendSee: # smart constructor.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.#amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendCreate a value of #" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:#, # - A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.#, # - The language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.#amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendCreate a value of #" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:#, ## - The response's http status code.#, # - A list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#, # - A list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehend#amazonka-comprehend##amazonka-comprehend#################################(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&';E#amazonka-comprehendSee: # smart constructor.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehendSee: # smart constructor.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.#amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendCreate a value of #" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:#, # - A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.#, # - The language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. The maximum size of each document is 5 KB.#amazonka-comprehendThe language of the input documents. You can specify any of the primary languages supported by Amazon Comprehend. All documents must be in the same language.#amazonka-comprehendCreate a value of #" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:#, ## - The response's http status code.#, # - A list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#, # - A list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehend The response's http status code.#amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.#amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.#amazonka-comprehend#amazonka-comprehend##amazonka-comprehend#################################(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; $amazonka-comprehendSee: $ smart constructor.$amazonka-comprehend The response's http status code.$amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.$amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.$amazonka-comprehendSee: $ smart constructor.$amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. Each document should contain at least 20 characters. The maximum size of each document is 5 KB.$amazonka-comprehendCreate a value of $" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:$, $ - A list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. Each document should contain at least 20 characters. The maximum size of each document is 5 KB.$amazonka-comprehendA list containing the UTF-8 encoded text of the input documents. The list can contain a maximum of 25 documents. Each document should contain at least 20 characters. The maximum size of each document is 5 KB.$amazonka-comprehendCreate a value of $" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:$, $# - The response's http status code.$, $ - A list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.$, $ - A list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.$amazonka-comprehend The response's http status code.$amazonka-comprehendA list of objects containing the results of the operation. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If all of the documents contain an error, the  ResultList is empty.$amazonka-comprehendA list containing one object for each document that contained an error. The results are sorted in ascending order by the Index field and match the order of the documents in the input list. If there are no errors in the batch, the  ErrorList is empty.$amazonka-comprehend$$amazonka-comprehend$$$$$$$$$$$$$$$$$$$$$$$$$$$$$(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; $amazonka-comprehendSee: $ smart constructor.$amazonka-comprehend The response's http status code.$amazonka-comprehendSee: $ smart constructor.$amazonka-comprehendThe Amazon Resource Name (ARN) of the given Amazon Comprehend resource from which you want to remove the tags.$amazonka-comprehendThe initial part of a key-value pair that forms a tag being removed from a given resource. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department. Keys must be unique and cannot be duplicated for a particular resource.$amazonka-comprehendCreate a value of $" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:$, $ - The Amazon Resource Name (ARN) of the given Amazon Comprehend resource from which you want to remove the tags.$, $ - The initial part of a key-value pair that forms a tag being removed from a given resource. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department. Keys must be unique and cannot be duplicated for a particular resource.$amazonka-comprehendThe Amazon Resource Name (ARN) of the given Amazon Comprehend resource from which you want to remove the tags.$amazonka-comprehendThe initial part of a key-value pair that forms a tag being removed from a given resource. For example, a tag with "Sales" as the key might be added to a resource to indicate its use by the sales department. Keys must be unique and cannot be duplicated for a particular resource.$amazonka-comprehendCreate a value of $" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:$, $# - The response's http status code.$amazonka-comprehend The response's http status code.$amazonka-comprehend$$amazonka-comprehend$ $$$$$$$$$$$$ $$$$$$$$$$$$(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred "%&'; L$amazonka-comprehendSee: $ smart constructor.$amazonka-comprehend The response's http status code.$amazonka-comprehendSee: $ smart constructor.$amazonka-comprehendData access role ARN to use in case the new model is encrypted with a customer CMK.$amazonka-comprehendThe desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.$amazonka-comprehendThe ARN of the new model to use when updating an existing endpoint.$amazonka-comprehend?The Amazon Resource Number (ARN) of the endpoint being updated.$amazonka-comprehendCreate a value of $" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:$, $ - Data access role ARN to use in case the new model is encrypted with a customer CMK.$, $ - The desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.$, $ - The ARN of the new model to use when updating an existing endpoint.$, $ - The Amazon Resource Number (ARN) of the endpoint being updated.$amazonka-comprehendData access role ARN to use in case the new model is encrypted with a customer CMK.$amazonka-comprehendThe desired number of inference units to be used by the model using this endpoint. Each inference unit represents of a throughput of 100 characters per second.$amazonka-comprehendThe ARN of the new model to use when updating an existing endpoint.$amazonka-comprehend?The Amazon Resource Number (ARN) of the endpoint being updated.$amazonka-comprehendCreate a value of $" with all optional fields omitted.Use  0https://hackage.haskell.org/package/generic-lens generic-lens or  *https://hackage.haskell.org/package/opticsoptics! to modify other optional fields.The following record fields are available, with the corresponding lenses provided for backwards compatibility:$, $# - The response's http status code.$amazonka-comprehend The response's http status code.$amazonka-comprehend$$amazonka-comprehend$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred & GHIJXYZijkl  !!!!!!!!!!!!!!!!!!!!!!!"""""""""""""""""""""""""""""#########################$$$$$$$$$$$$$$$$#########################"""""""""""""""""""""""""""""!!!!!!!!!!!!!!!!!!!!!!! $$$$$$$$   ijklGHIJXYZ       (c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred"% >(c) 2013-2023 Brendan HayMozilla Public License, v. 2.0. Brendan Hayauto-generatednon-portable (GHC extensions) Safe-Inferred (,+)*@AFRSWbchtu~  !!!!!!!!!!!!!!!!!!!!""""""""""""""""""""##########################$$$$$$$$$$$$$$$$$$$$$$$$########################""#""#"""""""""""""""!!"!!!!!!!!!!!!!!!!!! $$$$$$$$$$$$(,+)*,+    bch@AFRSWtu~       $                                                                                                               !!!!!!!!!!!!!!!!"""""""""""""""""#############$$$$$$$$$$$$$$%%%%%%%%%%%%%%&&&&&&&&&&&&&&&&&&&&&&&&'''''''''''''''''((((((((((((((((())))))))))))))))))))))))))))******************++++++++++++++++++,,,,,,,,,,,,,,,,,,------------------..................//////////////000000000000000000111111111111111111111111111111111122222222222222 2 2 2 2 2 2 2 2 2 2 2 2 2 2 2 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 3 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 4 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 5 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 7 7 7 7 7 7 7 7 7 7 7 7 7 7 7 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 8 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 9 : : : : : : : : : : : : : : : : ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; ; < < < < < < < < < < < < < < = = = = = = = = = = = = = = = = = = > > > > > > > > > > > > > > > > > > > > > > > > ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ @ A A A A A A A A A A A A A A A A A A B B B B B B B B B B B B B B C C C C C C C C C C C C C C D D D D D D D D D D D D D D E E E E E E E E E E E E E E F F F F F F F F F F F F F F F F F G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G G H H H H H H H H H H H H H H H H H H H H H H H I I I I I I I I I I I I I I J J J J J J J J J J J J J J J J J J J J J J K K K K K K K K K K K K K K K K K K L L L L L L L L L L L L L L L L L L M M M M M M M M M M M M M M M M M M M M M M M M M M N N N N N N N N N N N N N N O O O O O O O O O O O O O O O OPPPPPPPPPPPPPPPPPPPPPPPPQQQQQQQQQQQQQQQQQQQQQQQRRRRRRRRRRRRRRRRRRRRRRRSSSSSSSSSSSSSSSSSSSSSTTTTTTTTTTTTTTTTTTTTTTTTTTTTUUUUUUUUUUUUUUUUUUUUVVVVVVVVVVVVVVWWWWWWWWWWWWWWWXXXXXXXXXXXXXXXXXXYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYZZZZZZZZZZZZZZZZZZZZZZZZ[[[[[[[[[[[[[[\\\\\\\\\\\\\\]]]]]]]]]]]]]]]]]]^^^^^^^^^^^^^^^____________________________________````````````````````````````````````aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaabbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbcccccccccccccccccccccccccccccccccccccccccddddddddddddddddddddddddddddddddddddddeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeeefffffffffffffffffffffffffffffffffffffffffffffgggggggggggggggggggggggggggggggggggghhhhhhhhhhhhhhhhhiiiiiiiiiiiiiiiiiiiiiiiiiiiijjjjjjjjjjjjjjjjjjjjjjjjjjkkkkkkkkkkkkkkkkkkkkkkkkkkllllllllllllllllllllllllllllllmmmmmmmmmmmmmmmmmmmmmmmmmmmmmmnnnnnnnnnnnnnnnnnnnnnnnnnnnnnnooooooooooooooooooooooooooooooppppppppppppppppppppppppppppppqqqqqqqqqqqqqqqqqqqqqqqqqqqqqqrrrrrrrrrrrrrrrrrrrrrrrrrrrrrrssssssssssssssssssssssssssssssssssssssssssssssssttttttttttttttttttttttttttttttttttttttttttttttttuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuuvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvvwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwwxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyyzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzzz{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{{||||||||||||||||||||||||||||||||}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}}~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~                                                                                                                                !!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!!""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""""################################################################################################################################$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%%&&&&&&&&&&&&&&&&&&&&&&&&&&&.amazonka-comprehend-2.0-Ko6GCjAQF2RARapSdPn69F>Amazonka.Comprehend.Types.AugmentedManifestsDocumentTypeFormat(Amazonka.Comprehend.Types.BatchItemError#Amazonka.Comprehend.Types.BlockType%Amazonka.Comprehend.Types.BoundingBox$Amazonka.Comprehend.Types.ChildBlock(Amazonka.Comprehend.Types.BlockReference5Amazonka.Comprehend.Types.ClassifierEvaluationMetrics,Amazonka.Comprehend.Types.ClassifierMetadata'Amazonka.Comprehend.Types.DocumentClass6Amazonka.Comprehend.Types.DocumentClassifierDataFormat0Amazonka.Comprehend.Types.DocumentClassifierMode$sel:outputDataConfig:TargetedSentimentDetectionJobProperties'8$sel:submitTime:TargetedSentimentDetectionJobProperties'<$sel:volumeKmsKeyId:TargetedSentimentDetectionJobProperties'7$sel:vpcConfig:TargetedSentimentDetectionJobProperties'*newTargetedSentimentDetectionJobProperties9targetedSentimentDetectionJobProperties_dataAccessRoleArn/targetedSentimentDetectionJobProperties_endTime7targetedSentimentDetectionJobProperties_inputDataConfig.targetedSentimentDetectionJobProperties_jobArn-targetedSentimentDetectionJobProperties_jobId/targetedSentimentDetectionJobProperties_jobName1targetedSentimentDetectionJobProperties_jobStatus4targetedSentimentDetectionJobProperties_languageCode/targetedSentimentDetectionJobProperties_message8targetedSentimentDetectionJobProperties_outputDataConfig2targetedSentimentDetectionJobProperties_submitTime6targetedSentimentDetectionJobProperties_volumeKmsKeyId1targetedSentimentDetectionJobProperties_vpcConfig/$fNFDataTargetedSentimentDetectionJobProperties1$fHashableTargetedSentimentDetectionJobProperties1$fFromJSONTargetedSentimentDetectionJobProperties+$fEqTargetedSentimentDetectionJobProperties-$fReadTargetedSentimentDetectionJobProperties-$fShowTargetedSentimentDetectionJobProperties0$fGenericTargetedSentimentDetectionJobPropertiesSentimentDetectionJobProperties SentimentDetectionJobProperties'7$sel:dataAccessRoleArn:SentimentDetectionJobProperties'-$sel:endTime:SentimentDetectionJobProperties'5$sel:inputDataConfig:SentimentDetectionJobProperties',$sel:jobArn:SentimentDetectionJobProperties'+$sel:jobId:SentimentDetectionJobProperties'-$sel:jobName:SentimentDetectionJobProperties'/$sel:jobStatus:SentimentDetectionJobProperties'2$sel:languageCode:SentimentDetectionJobProperties'-$sel:message:SentimentDetectionJobProperties'6$sel:outputDataConfig:SentimentDetectionJobProperties'0$sel:submitTime:SentimentDetectionJobProperties'4$sel:volumeKmsKeyId:SentimentDetectionJobProperties'/$sel:vpcConfig:SentimentDetectionJobProperties'"newSentimentDetectionJobProperties1sentimentDetectionJobProperties_dataAccessRoleArn'sentimentDetectionJobProperties_endTime/sentimentDetectionJobProperties_inputDataConfig&sentimentDetectionJobProperties_jobArn%sentimentDetectionJobProperties_jobId'sentimentDetectionJobProperties_jobName)sentimentDetectionJobProperties_jobStatus,sentimentDetectionJobProperties_languageCode'sentimentDetectionJobProperties_message0sentimentDetectionJobProperties_outputDataConfig*sentimentDetectionJobProperties_submitTime.sentimentDetectionJobProperties_volumeKmsKeyId)sentimentDetectionJobProperties_vpcConfig'$fNFDataSentimentDetectionJobProperties)$fHashableSentimentDetectionJobProperties)$fFromJSONSentimentDetectionJobProperties#$fEqSentimentDetectionJobProperties%$fReadSentimentDetectionJobProperties%$fShowSentimentDetectionJobProperties($fGenericSentimentDetectionJobProperties KeyPhrasesDetectionJobProperties!KeyPhrasesDetectionJobProperties'8$sel:dataAccessRoleArn:KeyPhrasesDetectionJobProperties'.$sel:endTime:KeyPhrasesDetectionJobProperties'6$sel:inputDataConfig:KeyPhrasesDetectionJobProperties'-$sel:jobArn:KeyPhrasesDetectionJobProperties',$sel:jobId:KeyPhrasesDetectionJobProperties'.$sel:jobName:KeyPhrasesDetectionJobProperties'0$sel:jobStatus:KeyPhrasesDetectionJobProperties'3$sel:languageCode:KeyPhrasesDetectionJobProperties'.$sel:message:KeyPhrasesDetectionJobProperties'7$sel:outputDataConfig:KeyPhrasesDetectionJobProperties'1$sel:submitTime:KeyPhrasesDetectionJobProperties'5$sel:volumeKmsKeyId:KeyPhrasesDetectionJobProperties'0$sel:vpcConfig:KeyPhrasesDetectionJobProperties'#newKeyPhrasesDetectionJobProperties2keyPhrasesDetectionJobProperties_dataAccessRoleArn(keyPhrasesDetectionJobProperties_endTime0keyPhrasesDetectionJobProperties_inputDataConfig'keyPhrasesDetectionJobProperties_jobArn&keyPhrasesDetectionJobProperties_jobId(keyPhrasesDetectionJobProperties_jobName*keyPhrasesDetectionJobProperties_jobStatus-keyPhrasesDetectionJobProperties_languageCode(keyPhrasesDetectionJobProperties_message1keyPhrasesDetectionJobProperties_outputDataConfig+keyPhrasesDetectionJobProperties_submitTime/keyPhrasesDetectionJobProperties_volumeKmsKeyId*keyPhrasesDetectionJobProperties_vpcConfig($fNFDataKeyPhrasesDetectionJobProperties*$fHashableKeyPhrasesDetectionJobProperties*$fFromJSONKeyPhrasesDetectionJobProperties$$fEqKeyPhrasesDetectionJobProperties&$fReadKeyPhrasesDetectionJobProperties&$fShowKeyPhrasesDetectionJobProperties)$fGenericKeyPhrasesDetectionJobPropertiesEntityRecognizerPropertiesEntityRecognizerProperties'2$sel:dataAccessRoleArn:EntityRecognizerProperties'($sel:endTime:EntityRecognizerProperties'4$sel:entityRecognizerArn:EntityRecognizerProperties'0$sel:inputDataConfig:EntityRecognizerProperties'-$sel:languageCode:EntityRecognizerProperties'($sel:message:EntityRecognizerProperties'.$sel:modelKmsKeyId:EntityRecognizerProperties'3$sel:recognizerMetadata:EntityRecognizerProperties'/$sel:sourceModelArn:EntityRecognizerProperties''$sel:status:EntityRecognizerProperties'+$sel:submitTime:EntityRecognizerProperties'0$sel:trainingEndTime:EntityRecognizerProperties'2$sel:trainingStartTime:EntityRecognizerProperties',$sel:versionName:EntityRecognizerProperties'/$sel:volumeKmsKeyId:EntityRecognizerProperties'*$sel:vpcConfig:EntityRecognizerProperties'newEntityRecognizerProperties,entityRecognizerProperties_dataAccessRoleArn"entityRecognizerProperties_endTime.entityRecognizerProperties_entityRecognizerArn*entityRecognizerProperties_inputDataConfig'entityRecognizerProperties_languageCode"entityRecognizerProperties_message(entityRecognizerProperties_modelKmsKeyId-entityRecognizerProperties_recognizerMetadata)entityRecognizerProperties_sourceModelArn!entityRecognizerProperties_status%entityRecognizerProperties_submitTime*entityRecognizerProperties_trainingEndTime,entityRecognizerProperties_trainingStartTime&entityRecognizerProperties_versionName)entityRecognizerProperties_volumeKmsKeyId$entityRecognizerProperties_vpcConfig"$fNFDataEntityRecognizerProperties$$fHashableEntityRecognizerProperties$$fFromJSONEntityRecognizerProperties$fEqEntityRecognizerProperties $fShowEntityRecognizerProperties#$fGenericEntityRecognizerPropertiesEntitiesDetectionJobPropertiesEntitiesDetectionJobProperties'6$sel:dataAccessRoleArn:EntitiesDetectionJobProperties',$sel:endTime:EntitiesDetectionJobProperties'8$sel:entityRecognizerArn:EntitiesDetectionJobProperties'4$sel:inputDataConfig:EntitiesDetectionJobProperties'+$sel:jobArn:EntitiesDetectionJobProperties'*$sel:jobId:EntitiesDetectionJobProperties',$sel:jobName:EntitiesDetectionJobProperties'.$sel:jobStatus:EntitiesDetectionJobProperties'1$sel:languageCode:EntitiesDetectionJobProperties',$sel:message:EntitiesDetectionJobProperties'5$sel:outputDataConfig:EntitiesDetectionJobProperties'/$sel:submitTime:EntitiesDetectionJobProperties'3$sel:volumeKmsKeyId:EntitiesDetectionJobProperties'.$sel:vpcConfig:EntitiesDetectionJobProperties'!newEntitiesDetectionJobProperties0entitiesDetectionJobProperties_dataAccessRoleArn&entitiesDetectionJobProperties_endTime2entitiesDetectionJobProperties_entityRecognizerArn.entitiesDetectionJobProperties_inputDataConfig%entitiesDetectionJobProperties_jobArn$entitiesDetectionJobProperties_jobId&entitiesDetectionJobProperties_jobName(entitiesDetectionJobProperties_jobStatus+entitiesDetectionJobProperties_languageCode&entitiesDetectionJobProperties_message/entitiesDetectionJobProperties_outputDataConfig)entitiesDetectionJobProperties_submitTime-entitiesDetectionJobProperties_volumeKmsKeyId(entitiesDetectionJobProperties_vpcConfig&$fNFDataEntitiesDetectionJobProperties($fHashableEntitiesDetectionJobProperties($fFromJSONEntitiesDetectionJobProperties"$fEqEntitiesDetectionJobProperties$$fReadEntitiesDetectionJobProperties$$fShowEntitiesDetectionJobProperties'$fGenericEntitiesDetectionJobProperties&DominantLanguageDetectionJobProperties'DominantLanguageDetectionJobProperties'>$sel:dataAccessRoleArn:DominantLanguageDetectionJobProperties'4$sel:endTime:DominantLanguageDetectionJobProperties'<$sel:inputDataConfig:DominantLanguageDetectionJobProperties'3$sel:jobArn:DominantLanguageDetectionJobProperties'2$sel:jobId:DominantLanguageDetectionJobProperties'4$sel:jobName:DominantLanguageDetectionJobProperties'6$sel:jobStatus:DominantLanguageDetectionJobProperties'4$sel:message:DominantLanguageDetectionJobProperties'=$sel:outputDataConfig:DominantLanguageDetectionJobProperties'7$sel:submitTime:DominantLanguageDetectionJobProperties';$sel:volumeKmsKeyId:DominantLanguageDetectionJobProperties'6$sel:vpcConfig:DominantLanguageDetectionJobProperties')newDominantLanguageDetectionJobProperties8dominantLanguageDetectionJobProperties_dataAccessRoleArn.dominantLanguageDetectionJobProperties_endTime6dominantLanguageDetectionJobProperties_inputDataConfig-dominantLanguageDetectionJobProperties_jobArn,dominantLanguageDetectionJobProperties_jobId.dominantLanguageDetectionJobProperties_jobName0dominantLanguageDetectionJobProperties_jobStatus.dominantLanguageDetectionJobProperties_message7dominantLanguageDetectionJobProperties_outputDataConfig1dominantLanguageDetectionJobProperties_submitTime5dominantLanguageDetectionJobProperties_volumeKmsKeyId0dominantLanguageDetectionJobProperties_vpcConfig.$fNFDataDominantLanguageDetectionJobProperties0$fHashableDominantLanguageDetectionJobProperties0$fFromJSONDominantLanguageDetectionJobProperties*$fEqDominantLanguageDetectionJobProperties,$fReadDominantLanguageDetectionJobProperties,$fShowDominantLanguageDetectionJobProperties/$fGenericDominantLanguageDetectionJobPropertiesDocumentClassifierPropertiesDocumentClassifierProperties'5$sel:classifierMetadata:DocumentClassifierProperties'4$sel:dataAccessRoleArn:DocumentClassifierProperties'8$sel:documentClassifierArn:DocumentClassifierProperties'*$sel:endTime:DocumentClassifierProperties'2$sel:inputDataConfig:DocumentClassifierProperties'/$sel:languageCode:DocumentClassifierProperties'*$sel:message:DocumentClassifierProperties''$sel:mode:DocumentClassifierProperties'0$sel:modelKmsKeyId:DocumentClassifierProperties'3$sel:outputDataConfig:DocumentClassifierProperties'1$sel:sourceModelArn:DocumentClassifierProperties')$sel:status:DocumentClassifierProperties'-$sel:submitTime:DocumentClassifierProperties'2$sel:trainingEndTime:DocumentClassifierProperties'4$sel:trainingStartTime:DocumentClassifierProperties'.$sel:versionName:DocumentClassifierProperties'1$sel:volumeKmsKeyId:DocumentClassifierProperties',$sel:vpcConfig:DocumentClassifierProperties'newDocumentClassifierProperties/documentClassifierProperties_classifierMetadata.documentClassifierProperties_dataAccessRoleArn2documentClassifierProperties_documentClassifierArn$documentClassifierProperties_endTime,documentClassifierProperties_inputDataConfig)documentClassifierProperties_languageCode$documentClassifierProperties_message!documentClassifierProperties_mode*documentClassifierProperties_modelKmsKeyId-documentClassifierProperties_outputDataConfig+documentClassifierProperties_sourceModelArn#documentClassifierProperties_status'documentClassifierProperties_submitTime,documentClassifierProperties_trainingEndTime.documentClassifierProperties_trainingStartTime(documentClassifierProperties_versionName+documentClassifierProperties_volumeKmsKeyId&documentClassifierProperties_vpcConfig$$fNFDataDocumentClassifierProperties&$fHashableDocumentClassifierProperties&$fFromJSONDocumentClassifierProperties $fEqDocumentClassifierProperties"$fShowDocumentClassifierProperties%$fGenericDocumentClassifierProperties#DocumentClassificationJobProperties$DocumentClassificationJobProperties';$sel:dataAccessRoleArn:DocumentClassificationJobProperties'?$sel:documentClassifierArn:DocumentClassificationJobProperties'1$sel:endTime:DocumentClassificationJobProperties'9$sel:inputDataConfig:DocumentClassificationJobProperties'0$sel:jobArn:DocumentClassificationJobProperties'/$sel:jobId:DocumentClassificationJobProperties'1$sel:jobName:DocumentClassificationJobProperties'3$sel:jobStatus:DocumentClassificationJobProperties'1$sel:message:DocumentClassificationJobProperties':$sel:outputDataConfig:DocumentClassificationJobProperties'4$sel:submitTime:DocumentClassificationJobProperties'8$sel:volumeKmsKeyId:DocumentClassificationJobProperties'3$sel:vpcConfig:DocumentClassificationJobProperties'&newDocumentClassificationJobProperties5documentClassificationJobProperties_dataAccessRoleArn9documentClassificationJobProperties_documentClassifierArn+documentClassificationJobProperties_endTime3documentClassificationJobProperties_inputDataConfig*documentClassificationJobProperties_jobArn)documentClassificationJobProperties_jobId+documentClassificationJobProperties_jobName-documentClassificationJobProperties_jobStatus+documentClassificationJobProperties_message4documentClassificationJobProperties_outputDataConfig.documentClassificationJobProperties_submitTime2documentClassificationJobProperties_volumeKmsKeyId-documentClassificationJobProperties_vpcConfig+$fNFDataDocumentClassificationJobProperties-$fHashableDocumentClassificationJobProperties-$fFromJSONDocumentClassificationJobProperties'$fEqDocumentClassificationJobProperties)$fReadDocumentClassificationJobProperties)$fShowDocumentClassificationJobProperties,$fGenericDocumentClassificationJobPropertiesdefaultService _BatchSizeLimitExceededException _ConcurrentModificationException_InternalServerException_InvalidFilterException_InvalidRequestException_JobNotFoundException_KmsKeyValidationException_ResourceInUseException_ResourceLimitExceededException_ResourceNotFoundException_ResourceUnavailableException_TextSizeLimitExceededException_TooManyRequestsException_TooManyTagKeysException_TooManyTagsException_UnsupportedLanguageExceptionTagResourceResponseTagResourceResponse'$$sel:httpStatus:TagResourceResponse' TagResource TagResource'$sel:resourceArn:TagResource'$sel:tags:TagResource'newTagResourcetagResource_resourceArntagResource_tagsnewTagResourceResponsetagResourceResponse_httpStatus$fToQueryTagResource$fToPathTagResource$fToJSONTagResource$fToHeadersTagResource$fNFDataTagResource$fHashableTagResource$fNFDataTagResourceResponse$fAWSRequestTagResource$fEqTagResourceResponse$fReadTagResourceResponse$fShowTagResourceResponse$fGenericTagResourceResponse$fEqTagResource$fReadTagResource$fShowTagResource$fGenericTagResource$StopTrainingEntityRecognizerResponse%StopTrainingEntityRecognizerResponse'5$sel:httpStatus:StopTrainingEntityRecognizerResponse'StopTrainingEntityRecognizerStopTrainingEntityRecognizer'6$sel:entityRecognizerArn:StopTrainingEntityRecognizer'newStopTrainingEntityRecognizer0stopTrainingEntityRecognizer_entityRecognizerArn'newStopTrainingEntityRecognizerResponse/stopTrainingEntityRecognizerResponse_httpStatus%$fToQueryStopTrainingEntityRecognizer$$fToPathStopTrainingEntityRecognizer$$fToJSONStopTrainingEntityRecognizer'$fToHeadersStopTrainingEntityRecognizer$$fNFDataStopTrainingEntityRecognizer&$fHashableStopTrainingEntityRecognizer,$fNFDataStopTrainingEntityRecognizerResponse($fAWSRequestStopTrainingEntityRecognizer($fEqStopTrainingEntityRecognizerResponse*$fReadStopTrainingEntityRecognizerResponse*$fShowStopTrainingEntityRecognizerResponse-$fGenericStopTrainingEntityRecognizerResponse $fEqStopTrainingEntityRecognizer"$fReadStopTrainingEntityRecognizer"$fShowStopTrainingEntityRecognizer%$fGenericStopTrainingEntityRecognizer&StopTrainingDocumentClassifierResponse'StopTrainingDocumentClassifierResponse'7$sel:httpStatus:StopTrainingDocumentClassifierResponse'StopTrainingDocumentClassifierStopTrainingDocumentClassifier':$sel:documentClassifierArn:StopTrainingDocumentClassifier'!newStopTrainingDocumentClassifier4stopTrainingDocumentClassifier_documentClassifierArn)newStopTrainingDocumentClassifierResponse1stopTrainingDocumentClassifierResponse_httpStatus'$fToQueryStopTrainingDocumentClassifier&$fToPathStopTrainingDocumentClassifier&$fToJSONStopTrainingDocumentClassifier)$fToHeadersStopTrainingDocumentClassifier&$fNFDataStopTrainingDocumentClassifier($fHashableStopTrainingDocumentClassifier.$fNFDataStopTrainingDocumentClassifierResponse*$fAWSRequestStopTrainingDocumentClassifier*$fEqStopTrainingDocumentClassifierResponse,$fReadStopTrainingDocumentClassifierResponse,$fShowStopTrainingDocumentClassifierResponse/$fGenericStopTrainingDocumentClassifierResponse"$fEqStopTrainingDocumentClassifier$$fReadStopTrainingDocumentClassifier$$fShowStopTrainingDocumentClassifier'$fGenericStopTrainingDocumentClassifier)StopTargetedSentimentDetectionJobResponse*StopTargetedSentimentDetectionJobResponse'5$sel:jobId:StopTargetedSentimentDetectionJobResponse'9$sel:jobStatus:StopTargetedSentimentDetectionJobResponse':$sel:httpStatus:StopTargetedSentimentDetectionJobResponse'!StopTargetedSentimentDetectionJob"StopTargetedSentimentDetectionJob'-$sel:jobId:StopTargetedSentimentDetectionJob'$newStopTargetedSentimentDetectionJob'stopTargetedSentimentDetectionJob_jobId,newStopTargetedSentimentDetectionJobResponse/stopTargetedSentimentDetectionJobResponse_jobId3stopTargetedSentimentDetectionJobResponse_jobStatus4stopTargetedSentimentDetectionJobResponse_httpStatus*$fToQueryStopTargetedSentimentDetectionJob)$fToPathStopTargetedSentimentDetectionJob)$fToJSONStopTargetedSentimentDetectionJob,$fToHeadersStopTargetedSentimentDetectionJob)$fNFDataStopTargetedSentimentDetectionJob+$fHashableStopTargetedSentimentDetectionJob1$fNFDataStopTargetedSentimentDetectionJobResponse-$fAWSRequestStopTargetedSentimentDetectionJob-$fEqStopTargetedSentimentDetectionJobResponse/$fReadStopTargetedSentimentDetectionJobResponse/$fShowStopTargetedSentimentDetectionJobResponse2$fGenericStopTargetedSentimentDetectionJobResponse%$fEqStopTargetedSentimentDetectionJob'$fReadStopTargetedSentimentDetectionJob'$fShowStopTargetedSentimentDetectionJob*$fGenericStopTargetedSentimentDetectionJob!StopSentimentDetectionJobResponse"StopSentimentDetectionJobResponse'-$sel:jobId:StopSentimentDetectionJobResponse'1$sel:jobStatus:StopSentimentDetectionJobResponse'2$sel:httpStatus:StopSentimentDetectionJobResponse'StopSentimentDetectionJobStopSentimentDetectionJob'%$sel:jobId:StopSentimentDetectionJob'newStopSentimentDetectionJobstopSentimentDetectionJob_jobId$newStopSentimentDetectionJobResponse'stopSentimentDetectionJobResponse_jobId+stopSentimentDetectionJobResponse_jobStatus,stopSentimentDetectionJobResponse_httpStatus"$fToQueryStopSentimentDetectionJob!$fToPathStopSentimentDetectionJob!$fToJSONStopSentimentDetectionJob$$fToHeadersStopSentimentDetectionJob!$fNFDataStopSentimentDetectionJob#$fHashableStopSentimentDetectionJob)$fNFDataStopSentimentDetectionJobResponse%$fAWSRequestStopSentimentDetectionJob%$fEqStopSentimentDetectionJobResponse'$fReadStopSentimentDetectionJobResponse'$fShowStopSentimentDetectionJobResponse*$fGenericStopSentimentDetectionJobResponse$fEqStopSentimentDetectionJob$fReadStopSentimentDetectionJob$fShowStopSentimentDetectionJob"$fGenericStopSentimentDetectionJob#StopPiiEntitiesDetectionJobResponse$StopPiiEntitiesDetectionJobResponse'/$sel:jobId:StopPiiEntitiesDetectionJobResponse'3$sel:jobStatus:StopPiiEntitiesDetectionJobResponse'4$sel:httpStatus:StopPiiEntitiesDetectionJobResponse'StopPiiEntitiesDetectionJobStopPiiEntitiesDetectionJob''$sel:jobId:StopPiiEntitiesDetectionJob'newStopPiiEntitiesDetectionJob!stopPiiEntitiesDetectionJob_jobId&newStopPiiEntitiesDetectionJobResponse)stopPiiEntitiesDetectionJobResponse_jobId-stopPiiEntitiesDetectionJobResponse_jobStatus.stopPiiEntitiesDetectionJobResponse_httpStatus$$fToQueryStopPiiEntitiesDetectionJob#$fToPathStopPiiEntitiesDetectionJob#$fToJSONStopPiiEntitiesDetectionJob&$fToHeadersStopPiiEntitiesDetectionJob#$fNFDataStopPiiEntitiesDetectionJob%$fHashableStopPiiEntitiesDetectionJob+$fNFDataStopPiiEntitiesDetectionJobResponse'$fAWSRequestStopPiiEntitiesDetectionJob'$fEqStopPiiEntitiesDetectionJobResponse)$fReadStopPiiEntitiesDetectionJobResponse)$fShowStopPiiEntitiesDetectionJobResponse,$fGenericStopPiiEntitiesDetectionJobResponse$fEqStopPiiEntitiesDetectionJob!$fReadStopPiiEntitiesDetectionJob!$fShowStopPiiEntitiesDetectionJob$$fGenericStopPiiEntitiesDetectionJob"StopKeyPhrasesDetectionJobResponse#StopKeyPhrasesDetectionJobResponse'.$sel:jobId:StopKeyPhrasesDetectionJobResponse'2$sel:jobStatus:StopKeyPhrasesDetectionJobResponse'3$sel:httpStatus:StopKeyPhrasesDetectionJobResponse'StopKeyPhrasesDetectionJobStopKeyPhrasesDetectionJob'&$sel:jobId:StopKeyPhrasesDetectionJob'newStopKeyPhrasesDetectionJob stopKeyPhrasesDetectionJob_jobId%newStopKeyPhrasesDetectionJobResponse(stopKeyPhrasesDetectionJobResponse_jobId,stopKeyPhrasesDetectionJobResponse_jobStatus-stopKeyPhrasesDetectionJobResponse_httpStatus#$fToQueryStopKeyPhrasesDetectionJob"$fToPathStopKeyPhrasesDetectionJob"$fToJSONStopKeyPhrasesDetectionJob%$fToHeadersStopKeyPhrasesDetectionJob"$fNFDataStopKeyPhrasesDetectionJob$$fHashableStopKeyPhrasesDetectionJob*$fNFDataStopKeyPhrasesDetectionJobResponse&$fAWSRequestStopKeyPhrasesDetectionJob&$fEqStopKeyPhrasesDetectionJobResponse($fReadStopKeyPhrasesDetectionJobResponse($fShowStopKeyPhrasesDetectionJobResponse+$fGenericStopKeyPhrasesDetectionJobResponse$fEqStopKeyPhrasesDetectionJob $fReadStopKeyPhrasesDetectionJob $fShowStopKeyPhrasesDetectionJob#$fGenericStopKeyPhrasesDetectionJobStopEventsDetectionJobResponseStopEventsDetectionJobResponse'*$sel:jobId:StopEventsDetectionJobResponse'.$sel:jobStatus:StopEventsDetectionJobResponse'/$sel:httpStatus:StopEventsDetectionJobResponse'StopEventsDetectionJobStopEventsDetectionJob'"$sel:jobId:StopEventsDetectionJob'newStopEventsDetectionJobstopEventsDetectionJob_jobId!newStopEventsDetectionJobResponse$stopEventsDetectionJobResponse_jobId(stopEventsDetectionJobResponse_jobStatus)stopEventsDetectionJobResponse_httpStatus$fToQueryStopEventsDetectionJob$fToPathStopEventsDetectionJob$fToJSONStopEventsDetectionJob!$fToHeadersStopEventsDetectionJob$fNFDataStopEventsDetectionJob $fHashableStopEventsDetectionJob&$fNFDataStopEventsDetectionJobResponse"$fAWSRequestStopEventsDetectionJob"$fEqStopEventsDetectionJobResponse$$fReadStopEventsDetectionJobResponse$$fShowStopEventsDetectionJobResponse'$fGenericStopEventsDetectionJobResponse$fEqStopEventsDetectionJob$fReadStopEventsDetectionJob$fShowStopEventsDetectionJob$fGenericStopEventsDetectionJob StopEntitiesDetectionJobResponse!StopEntitiesDetectionJobResponse',$sel:jobId:StopEntitiesDetectionJobResponse'0$sel:jobStatus:StopEntitiesDetectionJobResponse'1$sel:httpStatus:StopEntitiesDetectionJobResponse'StopEntitiesDetectionJobStopEntitiesDetectionJob'$$sel:jobId:StopEntitiesDetectionJob'newStopEntitiesDetectionJobstopEntitiesDetectionJob_jobId#newStopEntitiesDetectionJobResponse&stopEntitiesDetectionJobResponse_jobId*stopEntitiesDetectionJobResponse_jobStatus+stopEntitiesDetectionJobResponse_httpStatus!$fToQueryStopEntitiesDetectionJob $fToPathStopEntitiesDetectionJob $fToJSONStopEntitiesDetectionJob#$fToHeadersStopEntitiesDetectionJob $fNFDataStopEntitiesDetectionJob"$fHashableStopEntitiesDetectionJob($fNFDataStopEntitiesDetectionJobResponse$$fAWSRequestStopEntitiesDetectionJob$$fEqStopEntitiesDetectionJobResponse&$fReadStopEntitiesDetectionJobResponse&$fShowStopEntitiesDetectionJobResponse)$fGenericStopEntitiesDetectionJobResponse$fEqStopEntitiesDetectionJob$fReadStopEntitiesDetectionJob$fShowStopEntitiesDetectionJob!$fGenericStopEntitiesDetectionJob(StopDominantLanguageDetectionJobResponse)StopDominantLanguageDetectionJobResponse'4$sel:jobId:StopDominantLanguageDetectionJobResponse'8$sel:jobStatus:StopDominantLanguageDetectionJobResponse'9$sel:httpStatus:StopDominantLanguageDetectionJobResponse' StopDominantLanguageDetectionJob!StopDominantLanguageDetectionJob',$sel:jobId:StopDominantLanguageDetectionJob'#newStopDominantLanguageDetectionJob&stopDominantLanguageDetectionJob_jobId+newStopDominantLanguageDetectionJobResponse.stopDominantLanguageDetectionJobResponse_jobId2stopDominantLanguageDetectionJobResponse_jobStatus3stopDominantLanguageDetectionJobResponse_httpStatus)$fToQueryStopDominantLanguageDetectionJob($fToPathStopDominantLanguageDetectionJob($fToJSONStopDominantLanguageDetectionJob+$fToHeadersStopDominantLanguageDetectionJob($fNFDataStopDominantLanguageDetectionJob*$fHashableStopDominantLanguageDetectionJob0$fNFDataStopDominantLanguageDetectionJobResponse,$fAWSRequestStopDominantLanguageDetectionJob,$fEqStopDominantLanguageDetectionJobResponse.$fReadStopDominantLanguageDetectionJobResponse.$fShowStopDominantLanguageDetectionJobResponse1$fGenericStopDominantLanguageDetectionJobResponse$$fEqStopDominantLanguageDetectionJob&$fReadStopDominantLanguageDetectionJob&$fShowStopDominantLanguageDetectionJob)$fGenericStopDominantLanguageDetectionJobStartTopicsDetectionJobResponse StartTopicsDetectionJobResponse',$sel:jobArn:StartTopicsDetectionJobResponse'+$sel:jobId:StartTopicsDetectionJobResponse'/$sel:jobStatus:StartTopicsDetectionJobResponse'0$sel:httpStatus:StartTopicsDetectionJobResponse'StartTopicsDetectionJobStartTopicsDetectionJob'0$sel:clientRequestToken:StartTopicsDetectionJob'%$sel:jobName:StartTopicsDetectionJob',$sel:numberOfTopics:StartTopicsDetectionJob'"$sel:tags:StartTopicsDetectionJob',$sel:volumeKmsKeyId:StartTopicsDetectionJob''$sel:vpcConfig:StartTopicsDetectionJob'-$sel:inputDataConfig:StartTopicsDetectionJob'.$sel:outputDataConfig:StartTopicsDetectionJob'/$sel:dataAccessRoleArn:StartTopicsDetectionJob'newStartTopicsDetectionJob*startTopicsDetectionJob_clientRequestTokenstartTopicsDetectionJob_jobName&startTopicsDetectionJob_numberOfTopicsstartTopicsDetectionJob_tags&startTopicsDetectionJob_volumeKmsKeyId!startTopicsDetectionJob_vpcConfig'startTopicsDetectionJob_inputDataConfig(startTopicsDetectionJob_outputDataConfig)startTopicsDetectionJob_dataAccessRoleArn"newStartTopicsDetectionJobResponse&startTopicsDetectionJobResponse_jobArn%startTopicsDetectionJobResponse_jobId)startTopicsDetectionJobResponse_jobStatus*startTopicsDetectionJobResponse_httpStatus $fToQueryStartTopicsDetectionJob$fToPathStartTopicsDetectionJob$fToJSONStartTopicsDetectionJob"$fToHeadersStartTopicsDetectionJob$fNFDataStartTopicsDetectionJob!$fHashableStartTopicsDetectionJob'$fNFDataStartTopicsDetectionJobResponse#$fAWSRequestStartTopicsDetectionJob#$fEqStartTopicsDetectionJobResponse%$fReadStartTopicsDetectionJobResponse%$fShowStartTopicsDetectionJobResponse($fGenericStartTopicsDetectionJobResponse$fEqStartTopicsDetectionJob$fReadStartTopicsDetectionJob$fShowStartTopicsDetectionJob $fGenericStartTopicsDetectionJob*StartTargetedSentimentDetectionJobResponse+StartTargetedSentimentDetectionJobResponse'7$sel:jobArn:StartTargetedSentimentDetectionJobResponse'6$sel:jobId:StartTargetedSentimentDetectionJobResponse':$sel:jobStatus:StartTargetedSentimentDetectionJobResponse';$sel:httpStatus:StartTargetedSentimentDetectionJobResponse'"StartTargetedSentimentDetectionJob#StartTargetedSentimentDetectionJob';$sel:clientRequestToken:StartTargetedSentimentDetectionJob'0$sel:jobName:StartTargetedSentimentDetectionJob'-$sel:tags:StartTargetedSentimentDetectionJob'7$sel:volumeKmsKeyId:StartTargetedSentimentDetectionJob'2$sel:vpcConfig:StartTargetedSentimentDetectionJob'8$sel:inputDataConfig:StartTargetedSentimentDetectionJob'9$sel:outputDataConfig:StartTargetedSentimentDetectionJob':$sel:dataAccessRoleArn:StartTargetedSentimentDetectionJob'5$sel:languageCode:StartTargetedSentimentDetectionJob'%newStartTargetedSentimentDetectionJob5startTargetedSentimentDetectionJob_clientRequestToken*startTargetedSentimentDetectionJob_jobName'startTargetedSentimentDetectionJob_tags1startTargetedSentimentDetectionJob_volumeKmsKeyId,startTargetedSentimentDetectionJob_vpcConfig2startTargetedSentimentDetectionJob_inputDataConfig3startTargetedSentimentDetectionJob_outputDataConfig4startTargetedSentimentDetectionJob_dataAccessRoleArn/startTargetedSentimentDetectionJob_languageCode-newStartTargetedSentimentDetectionJobResponse1startTargetedSentimentDetectionJobResponse_jobArn0startTargetedSentimentDetectionJobResponse_jobId4startTargetedSentimentDetectionJobResponse_jobStatus5startTargetedSentimentDetectionJobResponse_httpStatus+$fToQueryStartTargetedSentimentDetectionJob*$fToPathStartTargetedSentimentDetectionJob*$fToJSONStartTargetedSentimentDetectionJob-$fToHeadersStartTargetedSentimentDetectionJob*$fNFDataStartTargetedSentimentDetectionJob,$fHashableStartTargetedSentimentDetectionJob2$fNFDataStartTargetedSentimentDetectionJobResponse.$fAWSRequestStartTargetedSentimentDetectionJob.$fEqStartTargetedSentimentDetectionJobResponse0$fReadStartTargetedSentimentDetectionJobResponse0$fShowStartTargetedSentimentDetectionJobResponse3$fGenericStartTargetedSentimentDetectionJobResponse&$fEqStartTargetedSentimentDetectionJob($fReadStartTargetedSentimentDetectionJob($fShowStartTargetedSentimentDetectionJob+$fGenericStartTargetedSentimentDetectionJob"StartSentimentDetectionJobResponse#StartSentimentDetectionJobResponse'/$sel:jobArn:StartSentimentDetectionJobResponse'.$sel:jobId:StartSentimentDetectionJobResponse'2$sel:jobStatus:StartSentimentDetectionJobResponse'3$sel:httpStatus:StartSentimentDetectionJobResponse'StartSentimentDetectionJobStartSentimentDetectionJob'3$sel:clientRequestToken:StartSentimentDetectionJob'($sel:jobName:StartSentimentDetectionJob'%$sel:tags:StartSentimentDetectionJob'/$sel:volumeKmsKeyId:StartSentimentDetectionJob'*$sel:vpcConfig:StartSentimentDetectionJob'0$sel:inputDataConfig:StartSentimentDetectionJob'1$sel:outputDataConfig:StartSentimentDetectionJob'2$sel:dataAccessRoleArn:StartSentimentDetectionJob'-$sel:languageCode:StartSentimentDetectionJob'newStartSentimentDetectionJob-startSentimentDetectionJob_clientRequestToken"startSentimentDetectionJob_jobNamestartSentimentDetectionJob_tags)startSentimentDetectionJob_volumeKmsKeyId$startSentimentDetectionJob_vpcConfig*startSentimentDetectionJob_inputDataConfig+startSentimentDetectionJob_outputDataConfig,startSentimentDetectionJob_dataAccessRoleArn'startSentimentDetectionJob_languageCode%newStartSentimentDetectionJobResponse)startSentimentDetectionJobResponse_jobArn(startSentimentDetectionJobResponse_jobId,startSentimentDetectionJobResponse_jobStatus-startSentimentDetectionJobResponse_httpStatus#$fToQueryStartSentimentDetectionJob"$fToPathStartSentimentDetectionJob"$fToJSONStartSentimentDetectionJob%$fToHeadersStartSentimentDetectionJob"$fNFDataStartSentimentDetectionJob$$fHashableStartSentimentDetectionJob*$fNFDataStartSentimentDetectionJobResponse&$fAWSRequestStartSentimentDetectionJob&$fEqStartSentimentDetectionJobResponse($fReadStartSentimentDetectionJobResponse($fShowStartSentimentDetectionJobResponse+$fGenericStartSentimentDetectionJobResponse$fEqStartSentimentDetectionJob $fReadStartSentimentDetectionJob $fShowStartSentimentDetectionJob#$fGenericStartSentimentDetectionJob$StartPiiEntitiesDetectionJobResponse%StartPiiEntitiesDetectionJobResponse'1$sel:jobArn:StartPiiEntitiesDetectionJobResponse'0$sel:jobId:StartPiiEntitiesDetectionJobResponse'4$sel:jobStatus:StartPiiEntitiesDetectionJobResponse'5$sel:httpStatus:StartPiiEntitiesDetectionJobResponse'StartPiiEntitiesDetectionJobStartPiiEntitiesDetectionJob'5$sel:clientRequestToken:StartPiiEntitiesDetectionJob'*$sel:jobName:StartPiiEntitiesDetectionJob'2$sel:redactionConfig:StartPiiEntitiesDetectionJob''$sel:tags:StartPiiEntitiesDetectionJob'2$sel:inputDataConfig:StartPiiEntitiesDetectionJob'3$sel:outputDataConfig:StartPiiEntitiesDetectionJob''$sel:mode:StartPiiEntitiesDetectionJob'4$sel:dataAccessRoleArn:StartPiiEntitiesDetectionJob'/$sel:languageCode:StartPiiEntitiesDetectionJob'newStartPiiEntitiesDetectionJob/startPiiEntitiesDetectionJob_clientRequestToken$startPiiEntitiesDetectionJob_jobName,startPiiEntitiesDetectionJob_redactionConfig!startPiiEntitiesDetectionJob_tags,startPiiEntitiesDetectionJob_inputDataConfig-startPiiEntitiesDetectionJob_outputDataConfig!startPiiEntitiesDetectionJob_mode.startPiiEntitiesDetectionJob_dataAccessRoleArn)startPiiEntitiesDetectionJob_languageCode'newStartPiiEntitiesDetectionJobResponse+startPiiEntitiesDetectionJobResponse_jobArn*startPiiEntitiesDetectionJobResponse_jobId.startPiiEntitiesDetectionJobResponse_jobStatus/startPiiEntitiesDetectionJobResponse_httpStatus%$fToQueryStartPiiEntitiesDetectionJob$$fToPathStartPiiEntitiesDetectionJob$$fToJSONStartPiiEntitiesDetectionJob'$fToHeadersStartPiiEntitiesDetectionJob$$fNFDataStartPiiEntitiesDetectionJob&$fHashableStartPiiEntitiesDetectionJob,$fNFDataStartPiiEntitiesDetectionJobResponse($fAWSRequestStartPiiEntitiesDetectionJob($fEqStartPiiEntitiesDetectionJobResponse*$fReadStartPiiEntitiesDetectionJobResponse*$fShowStartPiiEntitiesDetectionJobResponse-$fGenericStartPiiEntitiesDetectionJobResponse $fEqStartPiiEntitiesDetectionJob"$fReadStartPiiEntitiesDetectionJob"$fShowStartPiiEntitiesDetectionJob%$fGenericStartPiiEntitiesDetectionJob#StartKeyPhrasesDetectionJobResponse$StartKeyPhrasesDetectionJobResponse'0$sel:jobArn:StartKeyPhrasesDetectionJobResponse'/$sel:jobId:StartKeyPhrasesDetectionJobResponse'3$sel:jobStatus:StartKeyPhrasesDetectionJobResponse'4$sel:httpStatus:StartKeyPhrasesDetectionJobResponse'StartKeyPhrasesDetectionJobStartKeyPhrasesDetectionJob'4$sel:clientRequestToken:StartKeyPhrasesDetectionJob')$sel:jobName:StartKeyPhrasesDetectionJob'&$sel:tags:StartKeyPhrasesDetectionJob'0$sel:volumeKmsKeyId:StartKeyPhrasesDetectionJob'+$sel:vpcConfig:StartKeyPhrasesDetectionJob'1$sel:inputDataConfig:StartKeyPhrasesDetectionJob'2$sel:outputDataConfig:StartKeyPhrasesDetectionJob'3$sel:dataAccessRoleArn:StartKeyPhrasesDetectionJob'.$sel:languageCode:StartKeyPhrasesDetectionJob'newStartKeyPhrasesDetectionJob.startKeyPhrasesDetectionJob_clientRequestToken#startKeyPhrasesDetectionJob_jobName startKeyPhrasesDetectionJob_tags*startKeyPhrasesDetectionJob_volumeKmsKeyId%startKeyPhrasesDetectionJob_vpcConfig+startKeyPhrasesDetectionJob_inputDataConfig,startKeyPhrasesDetectionJob_outputDataConfig-startKeyPhrasesDetectionJob_dataAccessRoleArn(startKeyPhrasesDetectionJob_languageCode&newStartKeyPhrasesDetectionJobResponse*startKeyPhrasesDetectionJobResponse_jobArn)startKeyPhrasesDetectionJobResponse_jobId-startKeyPhrasesDetectionJobResponse_jobStatus.startKeyPhrasesDetectionJobResponse_httpStatus$$fToQueryStartKeyPhrasesDetectionJob#$fToPathStartKeyPhrasesDetectionJob#$fToJSONStartKeyPhrasesDetectionJob&$fToHeadersStartKeyPhrasesDetectionJob#$fNFDataStartKeyPhrasesDetectionJob%$fHashableStartKeyPhrasesDetectionJob+$fNFDataStartKeyPhrasesDetectionJobResponse'$fAWSRequestStartKeyPhrasesDetectionJob'$fEqStartKeyPhrasesDetectionJobResponse)$fReadStartKeyPhrasesDetectionJobResponse)$fShowStartKeyPhrasesDetectionJobResponse,$fGenericStartKeyPhrasesDetectionJobResponse$fEqStartKeyPhrasesDetectionJob!$fReadStartKeyPhrasesDetectionJob!$fShowStartKeyPhrasesDetectionJob$$fGenericStartKeyPhrasesDetectionJobStartEventsDetectionJobResponse StartEventsDetectionJobResponse',$sel:jobArn:StartEventsDetectionJobResponse'+$sel:jobId:StartEventsDetectionJobResponse'/$sel:jobStatus:StartEventsDetectionJobResponse'0$sel:httpStatus:StartEventsDetectionJobResponse'StartEventsDetectionJobStartEventsDetectionJob'0$sel:clientRequestToken:StartEventsDetectionJob'%$sel:jobName:StartEventsDetectionJob'"$sel:tags:StartEventsDetectionJob'-$sel:inputDataConfig:StartEventsDetectionJob'.$sel:outputDataConfig:StartEventsDetectionJob'/$sel:dataAccessRoleArn:StartEventsDetectionJob'*$sel:languageCode:StartEventsDetectionJob'.$sel:targetEventTypes:StartEventsDetectionJob'newStartEventsDetectionJob*startEventsDetectionJob_clientRequestTokenstartEventsDetectionJob_jobNamestartEventsDetectionJob_tags'startEventsDetectionJob_inputDataConfig(startEventsDetectionJob_outputDataConfig)startEventsDetectionJob_dataAccessRoleArn$startEventsDetectionJob_languageCode(startEventsDetectionJob_targetEventTypes"newStartEventsDetectionJobResponse&startEventsDetectionJobResponse_jobArn%startEventsDetectionJobResponse_jobId)startEventsDetectionJobResponse_jobStatus*startEventsDetectionJobResponse_httpStatus $fToQueryStartEventsDetectionJob$fToPathStartEventsDetectionJob$fToJSONStartEventsDetectionJob"$fToHeadersStartEventsDetectionJob$fNFDataStartEventsDetectionJob!$fHashableStartEventsDetectionJob'$fNFDataStartEventsDetectionJobResponse#$fAWSRequestStartEventsDetectionJob#$fEqStartEventsDetectionJobResponse%$fReadStartEventsDetectionJobResponse%$fShowStartEventsDetectionJobResponse($fGenericStartEventsDetectionJobResponse$fEqStartEventsDetectionJob$fReadStartEventsDetectionJob$fShowStartEventsDetectionJob $fGenericStartEventsDetectionJob!StartEntitiesDetectionJobResponse"StartEntitiesDetectionJobResponse'.$sel:jobArn:StartEntitiesDetectionJobResponse'-$sel:jobId:StartEntitiesDetectionJobResponse'1$sel:jobStatus:StartEntitiesDetectionJobResponse'2$sel:httpStatus:StartEntitiesDetectionJobResponse'StartEntitiesDetectionJobStartEntitiesDetectionJob'2$sel:clientRequestToken:StartEntitiesDetectionJob'3$sel:entityRecognizerArn:StartEntitiesDetectionJob''$sel:jobName:StartEntitiesDetectionJob'$$sel:tags:StartEntitiesDetectionJob'.$sel:volumeKmsKeyId:StartEntitiesDetectionJob')$sel:vpcConfig:StartEntitiesDetectionJob'/$sel:inputDataConfig:StartEntitiesDetectionJob'0$sel:outputDataConfig:StartEntitiesDetectionJob'1$sel:dataAccessRoleArn:StartEntitiesDetectionJob',$sel:languageCode:StartEntitiesDetectionJob'newStartEntitiesDetectionJob,startEntitiesDetectionJob_clientRequestToken-startEntitiesDetectionJob_entityRecognizerArn!startEntitiesDetectionJob_jobNamestartEntitiesDetectionJob_tags(startEntitiesDetectionJob_volumeKmsKeyId#startEntitiesDetectionJob_vpcConfig)startEntitiesDetectionJob_inputDataConfig*startEntitiesDetectionJob_outputDataConfig+startEntitiesDetectionJob_dataAccessRoleArn&startEntitiesDetectionJob_languageCode$newStartEntitiesDetectionJobResponse(startEntitiesDetectionJobResponse_jobArn'startEntitiesDetectionJobResponse_jobId+startEntitiesDetectionJobResponse_jobStatus,startEntitiesDetectionJobResponse_httpStatus"$fToQueryStartEntitiesDetectionJob!$fToPathStartEntitiesDetectionJob!$fToJSONStartEntitiesDetectionJob$$fToHeadersStartEntitiesDetectionJob!$fNFDataStartEntitiesDetectionJob#$fHashableStartEntitiesDetectionJob)$fNFDataStartEntitiesDetectionJobResponse%$fAWSRequestStartEntitiesDetectionJob%$fEqStartEntitiesDetectionJobResponse'$fReadStartEntitiesDetectionJobResponse'$fShowStartEntitiesDetectionJobResponse*$fGenericStartEntitiesDetectionJobResponse$fEqStartEntitiesDetectionJob$fReadStartEntitiesDetectionJob$fShowStartEntitiesDetectionJob"$fGenericStartEntitiesDetectionJob)StartDominantLanguageDetectionJobResponse*StartDominantLanguageDetectionJobResponse'6$sel:jobArn:StartDominantLanguageDetectionJobResponse'5$sel:jobId:StartDominantLanguageDetectionJobResponse'9$sel:jobStatus:StartDominantLanguageDetectionJobResponse':$sel:httpStatus:StartDominantLanguageDetectionJobResponse'!StartDominantLanguageDetectionJob"StartDominantLanguageDetectionJob':$sel:clientRequestToken:StartDominantLanguageDetectionJob'/$sel:jobName:StartDominantLanguageDetectionJob',$sel:tags:StartDominantLanguageDetectionJob'6$sel:volumeKmsKeyId:StartDominantLanguageDetectionJob'1$sel:vpcConfig:StartDominantLanguageDetectionJob'7$sel:inputDataConfig:StartDominantLanguageDetectionJob'8$sel:outputDataConfig:StartDominantLanguageDetectionJob'9$sel:dataAccessRoleArn:StartDominantLanguageDetectionJob'$newStartDominantLanguageDetectionJob4startDominantLanguageDetectionJob_clientRequestToken)startDominantLanguageDetectionJob_jobName&startDominantLanguageDetectionJob_tags0startDominantLanguageDetectionJob_volumeKmsKeyId+startDominantLanguageDetectionJob_vpcConfig1startDominantLanguageDetectionJob_inputDataConfig2startDominantLanguageDetectionJob_outputDataConfig3startDominantLanguageDetectionJob_dataAccessRoleArn,newStartDominantLanguageDetectionJobResponse0startDominantLanguageDetectionJobResponse_jobArn/startDominantLanguageDetectionJobResponse_jobId3startDominantLanguageDetectionJobResponse_jobStatus4startDominantLanguageDetectionJobResponse_httpStatus*$fToQueryStartDominantLanguageDetectionJob)$fToPathStartDominantLanguageDetectionJob)$fToJSONStartDominantLanguageDetectionJob,$fToHeadersStartDominantLanguageDetectionJob)$fNFDataStartDominantLanguageDetectionJob+$fHashableStartDominantLanguageDetectionJob1$fNFDataStartDominantLanguageDetectionJobResponse-$fAWSRequestStartDominantLanguageDetectionJob-$fEqStartDominantLanguageDetectionJobResponse/$fReadStartDominantLanguageDetectionJobResponse/$fShowStartDominantLanguageDetectionJobResponse2$fGenericStartDominantLanguageDetectionJobResponse%$fEqStartDominantLanguageDetectionJob'$fReadStartDominantLanguageDetectionJob'$fShowStartDominantLanguageDetectionJob*$fGenericStartDominantLanguageDetectionJob&StartDocumentClassificationJobResponse'StartDocumentClassificationJobResponse'3$sel:jobArn:StartDocumentClassificationJobResponse'2$sel:jobId:StartDocumentClassificationJobResponse'6$sel:jobStatus:StartDocumentClassificationJobResponse'7$sel:httpStatus:StartDocumentClassificationJobResponse'StartDocumentClassificationJobStartDocumentClassificationJob'7$sel:clientRequestToken:StartDocumentClassificationJob',$sel:jobName:StartDocumentClassificationJob')$sel:tags:StartDocumentClassificationJob'3$sel:volumeKmsKeyId:StartDocumentClassificationJob'.$sel:vpcConfig:StartDocumentClassificationJob':$sel:documentClassifierArn:StartDocumentClassificationJob'4$sel:inputDataConfig:StartDocumentClassificationJob'5$sel:outputDataConfig:StartDocumentClassificationJob'6$sel:dataAccessRoleArn:StartDocumentClassificationJob'!newStartDocumentClassificationJob1startDocumentClassificationJob_clientRequestToken&startDocumentClassificationJob_jobName#startDocumentClassificationJob_tags-startDocumentClassificationJob_volumeKmsKeyId(startDocumentClassificationJob_vpcConfig4startDocumentClassificationJob_documentClassifierArn.startDocumentClassificationJob_inputDataConfig/startDocumentClassificationJob_outputDataConfig0startDocumentClassificationJob_dataAccessRoleArn)newStartDocumentClassificationJobResponse-startDocumentClassificationJobResponse_jobArn,startDocumentClassificationJobResponse_jobId0startDocumentClassificationJobResponse_jobStatus1startDocumentClassificationJobResponse_httpStatus'$fToQueryStartDocumentClassificationJob&$fToPathStartDocumentClassificationJob&$fToJSONStartDocumentClassificationJob)$fToHeadersStartDocumentClassificationJob&$fNFDataStartDocumentClassificationJob($fHashableStartDocumentClassificationJob.$fNFDataStartDocumentClassificationJobResponse*$fAWSRequestStartDocumentClassificationJob*$fEqStartDocumentClassificationJobResponse,$fReadStartDocumentClassificationJobResponse,$fShowStartDocumentClassificationJobResponse/$fGenericStartDocumentClassificationJobResponse"$fEqStartDocumentClassificationJob$$fReadStartDocumentClassificationJob$$fShowStartDocumentClassificationJob'$fGenericStartDocumentClassificationJobPutResourcePolicyResponsePutResourcePolicyResponse'0$sel:policyRevisionId:PutResourcePolicyResponse'*$sel:httpStatus:PutResourcePolicyResponse'PutResourcePolicyPutResourcePolicy'($sel:policyRevisionId:PutResourcePolicy'#$sel:resourceArn:PutResourcePolicy'&$sel:resourcePolicy:PutResourcePolicy'newPutResourcePolicy"putResourcePolicy_policyRevisionIdputResourcePolicy_resourceArn putResourcePolicy_resourcePolicynewPutResourcePolicyResponse*putResourcePolicyResponse_policyRevisionId$putResourcePolicyResponse_httpStatus$fToQueryPutResourcePolicy$fToPathPutResourcePolicy$fToJSONPutResourcePolicy$fToHeadersPutResourcePolicy$fNFDataPutResourcePolicy$fHashablePutResourcePolicy!$fNFDataPutResourcePolicyResponse$fAWSRequestPutResourcePolicy$fEqPutResourcePolicyResponse$fReadPutResourcePolicyResponse$fShowPutResourcePolicyResponse"$fGenericPutResourcePolicyResponse$fEqPutResourcePolicy$fReadPutResourcePolicy$fShowPutResourcePolicy$fGenericPutResourcePolicyListTopicsDetectionJobsResponse ListTopicsDetectionJobsResponse'/$sel:nextToken:ListTopicsDetectionJobsResponse'$sel:topicsDetectionJobPropertiesList:ListTopicsDetectionJobsResponse'0$sel:httpStatus:ListTopicsDetectionJobsResponse'ListTopicsDetectionJobsListTopicsDetectionJobs'%$sel:filter':ListTopicsDetectionJobs'($sel:maxResults:ListTopicsDetectionJobs''$sel:nextToken:ListTopicsDetectionJobs'newListTopicsDetectionJobslistTopicsDetectionJobs_filter"listTopicsDetectionJobs_maxResults!listTopicsDetectionJobs_nextToken"newListTopicsDetectionJobsResponse)listTopicsDetectionJobsResponse_nextTokenlistTopicsDetectionJobsResponse_topicsDetectionJobPropertiesList*listTopicsDetectionJobsResponse_httpStatus $fToQueryListTopicsDetectionJobs$fToPathListTopicsDetectionJobs$fToJSONListTopicsDetectionJobs"$fToHeadersListTopicsDetectionJobs$fNFDataListTopicsDetectionJobs!$fHashableListTopicsDetectionJobs!$fAWSPagerListTopicsDetectionJobs'$fNFDataListTopicsDetectionJobsResponse#$fAWSRequestListTopicsDetectionJobs#$fEqListTopicsDetectionJobsResponse%$fReadListTopicsDetectionJobsResponse%$fShowListTopicsDetectionJobsResponse($fGenericListTopicsDetectionJobsResponse$fEqListTopicsDetectionJobs$fReadListTopicsDetectionJobs$fShowListTopicsDetectionJobs $fGenericListTopicsDetectionJobs*ListTargetedSentimentDetectionJobsResponse+ListTargetedSentimentDetectionJobsResponse':$sel:nextToken:ListTargetedSentimentDetectionJobsResponse'$sel:targetedSentimentDetectionJobPropertiesList:ListTargetedSentimentDetectionJobsResponse';$sel:httpStatus:ListTargetedSentimentDetectionJobsResponse'"ListTargetedSentimentDetectionJobs#ListTargetedSentimentDetectionJobs'0$sel:filter':ListTargetedSentimentDetectionJobs'3$sel:maxResults:ListTargetedSentimentDetectionJobs'2$sel:nextToken:ListTargetedSentimentDetectionJobs'%newListTargetedSentimentDetectionJobs)listTargetedSentimentDetectionJobs_filter-listTargetedSentimentDetectionJobs_maxResults,listTargetedSentimentDetectionJobs_nextToken-newListTargetedSentimentDetectionJobsResponse4listTargetedSentimentDetectionJobsResponse_nextTokenlistTargetedSentimentDetectionJobsResponse_targetedSentimentDetectionJobPropertiesList5listTargetedSentimentDetectionJobsResponse_httpStatus+$fToQueryListTargetedSentimentDetectionJobs*$fToPathListTargetedSentimentDetectionJobs*$fToJSONListTargetedSentimentDetectionJobs-$fToHeadersListTargetedSentimentDetectionJobs*$fNFDataListTargetedSentimentDetectionJobs,$fHashableListTargetedSentimentDetectionJobs2$fNFDataListTargetedSentimentDetectionJobsResponse.$fAWSRequestListTargetedSentimentDetectionJobs.$fEqListTargetedSentimentDetectionJobsResponse0$fReadListTargetedSentimentDetectionJobsResponse0$fShowListTargetedSentimentDetectionJobsResponse3$fGenericListTargetedSentimentDetectionJobsResponse&$fEqListTargetedSentimentDetectionJobs($fReadListTargetedSentimentDetectionJobs($fShowListTargetedSentimentDetectionJobs+$fGenericListTargetedSentimentDetectionJobsListTagsForResourceResponseListTagsForResourceResponse'-$sel:resourceArn:ListTagsForResourceResponse'&$sel:tags:ListTagsForResourceResponse',$sel:httpStatus:ListTagsForResourceResponse'ListTagsForResourceListTagsForResource'%$sel:resourceArn:ListTagsForResource'newListTagsForResourcelistTagsForResource_resourceArnnewListTagsForResourceResponse'listTagsForResourceResponse_resourceArn listTagsForResourceResponse_tags&listTagsForResourceResponse_httpStatus$fToQueryListTagsForResource$fToPathListTagsForResource$fToJSONListTagsForResource$fToHeadersListTagsForResource$fNFDataListTagsForResource$fHashableListTagsForResource#$fNFDataListTagsForResourceResponse$fAWSRequestListTagsForResource$fEqListTagsForResourceResponse!$fReadListTagsForResourceResponse!$fShowListTagsForResourceResponse$$fGenericListTagsForResourceResponse$fEqListTagsForResource$fReadListTagsForResource$fShowListTagsForResource$fGenericListTagsForResource"ListSentimentDetectionJobsResponse#ListSentimentDetectionJobsResponse'2$sel:nextToken:ListSentimentDetectionJobsResponse'$sel:sentimentDetectionJobPropertiesList:ListSentimentDetectionJobsResponse'3$sel:httpStatus:ListSentimentDetectionJobsResponse'ListSentimentDetectionJobsListSentimentDetectionJobs'($sel:filter':ListSentimentDetectionJobs'+$sel:maxResults:ListSentimentDetectionJobs'*$sel:nextToken:ListSentimentDetectionJobs'newListSentimentDetectionJobs!listSentimentDetectionJobs_filter%listSentimentDetectionJobs_maxResults$listSentimentDetectionJobs_nextToken%newListSentimentDetectionJobsResponse,listSentimentDetectionJobsResponse_nextTokenlistSentimentDetectionJobsResponse_sentimentDetectionJobPropertiesList-listSentimentDetectionJobsResponse_httpStatus#$fToQueryListSentimentDetectionJobs"$fToPathListSentimentDetectionJobs"$fToJSONListSentimentDetectionJobs%$fToHeadersListSentimentDetectionJobs"$fNFDataListSentimentDetectionJobs$$fHashableListSentimentDetectionJobs$$fAWSPagerListSentimentDetectionJobs*$fNFDataListSentimentDetectionJobsResponse&$fAWSRequestListSentimentDetectionJobs&$fEqListSentimentDetectionJobsResponse($fReadListSentimentDetectionJobsResponse($fShowListSentimentDetectionJobsResponse+$fGenericListSentimentDetectionJobsResponse$fEqListSentimentDetectionJobs $fReadListSentimentDetectionJobs $fShowListSentimentDetectionJobs#$fGenericListSentimentDetectionJobs$ListPiiEntitiesDetectionJobsResponse%ListPiiEntitiesDetectionJobsResponse'4$sel:nextToken:ListPiiEntitiesDetectionJobsResponse'$sel:piiEntitiesDetectionJobPropertiesList:ListPiiEntitiesDetectionJobsResponse'5$sel:httpStatus:ListPiiEntitiesDetectionJobsResponse'ListPiiEntitiesDetectionJobsListPiiEntitiesDetectionJobs'*$sel:filter':ListPiiEntitiesDetectionJobs'-$sel:maxResults:ListPiiEntitiesDetectionJobs',$sel:nextToken:ListPiiEntitiesDetectionJobs'newListPiiEntitiesDetectionJobs#listPiiEntitiesDetectionJobs_filter'listPiiEntitiesDetectionJobs_maxResults&listPiiEntitiesDetectionJobs_nextToken'newListPiiEntitiesDetectionJobsResponse.listPiiEntitiesDetectionJobsResponse_nextTokenlistPiiEntitiesDetectionJobsResponse_piiEntitiesDetectionJobPropertiesList/listPiiEntitiesDetectionJobsResponse_httpStatus%$fToQueryListPiiEntitiesDetectionJobs$$fToPathListPiiEntitiesDetectionJobs$$fToJSONListPiiEntitiesDetectionJobs'$fToHeadersListPiiEntitiesDetectionJobs$$fNFDataListPiiEntitiesDetectionJobs&$fHashableListPiiEntitiesDetectionJobs&$fAWSPagerListPiiEntitiesDetectionJobs,$fNFDataListPiiEntitiesDetectionJobsResponse($fAWSRequestListPiiEntitiesDetectionJobs($fEqListPiiEntitiesDetectionJobsResponse*$fReadListPiiEntitiesDetectionJobsResponse*$fShowListPiiEntitiesDetectionJobsResponse-$fGenericListPiiEntitiesDetectionJobsResponse $fEqListPiiEntitiesDetectionJobs"$fReadListPiiEntitiesDetectionJobs"$fShowListPiiEntitiesDetectionJobs%$fGenericListPiiEntitiesDetectionJobs#ListKeyPhrasesDetectionJobsResponse$ListKeyPhrasesDetectionJobsResponse'$sel:keyPhrasesDetectionJobPropertiesList:ListKeyPhrasesDetectionJobsResponse'3$sel:nextToken:ListKeyPhrasesDetectionJobsResponse'4$sel:httpStatus:ListKeyPhrasesDetectionJobsResponse'ListKeyPhrasesDetectionJobsListKeyPhrasesDetectionJobs')$sel:filter':ListKeyPhrasesDetectionJobs',$sel:maxResults:ListKeyPhrasesDetectionJobs'+$sel:nextToken:ListKeyPhrasesDetectionJobs'newListKeyPhrasesDetectionJobs"listKeyPhrasesDetectionJobs_filter&listKeyPhrasesDetectionJobs_maxResults%listKeyPhrasesDetectionJobs_nextToken&newListKeyPhrasesDetectionJobsResponselistKeyPhrasesDetectionJobsResponse_keyPhrasesDetectionJobPropertiesList-listKeyPhrasesDetectionJobsResponse_nextToken.listKeyPhrasesDetectionJobsResponse_httpStatus$$fToQueryListKeyPhrasesDetectionJobs#$fToPathListKeyPhrasesDetectionJobs#$fToJSONListKeyPhrasesDetectionJobs&$fToHeadersListKeyPhrasesDetectionJobs#$fNFDataListKeyPhrasesDetectionJobs%$fHashableListKeyPhrasesDetectionJobs%$fAWSPagerListKeyPhrasesDetectionJobs+$fNFDataListKeyPhrasesDetectionJobsResponse'$fAWSRequestListKeyPhrasesDetectionJobs'$fEqListKeyPhrasesDetectionJobsResponse)$fReadListKeyPhrasesDetectionJobsResponse)$fShowListKeyPhrasesDetectionJobsResponse,$fGenericListKeyPhrasesDetectionJobsResponse$fEqListKeyPhrasesDetectionJobs!$fReadListKeyPhrasesDetectionJobs!$fShowListKeyPhrasesDetectionJobs$$fGenericListKeyPhrasesDetectionJobsListEventsDetectionJobsResponse ListEventsDetectionJobsResponse'$sel:eventsDetectionJobPropertiesList:ListEventsDetectionJobsResponse'/$sel:nextToken:ListEventsDetectionJobsResponse'0$sel:httpStatus:ListEventsDetectionJobsResponse'ListEventsDetectionJobsListEventsDetectionJobs'%$sel:filter':ListEventsDetectionJobs'($sel:maxResults:ListEventsDetectionJobs''$sel:nextToken:ListEventsDetectionJobs'newListEventsDetectionJobslistEventsDetectionJobs_filter"listEventsDetectionJobs_maxResults!listEventsDetectionJobs_nextToken"newListEventsDetectionJobsResponselistEventsDetectionJobsResponse_eventsDetectionJobPropertiesList)listEventsDetectionJobsResponse_nextToken*listEventsDetectionJobsResponse_httpStatus $fToQueryListEventsDetectionJobs$fToPathListEventsDetectionJobs$fToJSONListEventsDetectionJobs"$fToHeadersListEventsDetectionJobs$fNFDataListEventsDetectionJobs!$fHashableListEventsDetectionJobs'$fNFDataListEventsDetectionJobsResponse#$fAWSRequestListEventsDetectionJobs#$fEqListEventsDetectionJobsResponse%$fReadListEventsDetectionJobsResponse%$fShowListEventsDetectionJobsResponse($fGenericListEventsDetectionJobsResponse$fEqListEventsDetectionJobs$fReadListEventsDetectionJobs$fShowListEventsDetectionJobs $fGenericListEventsDetectionJobsListEntityRecognizersResponseListEntityRecognizersResponse'$sel:entityRecognizerPropertiesList:ListEntityRecognizersResponse'-$sel:nextToken:ListEntityRecognizersResponse'.$sel:httpStatus:ListEntityRecognizersResponse'ListEntityRecognizersListEntityRecognizers'#$sel:filter':ListEntityRecognizers'&$sel:maxResults:ListEntityRecognizers'%$sel:nextToken:ListEntityRecognizers'newListEntityRecognizerslistEntityRecognizers_filter listEntityRecognizers_maxResultslistEntityRecognizers_nextToken newListEntityRecognizersResponse$sel:httpStatus:DescribeTargetedSentimentDetectionJobResponse'%DescribeTargetedSentimentDetectionJob&DescribeTargetedSentimentDetectionJob'1$sel:jobId:DescribeTargetedSentimentDetectionJob'(newDescribeTargetedSentimentDetectionJob+describeTargetedSentimentDetectionJob_jobId0newDescribeTargetedSentimentDetectionJobResponsedescribeTargetedSentimentDetectionJobResponse_targetedSentimentDetectionJobProperties8describeTargetedSentimentDetectionJobResponse_httpStatus.$fToQueryDescribeTargetedSentimentDetectionJob-$fToPathDescribeTargetedSentimentDetectionJob-$fToJSONDescribeTargetedSentimentDetectionJob0$fToHeadersDescribeTargetedSentimentDetectionJob-$fNFDataDescribeTargetedSentimentDetectionJob/$fHashableDescribeTargetedSentimentDetectionJob5$fNFDataDescribeTargetedSentimentDetectionJobResponse1$fAWSRequestDescribeTargetedSentimentDetectionJob1$fEqDescribeTargetedSentimentDetectionJobResponse3$fReadDescribeTargetedSentimentDetectionJobResponse3$fShowDescribeTargetedSentimentDetectionJobResponse6$fGenericDescribeTargetedSentimentDetectionJobResponse)$fEqDescribeTargetedSentimentDetectionJob+$fReadDescribeTargetedSentimentDetectionJob+$fShowDescribeTargetedSentimentDetectionJob.$fGenericDescribeTargetedSentimentDetectionJob%DescribeSentimentDetectionJobResponse&DescribeSentimentDetectionJobResponse'$sel:sentimentDetectionJobProperties:DescribeSentimentDetectionJobResponse'6$sel:httpStatus:DescribeSentimentDetectionJobResponse'DescribeSentimentDetectionJobDescribeSentimentDetectionJob')$sel:jobId:DescribeSentimentDetectionJob' newDescribeSentimentDetectionJob#describeSentimentDetectionJob_jobId(newDescribeSentimentDetectionJobResponsedescribeSentimentDetectionJobResponse_sentimentDetectionJobProperties0describeSentimentDetectionJobResponse_httpStatus&$fToQueryDescribeSentimentDetectionJob%$fToPathDescribeSentimentDetectionJob%$fToJSONDescribeSentimentDetectionJob($fToHeadersDescribeSentimentDetectionJob%$fNFDataDescribeSentimentDetectionJob'$fHashableDescribeSentimentDetectionJob-$fNFDataDescribeSentimentDetectionJobResponse)$fAWSRequestDescribeSentimentDetectionJob)$fEqDescribeSentimentDetectionJobResponse+$fReadDescribeSentimentDetectionJobResponse+$fShowDescribeSentimentDetectionJobResponse.$fGenericDescribeSentimentDetectionJobResponse!$fEqDescribeSentimentDetectionJob#$fReadDescribeSentimentDetectionJob#$fShowDescribeSentimentDetectionJob&$fGenericDescribeSentimentDetectionJobDescribeResourcePolicyResponseDescribeResourcePolicyResponse'1$sel:creationTime:DescribeResourcePolicyResponse'5$sel:lastModifiedTime:DescribeResourcePolicyResponse'5$sel:policyRevisionId:DescribeResourcePolicyResponse'3$sel:resourcePolicy:DescribeResourcePolicyResponse'/$sel:httpStatus:DescribeResourcePolicyResponse'DescribeResourcePolicyDescribeResourcePolicy'($sel:resourceArn:DescribeResourcePolicy'newDescribeResourcePolicy"describeResourcePolicy_resourceArn!newDescribeResourcePolicyResponse+describeResourcePolicyResponse_creationTime/describeResourcePolicyResponse_lastModifiedTime/describeResourcePolicyResponse_policyRevisionId-describeResourcePolicyResponse_resourcePolicy)describeResourcePolicyResponse_httpStatus$fToQueryDescribeResourcePolicy$fToPathDescribeResourcePolicy$fToJSONDescribeResourcePolicy!$fToHeadersDescribeResourcePolicy$fNFDataDescribeResourcePolicy $fHashableDescribeResourcePolicy&$fNFDataDescribeResourcePolicyResponse"$fAWSRequestDescribeResourcePolicy"$fEqDescribeResourcePolicyResponse$$fReadDescribeResourcePolicyResponse$$fShowDescribeResourcePolicyResponse'$fGenericDescribeResourcePolicyResponse$fEqDescribeResourcePolicy$fReadDescribeResourcePolicy$fShowDescribeResourcePolicy$fGenericDescribeResourcePolicy'DescribePiiEntitiesDetectionJobResponse(DescribePiiEntitiesDetectionJobResponse'$sel:piiEntitiesDetectionJobProperties:DescribePiiEntitiesDetectionJobResponse'8$sel:httpStatus:DescribePiiEntitiesDetectionJobResponse'DescribePiiEntitiesDetectionJob DescribePiiEntitiesDetectionJob'+$sel:jobId:DescribePiiEntitiesDetectionJob'"newDescribePiiEntitiesDetectionJob%describePiiEntitiesDetectionJob_jobId*newDescribePiiEntitiesDetectionJobResponsedescribePiiEntitiesDetectionJobResponse_piiEntitiesDetectionJobProperties2describePiiEntitiesDetectionJobResponse_httpStatus($fToQueryDescribePiiEntitiesDetectionJob'$fToPathDescribePiiEntitiesDetectionJob'$fToJSONDescribePiiEntitiesDetectionJob*$fToHeadersDescribePiiEntitiesDetectionJob'$fNFDataDescribePiiEntitiesDetectionJob)$fHashableDescribePiiEntitiesDetectionJob/$fNFDataDescribePiiEntitiesDetectionJobResponse+$fAWSRequestDescribePiiEntitiesDetectionJob+$fEqDescribePiiEntitiesDetectionJobResponse-$fReadDescribePiiEntitiesDetectionJobResponse-$fShowDescribePiiEntitiesDetectionJobResponse0$fGenericDescribePiiEntitiesDetectionJobResponse#$fEqDescribePiiEntitiesDetectionJob%$fReadDescribePiiEntitiesDetectionJob%$fShowDescribePiiEntitiesDetectionJob($fGenericDescribePiiEntitiesDetectionJob&DescribeKeyPhrasesDetectionJobResponse'DescribeKeyPhrasesDetectionJobResponse'$sel:keyPhrasesDetectionJobProperties:DescribeKeyPhrasesDetectionJobResponse'7$sel:httpStatus:DescribeKeyPhrasesDetectionJobResponse'DescribeKeyPhrasesDetectionJobDescribeKeyPhrasesDetectionJob'*$sel:jobId:DescribeKeyPhrasesDetectionJob'!newDescribeKeyPhrasesDetectionJob$describeKeyPhrasesDetectionJob_jobId)newDescribeKeyPhrasesDetectionJobResponsedescribeKeyPhrasesDetectionJobResponse_keyPhrasesDetectionJobProperties1describeKeyPhrasesDetectionJobResponse_httpStatus'$fToQueryDescribeKeyPhrasesDetectionJob&$fToPathDescribeKeyPhrasesDetectionJob&$fToJSONDescribeKeyPhrasesDetectionJob)$fToHeadersDescribeKeyPhrasesDetectionJob&$fNFDataDescribeKeyPhrasesDetectionJob($fHashableDescribeKeyPhrasesDetectionJob.$fNFDataDescribeKeyPhrasesDetectionJobResponse*$fAWSRequestDescribeKeyPhrasesDetectionJob*$fEqDescribeKeyPhrasesDetectionJobResponse,$fReadDescribeKeyPhrasesDetectionJobResponse,$fShowDescribeKeyPhrasesDetectionJobResponse/$fGenericDescribeKeyPhrasesDetectionJobResponse"$fEqDescribeKeyPhrasesDetectionJob$$fReadDescribeKeyPhrasesDetectionJob$$fShowDescribeKeyPhrasesDetectionJob'$fGenericDescribeKeyPhrasesDetectionJob"DescribeEventsDetectionJobResponse#DescribeEventsDetectionJobResponse'$sel:eventsDetectionJobProperties:DescribeEventsDetectionJobResponse'3$sel:httpStatus:DescribeEventsDetectionJobResponse'DescribeEventsDetectionJobDescribeEventsDetectionJob'&$sel:jobId:DescribeEventsDetectionJob'newDescribeEventsDetectionJob describeEventsDetectionJob_jobId%newDescribeEventsDetectionJobResponse?describeEventsDetectionJobResponse_eventsDetectionJobProperties-describeEventsDetectionJobResponse_httpStatus#$fToQueryDescribeEventsDetectionJob"$fToPathDescribeEventsDetectionJob"$fToJSONDescribeEventsDetectionJob%$fToHeadersDescribeEventsDetectionJob"$fNFDataDescribeEventsDetectionJob$$fHashableDescribeEventsDetectionJob*$fNFDataDescribeEventsDetectionJobResponse&$fAWSRequestDescribeEventsDetectionJob&$fEqDescribeEventsDetectionJobResponse($fReadDescribeEventsDetectionJobResponse($fShowDescribeEventsDetectionJobResponse+$fGenericDescribeEventsDetectionJobResponse$fEqDescribeEventsDetectionJob $fReadDescribeEventsDetectionJob $fShowDescribeEventsDetectionJob#$fGenericDescribeEventsDetectionJob DescribeEntityRecognizerResponse!DescribeEntityRecognizerResponse'$sel:entityRecognizerProperties:DescribeEntityRecognizerResponse'1$sel:httpStatus:DescribeEntityRecognizerResponse'DescribeEntityRecognizerDescribeEntityRecognizer'2$sel:entityRecognizerArn:DescribeEntityRecognizer'newDescribeEntityRecognizer,describeEntityRecognizer_entityRecognizerArn#newDescribeEntityRecognizerResponse;describeEntityRecognizerResponse_entityRecognizerProperties+describeEntityRecognizerResponse_httpStatus!$fToQueryDescribeEntityRecognizer $fToPathDescribeEntityRecognizer $fToJSONDescribeEntityRecognizer#$fToHeadersDescribeEntityRecognizer $fNFDataDescribeEntityRecognizer"$fHashableDescribeEntityRecognizer($fNFDataDescribeEntityRecognizerResponse$$fAWSRequestDescribeEntityRecognizer$$fEqDescribeEntityRecognizerResponse&$fShowDescribeEntityRecognizerResponse)$fGenericDescribeEntityRecognizerResponse$fEqDescribeEntityRecognizer$fReadDescribeEntityRecognizer$fShowDescribeEntityRecognizer!$fGenericDescribeEntityRecognizer$DescribeEntitiesDetectionJobResponse%DescribeEntitiesDetectionJobResponse'$sel:entitiesDetectionJobProperties:DescribeEntitiesDetectionJobResponse'5$sel:httpStatus:DescribeEntitiesDetectionJobResponse'DescribeEntitiesDetectionJobDescribeEntitiesDetectionJob'($sel:jobId:DescribeEntitiesDetectionJob'newDescribeEntitiesDetectionJob"describeEntitiesDetectionJob_jobId'newDescribeEntitiesDetectionJobResponsedescribeEntitiesDetectionJobResponse_entitiesDetectionJobProperties/describeEntitiesDetectionJobResponse_httpStatus%$fToQueryDescribeEntitiesDetectionJob$$fToPathDescribeEntitiesDetectionJob$$fToJSONDescribeEntitiesDetectionJob'$fToHeadersDescribeEntitiesDetectionJob$$fNFDataDescribeEntitiesDetectionJob&$fHashableDescribeEntitiesDetectionJob,$fNFDataDescribeEntitiesDetectionJobResponse($fAWSRequestDescribeEntitiesDetectionJob($fEqDescribeEntitiesDetectionJobResponse*$fReadDescribeEntitiesDetectionJobResponse*$fShowDescribeEntitiesDetectionJobResponse-$fGenericDescribeEntitiesDetectionJobResponse $fEqDescribeEntitiesDetectionJob"$fReadDescribeEntitiesDetectionJob"$fShowDescribeEntitiesDetectionJob%$fGenericDescribeEntitiesDetectionJobDescribeEndpointResponseDescribeEndpointResponse'1$sel:endpointProperties:DescribeEndpointResponse')$sel:httpStatus:DescribeEndpointResponse'DescribeEndpointDescribeEndpoint'"$sel:endpointArn:DescribeEndpoint'newDescribeEndpointdescribeEndpoint_endpointArnnewDescribeEndpointResponse+describeEndpointResponse_endpointProperties#describeEndpointResponse_httpStatus$fToQueryDescribeEndpoint$fToPathDescribeEndpoint$fToJSONDescribeEndpoint$fToHeadersDescribeEndpoint$fNFDataDescribeEndpoint$fHashableDescribeEndpoint $fNFDataDescribeEndpointResponse$fAWSRequestDescribeEndpoint$fEqDescribeEndpointResponse$fReadDescribeEndpointResponse$fShowDescribeEndpointResponse!$fGenericDescribeEndpointResponse$fEqDescribeEndpoint$fReadDescribeEndpoint$fShowDescribeEndpoint$fGenericDescribeEndpoint,DescribeDominantLanguageDetectionJobResponse-DescribeDominantLanguageDetectionJobResponse'$sel:dominantLanguageDetectionJobProperties:DescribeDominantLanguageDetectionJobResponse'=$sel:httpStatus:DescribeDominantLanguageDetectionJobResponse'$DescribeDominantLanguageDetectionJob%DescribeDominantLanguageDetectionJob'0$sel:jobId:DescribeDominantLanguageDetectionJob''newDescribeDominantLanguageDetectionJob*describeDominantLanguageDetectionJob_jobId/newDescribeDominantLanguageDetectionJobResponsedescribeDominantLanguageDetectionJobResponse_dominantLanguageDetectionJobProperties7describeDominantLanguageDetectionJobResponse_httpStatus-$fToQueryDescribeDominantLanguageDetectionJob,$fToPathDescribeDominantLanguageDetectionJob,$fToJSONDescribeDominantLanguageDetectionJob/$fToHeadersDescribeDominantLanguageDetectionJob,$fNFDataDescribeDominantLanguageDetectionJob.$fHashableDescribeDominantLanguageDetectionJob4$fNFDataDescribeDominantLanguageDetectionJobResponse0$fAWSRequestDescribeDominantLanguageDetectionJob0$fEqDescribeDominantLanguageDetectionJobResponse2$fReadDescribeDominantLanguageDetectionJobResponse2$fShowDescribeDominantLanguageDetectionJobResponse5$fGenericDescribeDominantLanguageDetectionJobResponse($fEqDescribeDominantLanguageDetectionJob*$fReadDescribeDominantLanguageDetectionJob*$fShowDescribeDominantLanguageDetectionJob-$fGenericDescribeDominantLanguageDetectionJob"DescribeDocumentClassifierResponse#DescribeDocumentClassifierResponse'$sel:documentClassifierProperties:DescribeDocumentClassifierResponse'3$sel:httpStatus:DescribeDocumentClassifierResponse'DescribeDocumentClassifierDescribeDocumentClassifier'6$sel:documentClassifierArn:DescribeDocumentClassifier'newDescribeDocumentClassifier0describeDocumentClassifier_documentClassifierArn%newDescribeDocumentClassifierResponse?describeDocumentClassifierResponse_documentClassifierProperties-describeDocumentClassifierResponse_httpStatus#$fToQueryDescribeDocumentClassifier"$fToPathDescribeDocumentClassifier"$fToJSONDescribeDocumentClassifier%$fToHeadersDescribeDocumentClassifier"$fNFDataDescribeDocumentClassifier$$fHashableDescribeDocumentClassifier*$fNFDataDescribeDocumentClassifierResponse&$fAWSRequestDescribeDocumentClassifier&$fEqDescribeDocumentClassifierResponse($fShowDescribeDocumentClassifierResponse+$fGenericDescribeDocumentClassifierResponse$fEqDescribeDocumentClassifier $fReadDescribeDocumentClassifier $fShowDescribeDocumentClassifier#$fGenericDescribeDocumentClassifier)DescribeDocumentClassificationJobResponse*DescribeDocumentClassificationJobResponse'$sel:documentClassificationJobProperties:DescribeDocumentClassificationJobResponse':$sel:httpStatus:DescribeDocumentClassificationJobResponse'!DescribeDocumentClassificationJob"DescribeDocumentClassificationJob'-$sel:jobId:DescribeDocumentClassificationJob'$newDescribeDocumentClassificationJob'describeDocumentClassificationJob_jobId,newDescribeDocumentClassificationJobResponsedescribeDocumentClassificationJobResponse_documentClassificationJobProperties4describeDocumentClassificationJobResponse_httpStatus*$fToQueryDescribeDocumentClassificationJob)$fToPathDescribeDocumentClassificationJob)$fToJSONDescribeDocumentClassificationJob,$fToHeadersDescribeDocumentClassificationJob)$fNFDataDescribeDocumentClassificationJob+$fHashableDescribeDocumentClassificationJob1$fNFDataDescribeDocumentClassificationJobResponse-$fAWSRequestDescribeDocumentClassificationJob-$fEqDescribeDocumentClassificationJobResponse/$fReadDescribeDocumentClassificationJobResponse/$fShowDescribeDocumentClassificationJobResponse2$fGenericDescribeDocumentClassificationJobResponse%$fEqDescribeDocumentClassificationJob'$fReadDescribeDocumentClassificationJob'$fShowDescribeDocumentClassificationJob*$fGenericDescribeDocumentClassificationJobDeleteResourcePolicyResponseDeleteResourcePolicyResponse'-$sel:httpStatus:DeleteResourcePolicyResponse'DeleteResourcePolicyDeleteResourcePolicy'+$sel:policyRevisionId:DeleteResourcePolicy'&$sel:resourceArn:DeleteResourcePolicy'newDeleteResourcePolicy%deleteResourcePolicy_policyRevisionId deleteResourcePolicy_resourceArnnewDeleteResourcePolicyResponse'deleteResourcePolicyResponse_httpStatus$fToQueryDeleteResourcePolicy$fToPathDeleteResourcePolicy$fToJSONDeleteResourcePolicy$fToHeadersDeleteResourcePolicy$fNFDataDeleteResourcePolicy$fHashableDeleteResourcePolicy$$fNFDataDeleteResourcePolicyResponse $fAWSRequestDeleteResourcePolicy $fEqDeleteResourcePolicyResponse"$fReadDeleteResourcePolicyResponse"$fShowDeleteResourcePolicyResponse%$fGenericDeleteResourcePolicyResponse$fEqDeleteResourcePolicy$fReadDeleteResourcePolicy$fShowDeleteResourcePolicy$fGenericDeleteResourcePolicyDeleteEntityRecognizerResponseDeleteEntityRecognizerResponse'/$sel:httpStatus:DeleteEntityRecognizerResponse'DeleteEntityRecognizerDeleteEntityRecognizer'0$sel:entityRecognizerArn:DeleteEntityRecognizer'newDeleteEntityRecognizer*deleteEntityRecognizer_entityRecognizerArn!newDeleteEntityRecognizerResponse)deleteEntityRecognizerResponse_httpStatus$fToQueryDeleteEntityRecognizer$fToPathDeleteEntityRecognizer$fToJSONDeleteEntityRecognizer!$fToHeadersDeleteEntityRecognizer$fNFDataDeleteEntityRecognizer $fHashableDeleteEntityRecognizer&$fNFDataDeleteEntityRecognizerResponse"$fAWSRequestDeleteEntityRecognizer"$fEqDeleteEntityRecognizerResponse$$fReadDeleteEntityRecognizerResponse$$fShowDeleteEntityRecognizerResponse'$fGenericDeleteEntityRecognizerResponse$fEqDeleteEntityRecognizer$fReadDeleteEntityRecognizer$fShowDeleteEntityRecognizer$fGenericDeleteEntityRecognizerDeleteEndpointResponseDeleteEndpointResponse''$sel:httpStatus:DeleteEndpointResponse'DeleteEndpointDeleteEndpoint' $sel:endpointArn:DeleteEndpoint'newDeleteEndpointdeleteEndpoint_endpointArnnewDeleteEndpointResponse!deleteEndpointResponse_httpStatus$fToQueryDeleteEndpoint$fToPathDeleteEndpoint$fToJSONDeleteEndpoint$fToHeadersDeleteEndpoint$fNFDataDeleteEndpoint$fHashableDeleteEndpoint$fNFDataDeleteEndpointResponse$fAWSRequestDeleteEndpoint$fEqDeleteEndpointResponse$fReadDeleteEndpointResponse$fShowDeleteEndpointResponse$fGenericDeleteEndpointResponse$fEqDeleteEndpoint$fReadDeleteEndpoint$fShowDeleteEndpoint$fGenericDeleteEndpoint DeleteDocumentClassifierResponse!DeleteDocumentClassifierResponse'1$sel:httpStatus:DeleteDocumentClassifierResponse'DeleteDocumentClassifierDeleteDocumentClassifier'4$sel:documentClassifierArn:DeleteDocumentClassifier'newDeleteDocumentClassifier.deleteDocumentClassifier_documentClassifierArn#newDeleteDocumentClassifierResponse+deleteDocumentClassifierResponse_httpStatus!$fToQueryDeleteDocumentClassifier $fToPathDeleteDocumentClassifier $fToJSONDeleteDocumentClassifier#$fToHeadersDeleteDocumentClassifier $fNFDataDeleteDocumentClassifier"$fHashableDeleteDocumentClassifier($fNFDataDeleteDocumentClassifierResponse$$fAWSRequestDeleteDocumentClassifier$$fEqDeleteDocumentClassifierResponse&$fReadDeleteDocumentClassifierResponse&$fShowDeleteDocumentClassifierResponse)$fGenericDeleteDocumentClassifierResponse$fEqDeleteDocumentClassifier$fReadDeleteDocumentClassifier$fShowDeleteDocumentClassifier!$fGenericDeleteDocumentClassifierCreateEntityRecognizerResponseCreateEntityRecognizerResponse'8$sel:entityRecognizerArn:CreateEntityRecognizerResponse'/$sel:httpStatus:CreateEntityRecognizerResponse'CreateEntityRecognizerCreateEntityRecognizer'/$sel:clientRequestToken:CreateEntityRecognizer'*$sel:modelKmsKeyId:CreateEntityRecognizer'($sel:modelPolicy:CreateEntityRecognizer'!$sel:tags:CreateEntityRecognizer'($sel:versionName:CreateEntityRecognizer'+$sel:volumeKmsKeyId:CreateEntityRecognizer'&$sel:vpcConfig:CreateEntityRecognizer'+$sel:recognizerName:CreateEntityRecognizer'.$sel:dataAccessRoleArn:CreateEntityRecognizer',$sel:inputDataConfig:CreateEntityRecognizer')$sel:languageCode:CreateEntityRecognizer'newCreateEntityRecognizer)createEntityRecognizer_clientRequestToken$createEntityRecognizer_modelKmsKeyId"createEntityRecognizer_modelPolicycreateEntityRecognizer_tags"createEntityRecognizer_versionName%createEntityRecognizer_volumeKmsKeyId createEntityRecognizer_vpcConfig%createEntityRecognizer_recognizerName(createEntityRecognizer_dataAccessRoleArn&createEntityRecognizer_inputDataConfig#createEntityRecognizer_languageCode!newCreateEntityRecognizerResponse2createEntityRecognizerResponse_entityRecognizerArn)createEntityRecognizerResponse_httpStatus$fToQueryCreateEntityRecognizer$fToPathCreateEntityRecognizer$fToJSONCreateEntityRecognizer!$fToHeadersCreateEntityRecognizer$fNFDataCreateEntityRecognizer $fHashableCreateEntityRecognizer&$fNFDataCreateEntityRecognizerResponse"$fAWSRequestCreateEntityRecognizer"$fEqCreateEntityRecognizerResponse$$fReadCreateEntityRecognizerResponse$$fShowCreateEntityRecognizerResponse'$fGenericCreateEntityRecognizerResponse$fEqCreateEntityRecognizer$fReadCreateEntityRecognizer$fShowCreateEntityRecognizer$fGenericCreateEntityRecognizerCreateEndpointResponseCreateEndpointResponse'($sel:endpointArn:CreateEndpointResponse''$sel:httpStatus:CreateEndpointResponse'CreateEndpointCreateEndpoint''$sel:clientRequestToken:CreateEndpoint'&$sel:dataAccessRoleArn:CreateEndpoint'$sel:tags:CreateEndpoint'!$sel:endpointName:CreateEndpoint'$sel:modelArn:CreateEndpoint'*$sel:desiredInferenceUnits:CreateEndpoint'newCreateEndpoint!createEndpoint_clientRequestToken createEndpoint_dataAccessRoleArncreateEndpoint_tagscreateEndpoint_endpointNamecreateEndpoint_modelArn$createEndpoint_desiredInferenceUnitsnewCreateEndpointResponse"createEndpointResponse_endpointArn!createEndpointResponse_httpStatus$fToQueryCreateEndpoint$fToPathCreateEndpoint$fToJSONCreateEndpoint$fToHeadersCreateEndpoint$fNFDataCreateEndpoint$fHashableCreateEndpoint$fNFDataCreateEndpointResponse$fAWSRequestCreateEndpoint$fEqCreateEndpointResponse$fReadCreateEndpointResponse$fShowCreateEndpointResponse$fGenericCreateEndpointResponse$fEqCreateEndpoint$fReadCreateEndpoint$fShowCreateEndpoint$fGenericCreateEndpoint CreateDocumentClassifierResponse!CreateDocumentClassifierResponse'<$sel:documentClassifierArn:CreateDocumentClassifierResponse'1$sel:httpStatus:CreateDocumentClassifierResponse'CreateDocumentClassifierCreateDocumentClassifier'1$sel:clientRequestToken:CreateDocumentClassifier'#$sel:mode:CreateDocumentClassifier',$sel:modelKmsKeyId:CreateDocumentClassifier'*$sel:modelPolicy:CreateDocumentClassifier'/$sel:outputDataConfig:CreateDocumentClassifier'#$sel:tags:CreateDocumentClassifier'*$sel:versionName:CreateDocumentClassifier'-$sel:volumeKmsKeyId:CreateDocumentClassifier'($sel:vpcConfig:CreateDocumentClassifier'5$sel:documentClassifierName:CreateDocumentClassifier'0$sel:dataAccessRoleArn:CreateDocumentClassifier'.$sel:inputDataConfig:CreateDocumentClassifier'+$sel:languageCode:CreateDocumentClassifier'newCreateDocumentClassifier+createDocumentClassifier_clientRequestTokencreateDocumentClassifier_mode&createDocumentClassifier_modelKmsKeyId$createDocumentClassifier_modelPolicy)createDocumentClassifier_outputDataConfigcreateDocumentClassifier_tags$createDocumentClassifier_versionName'createDocumentClassifier_volumeKmsKeyId"createDocumentClassifier_vpcConfig/createDocumentClassifier_documentClassifierName*createDocumentClassifier_dataAccessRoleArn(createDocumentClassifier_inputDataConfig%createDocumentClassifier_languageCode#newCreateDocumentClassifierResponse6createDocumentClassifierResponse_documentClassifierArn+createDocumentClassifierResponse_httpStatus!$fToQueryCreateDocumentClassifier $fToPathCreateDocumentClassifier $fToJSONCreateDocumentClassifier#$fToHeadersCreateDocumentClassifier $fNFDataCreateDocumentClassifier"$fHashableCreateDocumentClassifier($fNFDataCreateDocumentClassifierResponse$$fAWSRequestCreateDocumentClassifier$$fEqCreateDocumentClassifierResponse&$fReadCreateDocumentClassifierResponse&$fShowCreateDocumentClassifierResponse)$fGenericCreateDocumentClassifierResponse$fEqCreateDocumentClassifier$fReadCreateDocumentClassifier$fShowCreateDocumentClassifier!$fGenericCreateDocumentClassifierContainsPiiEntitiesResponseContainsPiiEntitiesResponse'($sel:labels:ContainsPiiEntitiesResponse',$sel:httpStatus:ContainsPiiEntitiesResponse'ContainsPiiEntitiesContainsPiiEntities'$sel:text:ContainsPiiEntities'&$sel:languageCode:ContainsPiiEntities'newContainsPiiEntitiescontainsPiiEntities_text containsPiiEntities_languageCodenewContainsPiiEntitiesResponse"containsPiiEntitiesResponse_labels&containsPiiEntitiesResponse_httpStatus$fToQueryContainsPiiEntities$fToPathContainsPiiEntities$fToJSONContainsPiiEntities$fToHeadersContainsPiiEntities$fNFDataContainsPiiEntities$fHashableContainsPiiEntities#$fNFDataContainsPiiEntitiesResponse$fAWSRequestContainsPiiEntities$fEqContainsPiiEntitiesResponse!$fReadContainsPiiEntitiesResponse!$fShowContainsPiiEntitiesResponse$$fGenericContainsPiiEntitiesResponse$fEqContainsPiiEntities$fReadContainsPiiEntities$fShowContainsPiiEntities$fGenericContainsPiiEntitiesClassifyDocumentResponseClassifyDocumentResponse'&$sel:classes:ClassifyDocumentResponse'/$sel:documentMetadata:ClassifyDocumentResponse'+$sel:documentType:ClassifyDocumentResponse'%$sel:errors:ClassifyDocumentResponse'%$sel:labels:ClassifyDocumentResponse')$sel:httpStatus:ClassifyDocumentResponse'ClassifyDocumentClassifyDocument'$sel:bytes:ClassifyDocument'+$sel:documentReaderConfig:ClassifyDocument'$sel:text:ClassifyDocument'"$sel:endpointArn:ClassifyDocument'newClassifyDocumentclassifyDocument_bytes%classifyDocument_documentReaderConfigclassifyDocument_textclassifyDocument_endpointArnnewClassifyDocumentResponse classifyDocumentResponse_classes)classifyDocumentResponse_documentMetadata%classifyDocumentResponse_documentTypeclassifyDocumentResponse_errorsclassifyDocumentResponse_labels#classifyDocumentResponse_httpStatus$fToQueryClassifyDocument$fToPathClassifyDocument$fToJSONClassifyDocument$fToHeadersClassifyDocument$fNFDataClassifyDocument$fHashableClassifyDocument $fNFDataClassifyDocumentResponse$fAWSRequestClassifyDocument$fEqClassifyDocumentResponse$fShowClassifyDocumentResponse!$fGenericClassifyDocumentResponse$fEqClassifyDocument$fShowClassifyDocument$fGenericClassifyDocument$BatchDetectTargetedSentimentResponse%BatchDetectTargetedSentimentResponse'5$sel:httpStatus:BatchDetectTargetedSentimentResponse'5$sel:resultList:BatchDetectTargetedSentimentResponse'4$sel:errorList:BatchDetectTargetedSentimentResponse'BatchDetectTargetedSentimentBatchDetectTargetedSentiment'+$sel:textList:BatchDetectTargetedSentiment'/$sel:languageCode:BatchDetectTargetedSentiment'newBatchDetectTargetedSentiment%batchDetectTargetedSentiment_textList)batchDetectTargetedSentiment_languageCode'newBatchDetectTargetedSentimentResponse/batchDetectTargetedSentimentResponse_httpStatus/batchDetectTargetedSentimentResponse_resultList.batchDetectTargetedSentimentResponse_errorList%$fToQueryBatchDetectTargetedSentiment$$fToPathBatchDetectTargetedSentiment$$fToJSONBatchDetectTargetedSentiment'$fToHeadersBatchDetectTargetedSentiment$$fNFDataBatchDetectTargetedSentiment&$fHashableBatchDetectTargetedSentiment,$fNFDataBatchDetectTargetedSentimentResponse($fAWSRequestBatchDetectTargetedSentiment($fEqBatchDetectTargetedSentimentResponse*$fShowBatchDetectTargetedSentimentResponse-$fGenericBatchDetectTargetedSentimentResponse $fEqBatchDetectTargetedSentiment"$fShowBatchDetectTargetedSentiment%$fGenericBatchDetectTargetedSentimentBatchDetectSyntaxResponseBatchDetectSyntaxResponse'*$sel:httpStatus:BatchDetectSyntaxResponse'*$sel:resultList:BatchDetectSyntaxResponse')$sel:errorList:BatchDetectSyntaxResponse'BatchDetectSyntaxBatchDetectSyntax' $sel:textList:BatchDetectSyntax'$$sel:languageCode:BatchDetectSyntax'newBatchDetectSyntaxbatchDetectSyntax_textListbatchDetectSyntax_languageCodenewBatchDetectSyntaxResponse$batchDetectSyntaxResponse_httpStatus$batchDetectSyntaxResponse_resultList#batchDetectSyntaxResponse_errorList$fToQueryBatchDetectSyntax$fToPathBatchDetectSyntax$fToJSONBatchDetectSyntax$fToHeadersBatchDetectSyntax$fNFDataBatchDetectSyntax$fHashableBatchDetectSyntax!$fNFDataBatchDetectSyntaxResponse$fAWSRequestBatchDetectSyntax$fEqBatchDetectSyntaxResponse$fShowBatchDetectSyntaxResponse"$fGenericBatchDetectSyntaxResponse$fEqBatchDetectSyntax$fShowBatchDetectSyntax$fGenericBatchDetectSyntaxBatchDetectSentimentResponseBatchDetectSentimentResponse'-$sel:httpStatus:BatchDetectSentimentResponse'-$sel:resultList:BatchDetectSentimentResponse',$sel:errorList:BatchDetectSentimentResponse'BatchDetectSentimentBatchDetectSentiment'#$sel:textList:BatchDetectSentiment''$sel:languageCode:BatchDetectSentiment'newBatchDetectSentimentbatchDetectSentiment_textList!batchDetectSentiment_languageCodenewBatchDetectSentimentResponse'batchDetectSentimentResponse_httpStatus'batchDetectSentimentResponse_resultList&batchDetectSentimentResponse_errorList$fToQueryBatchDetectSentiment$fToPathBatchDetectSentiment$fToJSONBatchDetectSentiment$fToHeadersBatchDetectSentiment$fNFDataBatchDetectSentiment$fHashableBatchDetectSentiment$$fNFDataBatchDetectSentimentResponse $fAWSRequestBatchDetectSentiment $fEqBatchDetectSentimentResponse"$fShowBatchDetectSentimentResponse%$fGenericBatchDetectSentimentResponse$fEqBatchDetectSentiment$fShowBatchDetectSentiment$fGenericBatchDetectSentimentBatchDetectKeyPhrasesResponseBatchDetectKeyPhrasesResponse'.$sel:httpStatus:BatchDetectKeyPhrasesResponse'.$sel:resultList:BatchDetectKeyPhrasesResponse'-$sel:errorList:BatchDetectKeyPhrasesResponse'BatchDetectKeyPhrasesBatchDetectKeyPhrases'$$sel:textList:BatchDetectKeyPhrases'($sel:languageCode:BatchDetectKeyPhrases'newBatchDetectKeyPhrasesbatchDetectKeyPhrases_textList"batchDetectKeyPhrases_languageCode newBatchDetectKeyPhrasesResponse(batchDetectKeyPhrasesResponse_httpStatus(batchDetectKeyPhrasesResponse_resultList'batchDetectKeyPhrasesResponse_errorList$fToQueryBatchDetectKeyPhrases$fToPathBatchDetectKeyPhrases$fToJSONBatchDetectKeyPhrases $fToHeadersBatchDetectKeyPhrases$fNFDataBatchDetectKeyPhrases$fHashableBatchDetectKeyPhrases%$fNFDataBatchDetectKeyPhrasesResponse!$fAWSRequestBatchDetectKeyPhrases!$fEqBatchDetectKeyPhrasesResponse#$fShowBatchDetectKeyPhrasesResponse&$fGenericBatchDetectKeyPhrasesResponse$fEqBatchDetectKeyPhrases$fShowBatchDetectKeyPhrases$fGenericBatchDetectKeyPhrasesBatchDetectEntitiesResponseBatchDetectEntitiesResponse',$sel:httpStatus:BatchDetectEntitiesResponse',$sel:resultList:BatchDetectEntitiesResponse'+$sel:errorList:BatchDetectEntitiesResponse'BatchDetectEntitiesBatchDetectEntities'"$sel:textList:BatchDetectEntities'&$sel:languageCode:BatchDetectEntities'newBatchDetectEntitiesbatchDetectEntities_textList batchDetectEntities_languageCodenewBatchDetectEntitiesResponse&batchDetectEntitiesResponse_httpStatus&batchDetectEntitiesResponse_resultList%batchDetectEntitiesResponse_errorList$fToQueryBatchDetectEntities$fToPathBatchDetectEntities$fToJSONBatchDetectEntities$fToHeadersBatchDetectEntities$fNFDataBatchDetectEntities$fHashableBatchDetectEntities#$fNFDataBatchDetectEntitiesResponse$fAWSRequestBatchDetectEntities$fEqBatchDetectEntitiesResponse!$fShowBatchDetectEntitiesResponse$$fGenericBatchDetectEntitiesResponse$fEqBatchDetectEntities$fShowBatchDetectEntities$fGenericBatchDetectEntities#BatchDetectDominantLanguageResponse$BatchDetectDominantLanguageResponse'4$sel:httpStatus:BatchDetectDominantLanguageResponse'4$sel:resultList:BatchDetectDominantLanguageResponse'3$sel:errorList:BatchDetectDominantLanguageResponse'BatchDetectDominantLanguageBatchDetectDominantLanguage'*$sel:textList:BatchDetectDominantLanguage'newBatchDetectDominantLanguage$batchDetectDominantLanguage_textList&newBatchDetectDominantLanguageResponse.batchDetectDominantLanguageResponse_httpStatus.batchDetectDominantLanguageResponse_resultList-batchDetectDominantLanguageResponse_errorList$$fToQueryBatchDetectDominantLanguage#$fToPathBatchDetectDominantLanguage#$fToJSONBatchDetectDominantLanguage&$fToHeadersBatchDetectDominantLanguage#$fNFDataBatchDetectDominantLanguage%$fHashableBatchDetectDominantLanguage+$fNFDataBatchDetectDominantLanguageResponse'$fAWSRequestBatchDetectDominantLanguage'$fEqBatchDetectDominantLanguageResponse)$fShowBatchDetectDominantLanguageResponse,$fGenericBatchDetectDominantLanguageResponse$fEqBatchDetectDominantLanguage!$fShowBatchDetectDominantLanguage$$fGenericBatchDetectDominantLanguageUntagResourceResponseUntagResourceResponse'&$sel:httpStatus:UntagResourceResponse' UntagResourceUntagResource'$sel:resourceArn:UntagResource'$sel:tagKeys:UntagResource'newUntagResourceuntagResource_resourceArnuntagResource_tagKeysnewUntagResourceResponse untagResourceResponse_httpStatus$fToQueryUntagResource$fToPathUntagResource$fToJSONUntagResource$fToHeadersUntagResource$fNFDataUntagResource$fHashableUntagResource$fNFDataUntagResourceResponse$fAWSRequestUntagResource$fEqUntagResourceResponse$fReadUntagResourceResponse$fShowUntagResourceResponse$fGenericUntagResourceResponse$fEqUntagResource$fReadUntagResource$fShowUntagResource$fGenericUntagResourceUpdateEndpointResponseUpdateEndpointResponse''$sel:httpStatus:UpdateEndpointResponse'UpdateEndpointUpdateEndpoint'-$sel:desiredDataAccessRoleArn:UpdateEndpoint'*$sel:desiredInferenceUnits:UpdateEndpoint'$$sel:desiredModelArn:UpdateEndpoint' $sel:endpointArn:UpdateEndpoint'newUpdateEndpoint'updateEndpoint_desiredDataAccessRoleArn$updateEndpoint_desiredInferenceUnitsupdateEndpoint_desiredModelArnupdateEndpoint_endpointArnnewUpdateEndpointResponse!updateEndpointResponse_httpStatus$fToQueryUpdateEndpoint$fToPathUpdateEndpoint$fToJSONUpdateEndpoint$fToHeadersUpdateEndpoint$fNFDataUpdateEndpoint$fHashableUpdateEndpoint$fNFDataUpdateEndpointResponse$fAWSRequestUpdateEndpoint$fEqUpdateEndpointResponse$fReadUpdateEndpointResponse$fShowUpdateEndpointResponse$fGenericUpdateEndpointResponse$fEqUpdateEndpoint$fReadUpdateEndpoint$fShowUpdateEndpoint$fGenericUpdateEndpoint