comprehend {paws}R Documentation

Amazon Comprehend

Description

Amazon Comprehend is an Amazon Web Services service for gaining insight into the content of documents. Use these actions to determine the topics contained in your documents, the topics they discuss, the predominant sentiment expressed in them, the predominant language used, and more.

Usage

comprehend(
  config = list(),
  credentials = list(),
  endpoint = NULL,
  region = NULL
)

Arguments

config

Optional configuration of credentials, endpoint, and/or region.

  • credentials:

    • creds:

      • access_key_id: AWS access key ID

      • secret_access_key: AWS secret access key

      • session_token: AWS temporary session token

    • profile: The name of a profile to use. If not given, then the default profile is used.

    • anonymous: Set anonymous credentials.

  • endpoint: The complete URL to use for the constructed client.

  • region: The AWS Region used in instantiating the client.

  • close_connection: Immediately close all HTTP connections.

  • timeout: The time in seconds till a timeout exception is thrown when attempting to make a connection. The default is 60 seconds.

  • s3_force_path_style: Set this to true to force the request to use path-style addressing, i.e. ⁠http://s3.amazonaws.com/BUCKET/KEY⁠.

  • sts_regional_endpoint: Set sts regional endpoint resolver to regional or legacy https://docs.aws.amazon.com/sdkref/latest/guide/feature-sts-regionalized-endpoints.html

credentials

Optional credentials shorthand for the config parameter

  • creds:

    • access_key_id: AWS access key ID

    • secret_access_key: AWS secret access key

    • session_token: AWS temporary session token

  • profile: The name of a profile to use. If not given, then the default profile is used.

  • anonymous: Set anonymous credentials.

endpoint

Optional shorthand for complete URL to use for the constructed client.

region

Optional shorthand for AWS Region used in instantiating the client.

Value

A client for the service. You can call the service's operations using syntax like svc$operation(...), where svc is the name you've assigned to the client. The available operations are listed in the Operations section.

Service syntax

svc <- comprehend(
  config = list(
    credentials = list(
      creds = list(
        access_key_id = "string",
        secret_access_key = "string",
        session_token = "string"
      ),
      profile = "string",
      anonymous = "logical"
    ),
    endpoint = "string",
    region = "string",
    close_connection = "logical",
    timeout = "numeric",
    s3_force_path_style = "logical",
    sts_regional_endpoint = "string"
  ),
  credentials = list(
    creds = list(
      access_key_id = "string",
      secret_access_key = "string",
      session_token = "string"
    ),
    profile = "string",
    anonymous = "logical"
  ),
  endpoint = "string",
  region = "string"
)

Operations

batch_detect_dominant_language Determines the dominant language of the input text for a batch of documents
batch_detect_entities Inspects the text of a batch of documents for named entities and returns information about them
batch_detect_key_phrases Detects the key noun phrases found in a batch of documents
batch_detect_sentiment Inspects a batch of documents and returns an inference of the prevailing sentiment, POSITIVE, NEUTRAL, MIXED, or NEGATIVE, in each one
batch_detect_syntax Inspects the text of a batch of documents for the syntax and part of speech of the words in the document and returns information about them
batch_detect_targeted_sentiment Inspects a batch of documents and returns a sentiment analysis for each entity identified in the documents
classify_document Creates a classification request to analyze a single document in real-time
contains_pii_entities Analyzes input text for the presence of personally identifiable information (PII) and returns the labels of identified PII entity types such as name, address, bank account number, or phone number
create_dataset Creates a dataset to upload training or test data for a model associated with a flywheel
create_document_classifier Creates a new document classifier that you can use to categorize documents
create_endpoint Creates a model-specific endpoint for synchronous inference for a previously trained custom model For information about endpoints, see Managing endpoints
create_entity_recognizer Creates an entity recognizer using submitted files
create_flywheel A flywheel is an Amazon Web Services resource that orchestrates the ongoing training of a model for custom classification or custom entity recognition
delete_document_classifier Deletes a previously created document classifier
delete_endpoint Deletes a model-specific endpoint for a previously-trained custom model
delete_entity_recognizer Deletes an entity recognizer
delete_flywheel Deletes a flywheel
delete_resource_policy Deletes a resource-based policy that is attached to a custom model
describe_dataset Returns information about the dataset that you specify
describe_document_classification_job Gets the properties associated with a document classification job
describe_document_classifier Gets the properties associated with a document classifier
describe_dominant_language_detection_job Gets the properties associated with a dominant language detection job
describe_endpoint Gets the properties associated with a specific endpoint
describe_entities_detection_job Gets the properties associated with an entities detection job
describe_entity_recognizer Provides details about an entity recognizer including status, S3 buckets containing training data, recognizer metadata, metrics, and so on
describe_events_detection_job Gets the status and details of an events detection job
describe_flywheel Provides configuration information about the flywheel
describe_flywheel_iteration Retrieve the configuration properties of a flywheel iteration
describe_key_phrases_detection_job Gets the properties associated with a key phrases detection job
describe_pii_entities_detection_job Gets the properties associated with a PII entities detection job
describe_resource_policy Gets the details of a resource-based policy that is attached to a custom model, including the JSON body of the policy
describe_sentiment_detection_job Gets the properties associated with a sentiment detection job
describe_targeted_sentiment_detection_job Gets the properties associated with a targeted sentiment detection job
describe_topics_detection_job Gets the properties associated with a topic detection job
detect_dominant_language Determines the dominant language of the input text
detect_entities Detects named entities in input text when you use the pre-trained model
detect_key_phrases Detects the key noun phrases found in the text
detect_pii_entities Inspects the input text for entities that contain personally identifiable information (PII) and returns information about them
detect_sentiment Inspects text and returns an inference of the prevailing sentiment (POSITIVE, NEUTRAL, MIXED, or NEGATIVE)
detect_syntax Inspects text for syntax and the part of speech of words in the document
detect_targeted_sentiment Inspects the input text and returns a sentiment analysis for each entity identified in the text
detect_toxic_content Performs toxicity analysis on the list of text strings that you provide as input
import_model Creates a new custom model that replicates a source custom model that you import
list_datasets List the datasets that you have configured in this Region
list_document_classification_jobs Gets a list of the documentation classification jobs that you have submitted
list_document_classifiers Gets a list of the document classifiers that you have created
list_document_classifier_summaries Gets a list of summaries of the document classifiers that you have created
list_dominant_language_detection_jobs Gets a list of the dominant language detection jobs that you have submitted
list_endpoints Gets a list of all existing endpoints that you've created
list_entities_detection_jobs Gets a list of the entity detection jobs that you have submitted
list_entity_recognizers Gets a list of the properties of all entity recognizers that you created, including recognizers currently in training
list_entity_recognizer_summaries Gets a list of summaries for the entity recognizers that you have created
list_events_detection_jobs Gets a list of the events detection jobs that you have submitted
list_flywheel_iteration_history Information about the history of a flywheel iteration
list_flywheels Gets a list of the flywheels that you have created
list_key_phrases_detection_jobs Get a list of key phrase detection jobs that you have submitted
list_pii_entities_detection_jobs Gets a list of the PII entity detection jobs that you have submitted
list_sentiment_detection_jobs Gets a list of sentiment detection jobs that you have submitted
list_tags_for_resource Lists all tags associated with a given Amazon Comprehend resource
list_targeted_sentiment_detection_jobs Gets a list of targeted sentiment detection jobs that you have submitted
list_topics_detection_jobs Gets a list of the topic detection jobs that you have submitted
put_resource_policy Attaches a resource-based policy to a custom model
start_document_classification_job Starts an asynchronous document classification job using a custom classification model
start_dominant_language_detection_job Starts an asynchronous dominant language detection job for a collection of documents
start_entities_detection_job Starts an asynchronous entity detection job for a collection of documents
start_events_detection_job Starts an asynchronous event detection job for a collection of documents
start_flywheel_iteration Start the flywheel iteration
start_key_phrases_detection_job Starts an asynchronous key phrase detection job for a collection of documents
start_pii_entities_detection_job Starts an asynchronous PII entity detection job for a collection of documents
start_sentiment_detection_job Starts an asynchronous sentiment detection job for a collection of documents
start_targeted_sentiment_detection_job Starts an asynchronous targeted sentiment detection job for a collection of documents
start_topics_detection_job Starts an asynchronous topic detection job
stop_dominant_language_detection_job Stops a dominant language detection job in progress
stop_entities_detection_job Stops an entities detection job in progress
stop_events_detection_job Stops an events detection job in progress
stop_key_phrases_detection_job Stops a key phrases detection job in progress
stop_pii_entities_detection_job Stops a PII entities detection job in progress
stop_sentiment_detection_job Stops a sentiment detection job in progress
stop_targeted_sentiment_detection_job Stops a targeted sentiment detection job in progress
stop_training_document_classifier Stops a document classifier training job while in progress
stop_training_entity_recognizer Stops an entity recognizer training job while in progress
tag_resource Associates a specific tag with an Amazon Comprehend resource
untag_resource Removes a specific tag associated with an Amazon Comprehend resource
update_endpoint Updates information about the specified endpoint
update_flywheel Update the configuration information for an existing flywheel

Examples

## Not run: 
svc <- comprehend()
svc$batch_detect_dominant_language(
  Foo = 123
)

## End(Not run)


[Package paws version 0.6.0 Index]