LogoLogo
Release notesPython SDK docsDocs homeTextual CloudTonic.ai
  • Tonic Textual guide
  • Getting started with Textual
  • Previewing Textual detection and redaction
  • Entity types that Textual detects
    • Built-in entity types
    • Managing custom entity types
  • Language support in Textual
  • Datasets - Create redacted files
    • Datasets workflow for text redaction
    • Creating and managing datasets
    • Assigning tags to datasets
    • Adding and removing dataset files
    • Reviewing the sensitivity detection results
    • Configuring the redaction
      • Configuring added and excluded values for built-in entity types
      • Working with custom entity types
      • Selecting the handling option for entity types
      • Configuring synthesis options
      • Configuring handling of file components
    • Adding manual overrides to PDF files
      • Editing an individual PDF file
      • Creating templates to apply to PDF files
    • Sharing dataset access
    • Previewing the original and redacted data in a file
    • Downloading redacted data
  • Pipelines - Prepare LLM content
    • Pipelines workflow for LLM preparation
    • Viewing pipeline lists and details
    • Assigning tags to pipelines
    • Setting up pipelines
      • Creating and editing pipelines
      • Supported file types for pipelines
      • Creating custom entity types from a pipeline
      • Configuring file synthesis for a pipeline
      • Configuring an Amazon S3 pipeline
      • Configuring a Databricks pipeline
      • Configuring an Azure pipeline
      • Configuring a Sharepoint pipeline
      • Selecting files for an uploaded file pipeline
    • Starting a pipeline run
    • Sharing pipeline access
    • Viewing pipeline results
      • Viewing pipeline files, runs, and statistics
      • Displaying details for a processed file
      • Structure of the pipeline output file JSON
    • Downloading and using pipeline output
  • Textual Python SDK
    • Installing the Textual SDK
    • Creating and revoking Textual API keys
    • Obtaining JWT tokens for authentication
    • Instantiating the SDK client
    • Datasets and redaction
      • Create and manage datasets
      • Redact individual strings
      • Redact individual files
      • Transcribe and redact an audio file
      • Configure entity type handling for redaction
      • Record and review redaction requests
    • Pipelines and parsing
      • Create and manage pipelines
      • Parse individual files
  • Textual REST API
    • About the Textual REST API
    • REST API authentication
    • Redaction
      • Redact text strings
  • Datasets
    • Manage datasets
    • Manage dataset files
  • Snowflake Native App and SPCS
    • About the Snowflake Native App
    • Setting up the app
    • Using the app
    • Using Textual with Snowpark Container Services directly
  • Install and administer Textual
    • Textual architecture
    • Setting up and managing a Textual Cloud pay-as-you-go subscription
    • Deploying a self-hosted instance
      • System requirements
      • Deploying with Docker Compose
      • Deploying on Kubernetes with Helm
    • Configuring Textual
      • How to configure Textual environment variables
      • Configuring the number of textual-ml workers
      • Configuring the number of jobs to run concurrently
      • Configuring the format of Textual logs
      • Setting a custom certificate
      • Configuring endpoint URLs for calls to AWS
      • Enabling PDF and image processing
      • Setting the S3 bucket for file uploads and redactions
      • Required IAM role permissions for Amazon S3
      • Configuring model preferences
    • Viewing model specifications
    • Managing user access to Textual
      • Textual organizations
      • Creating a new account in an existing organization
      • Single sign-on (SSO)
        • Viewing the list of SSO groups in Textual
        • Azure
        • GitHub
        • Google
        • Keycloak
        • Okta
      • Managing Textual users
      • Managing permissions
        • About permissions and permission sets
        • Built-in permission sets and available permissions
        • Viewing the lists of permission sets
        • Configuring custom permission sets
        • Configuring access to global permission sets
        • Setting initial access to all global permissions
    • Textual monitoring
      • Downloading a usage report
      • Tracking user access to Textual
Powered by GitBook
On this page
  • Viewing the list of all files for a pipeline
  • Viewing file statistics for the pipeline
  • Summary file statistics
  • Entity type value counts
  • Topics list
  • Viewing the list of pipeline runs
  • Viewing the list of pipeline run files
  • Information in a file list

Was this helpful?

Export as PDF
  1. Pipelines - Prepare LLM content
  2. Viewing pipeline results

Viewing pipeline files, runs, and statistics

Last updated 1 month ago

Was this helpful?

Required pipeline permission: View pipeline settings

For an uploaded file pipeline, the Files tab contains the list of all of the pipeline files.

For cloud storage pipelines, you use the pipeline details page to track processed files and pipeline runs.

For pipelines that are configured to also redact files, you can configure the redaction for the detected entity types. For more information, go to Selecting the handling option for entity types.

Viewing the list of all files for a pipeline

For uploaded file pipelines, when you add a file to the pipeline, it is automatically added to the file list.

For cloud storage pipelines, the file list is not populated until you run the pipeline. The list only contains processed files.

Viewing file statistics for the pipeline

The statistics panels at the right of the pipeline details page provide a summary of information about the pipeline files, the detected entities, and the detected topics.

Summary file statistics

The File Statistics panel displays the following values.

  • Total # of files - The number of files in the pipeline.

  • Total # of words - The number of words that the files contain.

  • Entities detected - The number of entity types for which Textual detected values in the files.

  • Topics detected - The number of topics that the files contain. A topic is a subject area that is common across multiple files. If the pipeline files contain completely unrelated content, then Textual might not detect any topics.

Entity type value counts

The entity types panel displays the 5 entity types that have the largest number of values in the pipeline files.

For each entity type, the panel displays the value count.

If there are more than 5 detected entity types, to display the full list of detected entity types, click View All.

Topics list

The topics panel displays the 5 topics that are present in the most files.

For each topic, the panel displays the number of files that include that topic.

If there are more than 5 detected topics, to display the full list of detected topics, click View All.

Viewing the list of pipeline runs

On the pipeline details page for a cloud storage pipeline, the Pipeline Runs tab displays the list of pipeline runs.

Required pipeline permission: View pipeline settings

For each run, the list includes:

  • Run identifier

  • When the run was started

  • The current status of the pipeline run. The possible statuses are:

    • Queued - The pipeline run has not started to run yet.

    • Running - The pipeline run is in progress.

    • Completed - The pipeline run completed successfully.

    • Failed - The pipeline run failed.

Viewing the list of pipeline run files

For a pipeline run, to display the list of files that the pipeline run includes, click View Run.

Information in a file list

For each file, the list includes the following information:

  • File name

  • For cloud storage files, the path to the file

  • The status of the file processing. The possible status are:

    • Unprocessed - The file is added, but a pipeline run to process it has not yet started. This only applies to uploaded files that were added since the most recent pipeline run.

    • Queued - A pipeline run was started but the file is not yet processed.

    • Running - The file is being processed.

    • Completed - The file was processed successfully.

    • Failed - The file could not be processed.

Files list for an uploaded file pipeline
Files tab on the pipeline Overview page
File statistics for a pipeline
Pipeline Runs tab on the pipeline Overview page
File list for a pipeline run