Supametas.AI
Official website
English
English
  • Welcome to Supametas.AI
  • Cloud Service
    • Supametas.AI Cloud Service
      • Detailed Pricing Comparison
    • Guide
      • Create Dataset
      • Import Metadata
        • Fetching webpage data
        • Importing local text processing
        • Importing local image processing
        • Importing local audio processing
        • Importing local video processing
      • Cleaning data query
      • Export Cleaned Data
      • Dataset Configuration
  • Developer
    • Integration Process
    • Create API Key
    • Standard Request and Response
    • API
      • Import Text File
      • Import Web Page Task
      • Import Image Files
      • Import Audio Files
      • Import Video Files
      • View Import Task Details
      • Delete File Processing Task
    • Webhook
    • Error Code
    • Demo
  • Other
    • Community
    • Technical Support
Powered by GitBook
On this page
  • 1. Create a New Task
  • 2. Upload Local Text Files:
  • 3. Task Settings
  • 4. Get parameters
  • 5. Output Settings
  • 6. Save or Execute Task Immediately
  1. Cloud Service
  2. Guide
  3. Import Metadata

Importing local text processing

If your original data is text, then follow this document for processing.

PreviousFetching webpage dataNextImporting local image processing

Last updated 4 months ago

Select 'From Local Text' and 'New Task' in the 'Import Data Source' section on the dataset details page.

All imports are referred to as tasks, where each task can include multiple similar data entries for processing. In this section, local text tasks can accommodate several text data entries.

1. Create a New Task

On the task creation page, provide a name for your task (up to 20 characters). This name will assist you in quickly locating and managing this task in the task list.

2. Upload Local Text Files:

  1. Click on the upload area, drag and drop the local files you want to import into the upload box, or click the upload button to select files for uploading.

  2. Supported file formats include: .docx、.pdf、.txt、.md、.json.

  3. You can upload up to 50 files per task, with each file not exceeding 200MB in size(in some cases, the CDN we use may only allow file uploads of around 100MB).

  4. Ensure that multiple files uploaded within a task have similar content for parameter extraction and output processing.

3. Task Settings

  1. Task settings are similar to importing tasks from the web and involve configuring fields and content extraction.

  2. Select an appropriate parsing method based on the file type to ensure that the system can correctly process the uploaded text files.

4. Get parameters

  1. Default Field Types:

    • Title: The system will attempt to extract title information from the file content.

    • Content Details: The system will capture and store the main content of the file.

  2. Custom Fields:

    • If you need to categorize specific extracted data into designated fields, you can click on "+ Add Field" and add field names and descriptions.

    • For example, if there is a nickname to be extracted from the text, the field name key could be: nickname; field description: user nickname.

    • Please use English when adding custom fields; more detailed descriptions lead to more accurate extraction.

5. Output Settings

Once you have configured the fetch parameters, you will need to set up output settings to determine how the extracted data will be saved and exported.

  1. Output Format Settings

    • You can choose to save the retrieved data in either JSON or Markdown format. JSON format is more suitable for subsequent API program calls, while Markdown format is better suited for knowledge base data processing.

6. Save or Execute Task Immediately

  1. Save and manually execute the task later:

    • If you wish to configure the task first without initiating the scraping process immediately, you can click on the "Save and manually execute the task later" button. The task will be saved in the task list for manual initiation at a later time.

  2. Execute the task immediately:

    • If you are prepared to instantly scrape webpage data, click the "Execute the task immediately" button. The system will commence data scraping and import it into the specified dataset.

Create a new task to import from local text for the dataset