Pundi
Pundi AIPundi X
  • 🤖Pundi AI overview
  • 🈁Pundi AI Data Platform
    • Data stakeholders
    • Data Labelling
    • Pundi AI Data Platform (Alpha Testnet) - User Guide
    • FAQ
  • 🦾Pundi AIFX Omnilayer
    • Pundi AIFX
    • Installation Pundi AIFX
    • Setup Node
      • Full node with Binaries
      • Full node with Docker
      • Snapshot Guide
      • Node Monitoring Device
    • Validators
      • Validator Overview
      • Setting Up a Validator for Pundi AIFX
      • Validator Recovery
      • Validator FAQ
      • Validator Security Notice
      • Migration Best Practices
      • Transfer Validator Permissions
    • Delegators
      • Delegators FAQ
      • Delegator CLI Guide
      • Delegator Security Notice
    • Pundi AIFX Tutorials
      • Pundi AIFX CLI Commands
      • Cloud Setup
      • Testnet faucet
      • Ledger Integration for fxcored
      • Sentry Nodes
      • Account Migration Guide (CLI)
    • Upgrade Instructions
      • Cosmovisor Upgrade Guide v8.5.x
        • Cosmovisor Integration - Binaries
        • Cosmovisor Integration - Docker
      • Manual Upgrade Guide v8.5.x
        • Binaries - Upgrading Your Node
        • Docker - Upgrading Your Node
      • Upgrade Versions
        • v2.2.0 Upgrade Instructions
        • v3.1.0 Upgrade Instructions
        • v4.2.1 Upgrade Instructions
        • v5.0.0 Upgrade Instructions
      • Upgrade FAQ
    • Deploying on Pundi AIFX EVM
      • Products
      • MetaMask
        • Download and Install MetaMask
        • Add Network
        • Configure Custom Tokens
        • Create and Import Accounts
      • Connect Wallet to Dapps
        • Connect using MetaMask
        • Connect using Pundi Wallet (Wallet Connect)
      • With Remix
      • With Truffle
      • With Hardhat
      • Cross-Chain Bridges
        • Bridging Tokens from other networks
        • Pundi Gravity Bridge
    • DApps and Infra
      • Margin X Swap
      • Pundi Wallet
      • PundiScan
      • Baklava Space
      • Safe Multisig
      • PortfolioX
      • Token Factory
    • Developers
      • Pundi AIFX Network
      • Pundi AIFX Modules
      • Pundi AIFX JSON RPC
      • Pundi AIFX REST API
      • Web3 JSON RPC
        • JSON RPC Server
        • Namespaces
        • JSON RPC Methods
        • Events
      • Support Leap
      • Third Party Price Oracles
      • Contract Deployments
      • Pundi AIFX Cross Chain
        • sendToFx
        • f(x)Core
        • ibc
        • Target
      • Precompiled Contracts
        • CrossChain Precompiled
        • Staking-V2 Precompiled
      • Pundi AIFX SDKs
        • Python SDK
        • JavaScript SDK
      • Contract Monitoring
    • Pundi X Chain (under Pundi X)
      • Getting Started
        • Install Pundi X (PundiXChain)
        • Setup Node
          • Full node with Binaries
          • Full node with Docker
          • Snapshot Guide
          • Node Monitoring Device
          • Node Peers
      • Validators
        • Validator Overview
        • Setting Up a Validator for PundiXChain
        • Validator Recovery
        • Validator FAQ
        • Validator Security
        • Sentry Nodes
      • Delegators
        • Delegator FAQ
        • Delegator Overview
        • Delegator CLI Guide
      • Pundi X Tutorials
        • Pundi X CLI Guide
        • Ledger Integration for pundixd
        • Testnet Faucet
        • Cloud Setup
      • Upgrade Instructions
        • Cosmovisor Integration - Binaries
        • Cosmovisor Integration - Docker
        • Support keplr
      • Developers
        • Pundi X Network
        • Pundi X Protobuf
        • Pundi X gRPC & REST
        • Pundi X JSON RPC
        • Pundi X Cross Chain
          • ibc
  • 👛PURSE+
    • Purse Box
      • How many NFTs can I mint?
      • Minting (Etherscan)
      • Minting (Pursetoken.com)
    • PURSE+
      • Links
    • Bridging
      • BSC -> Ethereum
    • Protocols
      • LP Restaking Farms
        • How to Use Farms
        • Contract on BscScan
      • PURSE Staking
      • PURSE Staking Rewards
        • Campaigns
      • PURSE Retroactive Rewards
        • Campaigns
    • Contracts
      • Ethereum
      • Binance
      • Function X
    • IBC
      • PUNDIX Chain
    • PURSE Token
      • Background
      • ERC404 intro
      • Accounting of Balances
      • Minting PURSE NFT(s)
      • Transferring PURSE
      • Maintaining PURSE NFT(s)
      • NFTs in Queue
      • Token IDs
  • 🎡Pundi AI MM Agent
    • Why Market Making
    • Traditional MM vs. AMM
    • AI MM Agent: Bridging the Gap
    • Use Cases
    • Train & Launch (Step 1 & 2)
    • Flywheel (Step 3)
    • How the AI MM Agent Works in Detail
    • Why This is Powerful
  • 🛍️Pundi AI Data Marketplace (soon)
  • 🥇PUNDIAI Token Overview
    • Revenue generation flywheel
    • vePUNDIAI
      • how to bribe with vePUNDIAI
    • Protocol pool
  • Governance
    • Governance Proposal Information
      • EGF Info
      • EGF Grants Program
      • Application Template
      • Successful Grant Applicants
      • Launching an EGF Proposal
  • Links
Powered by GitBook
On this page
  1. Pundi AI Data Platform

FAQ

Q: What do the different types of annotations such as Single-label Text Classification and Entity Annotation on the Pundi AI Data Platform mean?

A: These are common data annotation tasks used in machine learning, particularly in supervised learning, to prepare labeled datasets for training models.

  • Single-label Text Classification:

    • Assigns a single label to a given text input.

    • Example: Categorizing an email as "Spam" or "Not Spam."

  • Multi-label Text Classification:

    • Allows assigning multiple labels to a single text input.

    • Example: Tagging a movie review as both "Comedy" and "Romance."

  • Entity Annotation:

    • Involves identifying and labeling specific entities in a text, such as names, dates, or locations.

    • Example: Marking "John" as a person and "New York" as a location in "John visited New York."

  • Single-label Image Classification:

    • Assigns a single label to an image.

    • Example: Classifying an image as either "Dog" or "Cat."

  • Multi-label Image Classification:

    • Allows assigning multiple labels to a single image.

    • Example: Tagging an image as both "Beach" and "Sunset."

  • Bounding Box Annotation:

    • Involves drawing rectangular boxes around objects in an image and labeling them.

    • Example: Marking a car in a street image with a bounding box labeled "Car."

  • Image OCR Annotation:

    • Focuses on recognizing and extracting text from images.

    • Example: Extracting the text "Sale 50% Off" from a billboard photo.

Q: Do tasks go through multiple iterations of data labelling?

A: Yes, tasks go through multiple iterations of data labeling, where they are reviewed or annotated by several individuals or AI agents, to ensure the highest levels of accuracy and consistency. Typically, tasks are labeled by one or more annotators, enabling the comparison of results and the identification of any discrepancies. This practice of cross-labeling helps to minimize bias and errors that may arise from individual interpretation.

Once the initial labeling is completed, tasks enter a review phase, where they are further evaluated by data verifiers or reviewers. This additional review step makes the labeling process more reliable and ensures that the final results align with project goals.

Q: Is there any incentive for reviewers to review labeled data correctly? For example, what is stopping a reviewer from rapidly accepting or denying a number of labeled tasks to try and earn as much as possible?

A: The labeled data will be reviewed and verified by data verifiers, who must approve the results before rewards are distributed. This process helps minimize the involvement of "bad actors" in the annotation process and ensures that the results meet quality standards.

There will be incentive for reviewers and they will only get paid when their “decision” is acknowledged by other parties and mechanics.

Q: Where are you currently sourcing the data from?

A: Currently, our data is sourced from a variety of providers, including internally curated datasets and medical datasets supplied by our client as part of the healthcare project. For text datasets, we are conducting experimental tests on data annotation using content generated by AI agents (Truth Terminal). These posts are used in their original, unfiltered form as provided by the agents, which may explain the presence of certain errors or inappropriate language.

Q: How do you expect you'll be able to provide a consistent flow of labeled data to protocols?

A: By addressing two sides of the equation, the demand side (buyers) and the supply side (annotators and reviewers), in the following ways:

Demand Side: Driving Task Publishers (Buyers)

  1. Quality Speaks for Itself:

    • Core Proposition: Emphasize the high-quality datasets produced by the PUNDI AIDATA platform due to its robust multi-layered review system (peer, AI, and expert validation).

    • Data Validation Assurance:

      • Market the platform as offering ready-to-use datasets for AI model training and validation.

      • Highlight the ⅔ consensus mechanism and expert validation to assure buyers of accuracy and reliability.

  2. Partnership with Other AI Projects:

    • Collaborate with AI initiatives that rely on high-quality datasets.

      • Examples: Autonomous vehicles, NLP, robotics, and medical AI applications.

    • API Integration: Allow seamless integration for direct dataset uploads or requests to/from other AI platforms.

  3. Hugging Face Partnership:

    • Strategic Synergy:

      • Hugging Face's vast community of developers and data contributors aligns perfectly with PUNDI AIDATA's value proposition.

      • Enable Hugging Face contributors to earn extra income by participating in PUNDI’s ecosystem as annotators or reviewers.

    • How It Works:

      • Set up a reward-sharing program where datasets contributed to Hugging Face are tokenized and made available for monetization on PUNDI.

      • Use dual-platform recognition (e.g., co-branding datasets).

  4. Customizable Incentives for Publishers:

    • Task publishers have full control over reward distribution and task structuring, making the platform flexible for diverse AI projects.

    • Offer dynamic pricing models for tasks based on complexity and urgency.

Supply Side: Incentivizing Annotators and Reviewers

  1. Gamified Incentive Models:

    • Introduce achievement levels or badges for annotators and reviewers based on:

  2. Number of tasks completed.

    • Quality of work (peer and AI-reviewed).

    • Unlock higher-paying tasks for top performers.

  3. Income Potential:

    • Market the platform as a side-income opportunity for contributors, with examples like:

    • Annotators who complete X tasks in Y time earn Z tokens.

    • Reviewers with top-tier reputations earning bonuses per task.

    • Demonstrate earning potential compared to similar platforms.

  4. Skill Building:

    • Provide free training materials or tutorials to help contributors improve their skills (e.g., annotating medical images or reviewing NLP datasets).

    • Offer certifications for expert reviewers, which can further boost their earnings.

  5. Broad Accessibility:

    • Ensure tasks are available to contributors of all skill levels:

      • Entry-Level Tasks: Simpler annotations with lower rewards.

      • Advanced Tasks: Higher rewards for tasks requiring domain expertise.

  6. Integration with Web3 Ecosystem:

    • Incentivize contributors by:

      • Providing staking opportunities with rewards linked to platform tokens.

      • Offering NFTs or badges for top-performing contributors that can have monetary or platform-specific value.

  7. Community-Building Initiatives:

    • Host data challenges or hackathons with token prizes to attract and engage contributors.

    • Encourage contributors to share their work on social media for extra rewards, growing the platform's visibility.

PreviousPundi AI Data Platform (Alpha Testnet) - User GuideNextPundi AIFX Omnilayer

Last updated 4 months ago

🈁