Skip to content
  • Categories
  • Newsletter
  • Recent
  • AI Insights
  • Tags
  • Popular
  • World
  • Groups
Skins
  • Light
  • Brite
  • Cerulean
  • Cosmo
  • Flatly
  • Journal
  • Litera
  • Lumen
  • Lux
  • Materia
  • Minty
  • Morph
  • Pulse
  • Sandstone
  • Simplex
  • Sketchy
  • Spacelab
  • United
  • Yeti
  • Zephyr
  • Dark
  • Cyborg
  • Darkly
  • Quartz
  • Slate
  • Solar
  • Superhero
  • Vapor

  • Default (No Skin)
  • No Skin
Collapse
  1. Home
  2. AI Insights
  3. Only Large Tech Companies Have the Capacity to Bear the Costs of AI Training Data
uSpeedo.ai - AI marketing assistant
Try uSpeedo.ai — Boost your marketing

Only Large Tech Companies Have the Capacity to Bear the Costs of AI Training Data

Scheduled Pinned Locked Moved AI Insights
techinteligencia-ar
1 Posts 1 Posters 3 Views 1 Watching
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
Log in to reply
This topic has been deleted. Only users with topic management privileges can see it.
  • baoshi.raoB Offline
    baoshi.raoB Offline
    baoshi.rao
    wrote on last edited by
    #1

    The development of AI relies on data, and the cost of such data is increasing, making it difficult for companies other than the wealthiest tech firms to bear these costs. According to an article by OpenAI researcher James Betker last year, the training data for AI models is a key factor determining their capabilities. Traditional AI systems are primarily based on statistical machines, using vast examples to guess the most 'reasonable' data distribution. Therefore, the larger the dataset a model relies on, the better its performance.

    Kyle Lo, a senior research scientist at the nonprofit AI research organization AI2, pointed out that Meta's Llama3 model significantly outperforms AI2's OLMo model in terms of data volume, which explains its advantage in many popular AI benchmarks. However, larger datasets do not always lead to linear improvements in model performance; data quality and organization are equally important, sometimes even more so than quantity. Some AI models are trained using human-annotated data, and higher-quality annotations have a significant impact on model performance.

    However, Lo and other experts worry that the demand for large, high-quality training datasets is concentrating AI development in the hands of a few companies with multi-billion-dollar budgets. Although some illegal or even criminal activities may raise questions about data acquisition methods, tech giants can secure data licenses due to their financial strength. These data transactions do not foster a fair and open generative AI ecosystem, harming the broader AI research community.

    Some independent, nonprofit organizations, such as EleutherAI and Hugging Face, are attempting to open large-scale datasets, but whether they can keep up with the pace of large tech companies remains uncertain. Only when research breakthroughs eliminate technical barriers and the costs of data collection and organization are no longer an issue will these open datasets have a chance to compete with tech giants.

    1 Reply Last reply
    0
    Reply
    • Reply as topic
    Log in to reply
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes


    • Login

    • Don't have an account? Register

    • Login or register to search.
    • First post
      Last post
    0
    • Categories
    • Newsletter
    • Recent
    • AI Insights
    • Tags
    • Popular
    • World
    • Groups