Close Menu
OpenWing – Agent Store for AIoT Devices

    Subscribe to Updates

    Get the latest creative news from FooBar about art, design and business.

    What's Hot

    Build AI in Wearables – OpenWing DevPack

    April 13, 2025

    DevPack AI Notelet – “Capture. Transcribe. Summarize. In Your Pocket.”

    April 9, 2025

    Gemini Robotics Revolutionizes AI Integration in Robotics

    April 8, 2025
    Facebook X (Twitter) Instagram
    OpenWing – Agent Store for AIoT DevicesOpenWing – Agent Store for AIoT Devices
    • AIoT Hotline
    • AGENT STORE
    • DEV CENTER
      • AIoT Agents
      • Hot Devices
      • AI on Devices
      • AI Developer Community
    • MARKETPLACE
      • HikmaVerse AI Products
      • Biz Device Builder
      • Global Marketing
        • Oversea Marketing Strategy
        • Customer Acquisitions
        • Product Launch Campaigns
      • Startup CFO Services
      • Partner Onboarding
        • Media Affiliate Program
    Facebook X (Twitter) Instagram
    OpenWing – Agent Store for AIoT Devices
    Home»News»Geekbench releases AI benchmarking app
    News

    Geekbench releases AI benchmarking app

    No Comments1 Min Read
    Facebook Twitter Pinterest LinkedIn Tumblr Email Reddit Copy Link VKontakte
    Share
    Facebook Twitter LinkedIn Pinterest Email Reddit Copy Link VKontakte Telegram WhatsApp

    [ad_1]

    Benchmarking stalwarts Primate Labs on Thursday released Geekbench AI 1.0. The app, which is currently available for Android, Linux, MacOS and Windows, applies Geekbench’s principles to machine learning, deep learning and other AI workloads, in a bid to standardize performance ratings across platforms. It’s a successor to Geekbench ML (machine learning), which was announced in 2021 and is currently on version 0.6.

    “[I]n recent years, companies have coalesced around using the term ‘AI’ in these kinds of workloads (and in their related marketing),” Primate Labs says of the name change. “To ensure that everyone, from engineers to performance enthusiasts, understands what this benchmark does and how it works, we felt it was time for an update.”

    Earlier this week, ChatGPT-maker OpenAI announced a new version of its own AI model benchmark. SWE-bench Verified is a “human-validated” offering that uses human validation to determine models’ efficacy in solving “real world issues.”

    [ad_2]

    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email Reddit Copy Link

    Related Posts

    Gemini Robotics Revolutionizes AI Integration in Robotics

    April 8, 2025

    Hyundai Amplifies Robotics Partnership with Boston Dynamics, Eyeing Mass Deployment of Humanoid Robots

    April 8, 2025

    Unitree G1: The World’s First Side-Flipping Humanoid Robot Astonishes with Acrobatic Feats

    April 8, 2025

    The Rise of AI Mental Health Chatbots for Children: Navigating the Ethical Labyrinth

    April 8, 2025
    Add A Comment

    Comments are closed.

    OpenWing – Agent Store for AIoT Devices
    Facebook X (Twitter) Instagram Pinterest Vimeo YouTube
    • Home
    • ABOUT US
    • CONTACT US
    • TERMS
    • PRIVACY
    © 2025 OpenWing.AI, all rights reserved.

    Type above and press Enter to search. Press Esc to cancel.