Freeplay’s cover photo
Freeplay

Freeplay

Software Development

Boulder, Colorado 987 followers

A better way to build with LLMs. Prompt engineering, testing & evaluation tools for your whole team.

About us

A better way to build with LLMs. Bridge the gap between domain experts & developers. Prompt engineering, testing & evaluation tools for your whole team. Now in private beta.

Industry
Software Development
Company size
2-10 employees
Headquarters
Boulder, Colorado
Type
Privately Held
Founded
2022
Specialties
Artificial Intelligence and Developer Tools

Locations

Employees at Freeplay

Updates

  • Freeplay reposted this

    The next AI Builders meetup is March 19 in Denver, and it's almost at 250 RSVPs! We've got room for up to 300 so sign up soon. ⏱️ These have turned into one of the most fun events I've been part of in Colorado. Come see what AI products people are building, from big companies to local startups. The last one was hosted at Google Boulder and featured we packed it out, so we've got a bigger space this time and won't have to turn anyone away. Big thanks to everyone who helps make these happen including our sponsors Silicon Valley Bank and Technical Integrity, and our co-organizers at Matchstick Ventures, Ombud and Freeplay. 🤘 Sign up here: https://lu.ma/n0g16hzk

    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
  • We’ve been cooking on this for a while. Much better way to run a feedback loop for AI products. 🤘 Let us know what you think!

    🔍 Scale Your AI Data Review Process & Build A Better Learning Loop with Freeplay's New Review Queues One thing we've consistently seen from the best AI product teams is that they look at LOTS of row-level data. They build intuition about what's working and what isn't by reviewing real user interactions with their systems. But as teams scale, coordinating this review work gets complicated fast. Some of our customers now have 30-40 people involved in reviewing AI outputs for a single feature or agent. That's why we've launched Review Queues in Freeplay – tools to help teams manage human review at scale: 🧑💻 Create bounded data sets from production completions for targeted review ✅ Assign work to many team members with due dates and track progress 📊 Auto-generate insights reports that make it easy to share findings with stakeholders 🧪 Turn reviewed data directly into evaluation or fine-tuning datasets 🔒 Support sensitive data with private projects and granular user roles Here's what one customer at a major financial institution told us: 📣 "Freeplay's Reviews feature has been a game-changer for operationalizing LLMs within our prompt ops workflow. Analysts can now efficiently review interactions, surface insights, and refine prompts—all in a seamless, closed-loop system." I'm particularly excited about how Review Queues help AI teams close the loop between spotting issues and taking action. No more spreadsheets and screenshots flying around in Slack and Jira. Check out our latest blog post (linked in comments) for a deeper dive into how Review Queues work and a demo video showing some of the real-world problems they solve. If you're thinking about these problems our team Freeplay would love to chat. #AIQuality #ProductDevelopment #DataOps #MachineLearning

    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
  • New podcast is up with the CTO of Box on building agents in (and for) the enterprise. Check it out!

    🎯 Fresh conversation with Box CTO Ben Kus about building AI agents for the enterprise. Good timing with the AI Engineer Summit on agents this week! Box has 100K+ customers and manages petabytes of unstructured data, and generative AI gave them superpowers to help customers make sense of all that data. They've invested in generative AI search and summary features from early on, and they've been getting into more complex agentic behavior for a while now. Ben shares a bunch of good highlights and practical lessons learned from their experience that will be interesting to Freeplay customers and anyone else building agents at scale: 🎯 How they got started with generative AI, built teams, and decided what to build 🤖 What's going on with agents these days, and what's involved to bring them to true enterprise customers 💬 Moving beyond chat and conversational UX to apps that just do the work for you 🛠️ Foundational tools they've invested in, including the essential role of LLM judges to build high quality products (especially since they never look at customer data) Check out the link in the comments to get the full episode on Spotify, Apple or YouTube. #EnterpriseAI #ProductDevelopment #AIAgents #Engineering

  • Freeplay reposted this

    View organization page for KO Law®

    1,062 followers

    🚀 Colorado Startups recently released its updated Leaderboard, featuring the fastest growing #startups and companies in Colorado as of Jan. 31, 2025. Congrats to all the companies on the list, including our clients Freeplay, Pomp, Silvis Materials, Spekit 🐙, StackHawk, Strata Identity, Suite Studios, and many other friends! #startup #fastestgrowing #entrepreneurship

    View organization page for Colorado Startups

    1,447 followers

    We are STOKED to present The Leaderboard, a showcase of the 50 fastest-growing startups and companies in Colorado as of January 31, 2025. 🏆 The Leaderboard is our way of shining a light on the top startups and their founders who are building great companies in Colorado! Check it out at https://lnkd.in/eaUcJni ! Congrats to the top 3 companies in each category and all 100 that made either leaderboard! Overall Leaderboard (Founded in last 10 years) 🥇 Sierra Space (sierraspace.com) 🥈 Quantinuum (quantinuum.com) 🥉 Transcarent (transcarent.com) Startup Leaderboard (Founded in last 5 years and raised less than $10m) 🥇 SurePath AI (surepath.ai) 🥈 Cloverleaf AI (cloverleaf.ai) 🥉 ThinkOrbital (thinkorbital.com) Please join me in celebrating these companies, the full leaderboard, and all the folks building great companies and our amazing entrepreneurial ecosystem in #Colorado!

    The Leaderboard — Colorado Startups

    The Leaderboard — Colorado Startups

    coloradostartups.org

  • Freeplay reposted this

    We've gotten lots of great feedback from people working on agents about the latest Deployed podcast with Arya Asemanfar from Sierra. He talks straight about practical details and tactics, including lessons learned along the way. If you haven't heard/seen it yet, check it out. 🙌 Here's the link to all the links -- Spotify, Apple, YouTube, etc: https://lnkd.in/g8dK763p

  • We love supporting CU! 🦬 Many of our Freeplay teammates are graduates, so it's a no-brainer to support programs like HackCU. We're looking forward to seeing what this year's hackers build!

    View profile for Tom Yeh

    CS Prof | AI by Hand ✍️ | CU Boulder

    My university's largest hackathon is in its 11th year! When my students started HackCU back then, we didn't know it would keep running and growing so big! Kudos to all the organizers who keep passing down the torch, now to Andy Strong and Co, and all the students who participate in each year's HackCU event! 🙌 A big shout-out to Freeplay who generously sponsored this year's hackathon. Freeplay is working on a powerful platform to build, test, and evaluate LLM applications. Check it out: https://www.freeplay.ai/ We are still looking for more sponsors. If you are interested in supporting us, please reach out to Andy Strong!

    • No alternative text description for this image
    • No alternative text description for this image
  • Freeplay reposted this

    🔮 Building Enterprise AI Agents That Actually Work This week on our Deployed podcast, Arya Asemanfar from Sierra shared insights from building AI agents that manage customer service for enterprise customers. The conversation is packed with practical learnings for teams working on (or wanting to work on) complex AI agents. We started Deployed so that other builders could learn from leaders who are building AI systems at scale. Arya shares a ton of practical wisdom specific to building agents, which we know lots of teams are curious about right now. Key lessons from Sierra's journey building enterprise agents include: ✅ They treat each agent as its own product, with dedicated team members responsible for it ✅ Build comprehensive eval systems that combine single prompt "unit tests", multi-step "integration tests" for an agent skill or flow, and human review as a backstop ✅ Carefully design feedback loops for different types of issues, based on the types of improvements you need to make (tone, behavior, system-level) ✅ Invest in tooling - monitoring, observability, eval, and testing infrastructure are crucial Check out the full episode (linked in comments) for a deeper dive into Sierra's approach to building enterprise-grade AI agents. We'd love to hear from others working on similar challenges - what's the most valuable thing you've done to get agents to work well? #AIAgents #EnterpriseAI #ProductDevelopment #ArtificialIntelligence

  • The latest Deployed podcast is live! This week we talked with Arya Asemanfar from Sierra about building AI agents for enterprise customers. Check it out. 👇

    🔮 Building Enterprise AI Agents That Actually Work This week on our Deployed podcast, Arya Asemanfar from Sierra shared insights from building AI agents that manage customer service for enterprise customers. The conversation is packed with practical learnings for teams working on (or wanting to work on) complex AI agents. We started Deployed so that other builders could learn from leaders who are building AI systems at scale. Arya shares a ton of practical wisdom specific to building agents, which we know lots of teams are curious about right now. Key lessons from Sierra's journey building enterprise agents include: ✅ They treat each agent as its own product, with dedicated team members responsible for it ✅ Build comprehensive eval systems that combine single prompt "unit tests", multi-step "integration tests" for an agent skill or flow, and human review as a backstop ✅ Carefully design feedback loops for different types of issues, based on the types of improvements you need to make (tone, behavior, system-level) ✅ Invest in tooling - monitoring, observability, eval, and testing infrastructure are crucial Check out the full episode (linked in comments) for a deeper dive into Sierra's approach to building enterprise-grade AI agents. We'd love to hear from others working on similar challenges - what's the most valuable thing you've done to get agents to work well? #AIAgents #EnterpriseAI #ProductDevelopment #ArtificialIntelligence

  • Don’t miss the next meetup in Denver! Sign up for updates https://lnkd.in/gafEXR9A

    Last night’s AI Builders meetup at Google Boulder office was amazing. Google were incredible hosts, showed up like the rest of the community and showed off some very early software, and a couple hundred of our closest friends showed up. ✨🏔️ We were super bummed we had to turn some people away because of capacity issues. 😢 Not cool, and we feel awful about it. We hope you’ll forgive us, and we’re sorry to folks who were impacted (there were ~20-30). We’re following up separately with them today. Some growing pains as this community scales, and we'll get better for the future — we promise. 🙏 Check out the photos below (and Google’s incredible space). Many many thanks to the 6 builders who demoed what they’ve been working on, inside and outside of work: * Rustin Banks of Google 🤐 * Jane Fine showed off new Google Labs coding agents in Colab and Jules (and made the whole event possible 🙌) * Joey McDonald showed us an epic hardware demo of his Core Dispatch project (raspberry pi, walkie talkies, radio receivers, etc.) * Danielle Dannenberg featured Atlassian’s newest Rovo AI features (there are many!) * Kevin Walkup previewed what’s coming at Roam * Gabe Monroy from Google Cloud was last but not least: GCP Cloud Run on-demand GPUs with a 5s cold start 🤯 Lots of great folks announced open AI engineering and related roles too: Brightwave, Freeplay (that’s us!), NVIDIA, Heroku, Biltwise, KamiwazaAI, 4everoceans (seeking a co-founder), and the many companies represented by our partners, Technical Integrity. Check out their websites for more! These events are a team effort and we’re endlessly grateful to our long-standing co-organizers, Matchstick Ventures and Ombud, as well as our generous sponsors at Silicon Valley Bank and Technical Integrity. We want to give a very special shout out to our friend Jane Fine who led the charge and the long list of other Googlers (Kelly Schaefer 🇺🇦, Gabe Monroy, Kyle Duffy, and so many more!) who turned out to support the event. Next up, Denver -- date TBD but likely late Feb/early March. Sign up here for more info and future updates: boulderaibuilders.org

    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
    • No alternative text description for this image
  • Freeplay reposted this

    🚀 Speed up AI product development with an "AI Quality Lead" Role Last week on our Deployed podcast, our friends Luis Morales and Nick Francis at Help Scout shared how they transformed their customer service platform into an AI-native product company. One of the tactics that made a big impact on their product experience was introducing a role they call the AI Quality Lead. It’s a strategy we’ve seen work across several of our customers, and it’s one we think more teams should explore. People in this role (and roles like it) act as a critical bridge between domain expertise and AI development — enabling tighter feedback loops, more trustworthy product quality and performance, and faster time to market. What does an AI Quality Lead do? ✅ Reviews production data for quality issues ✅ Defines evaluation criteria and writes evals to improve future quality ✅ Manages test datasets for comprehensive testing ✅ Improves prompts to maintain tone and quality ✅ Collaborates with engineering and PM teams to prioritize issues ✅ Trains teammates on evaluating AI outputs Luis describes the impact more in this video. In just one month, their AI Quality Lead — who came from a customer success background — transformed the quality AI-generated email drafts, making them indistinguishable from those written by their human team. I'll drop a couple more links in the comments, both to our Freeplay blog post about this role and to a draft job description for teams who might want to consider something similar. Let us know what you think! Especially if you've got a similar role on your team.

Similar pages

Browse jobs

Funding