Top Builders

Explore the top contributors showcasing the highest number of app submissions within our community.

Generative Agents

Generative Agents are computer programs designed to replicate human actions and responses within interactive software. To create believable individual and group behavior, they utilize memory, reflection, and planning in combination. These agents have the ability to recall past experiences, make inferences about themselves and others, and devise strategies based on their surroundings. They have a wide range of applications, including creating immersive environments, rehearsing interpersonal communication, and prototyping. In a simulated world resembling The Sims, automated agents can interact, build relationships, and collaborate on group tasks while users watch and intervene as necessary.

General
Relese dateApril 7, 2023
TypeAutonomous Agent Simulation

Start building with Generative Agents

We have collected the best Generative Agents libraries and resources to help you get started to build with Generative Agents today. To see what others are building with Generative Agents, check out the community built Generative Agents Use Cases and Applications.

Generative Agents Tutorials


Generative Agents resources

Kickstart your development with Generative Agents.


Generative Agents AI technology page Hackathon projects

Discover innovative solutions crafted with Generative Agents AI technology page, developed by our community members during our engaging hackathons.

Ladybug - The Robot Reader

Ladybug - The Robot Reader

There are 240 million children worldwide living with learning disabilities, and many struggle to access physical books independently. Ladybug: The Robot Reader was built to change that. Ladybug is an autonomous robotic system that reads physical books aloud from cover to cover with no human intervention. Built on the SO-101 robotic arm, it uses a perception-action loop powered by Claude Vision to assess the workspace, decide what to do next, and execute — opening a closed book, reading each page spread, turning pages, and closing the book when finished. Claude Vision analyzes camera frames to classify page types (content, title page, table of contents, index, blank) and extract text. ElevenLabs then streams natural-sounding speech in real time using a sentence-level prefetch pipeline so audio plays continuously without pauses. Motor skills — opening, closing, and page turning — are trained using ACT (Action Chunking with Transformers) policies. The system includes intelligent retry logic with frame hashing to detect failed page turns and automatically retry them. Ladybug supports multiple reading modes: verbose (reads everything), skim (headers and titles only), and silent (text extraction only). It also features a web dashboard for remote monitoring and a dry-run mode for testing without hardware. Our mission is accessibility in education — putting an autonomous reading companion in every special education classroom. We want 1,000,000 lady bug robot readers available to children around the world.

ClutterBot

ClutterBot

ClutterBot is a proof-of-concept simulation platform that bridges natural language understanding and robotic task execution for household cleanup tasks. Users issue commands like "pick up the phone and the toy train," which Gemini 3 Flash parses into structured task lists. The system generates complete execution plans upfront, with Gemini deciding the sequence of pick-and-place operations for each object. The architecture combines a FastAPI backend hosted on Vultr (central system of record), a Next.js frontend for real-time monitoring, and a MuJoCo physics simulation featuring a Franka FR3 manipulator in a room environment with everyday objects. The robot executes inverse kinematics motions to relocate objects from scattered positions on a table to a collection bin, with each action streamed via WebSocket for live visualization. This prototype validates the feasibility of integrating large language models with robotic simulation pipelines, demonstrating how AI can translate high-level human intent into executable robot behaviors. While the current implementation uses deterministic motion planning with hardcoded inverse kinematics rather than learned policies, the framework establishes foundational patterns for future work incorporating adaptive control, real hardware integration, and expanded object manipulation capabilities. The plan-first approach (Gemini generates the full task plan in a single API call) shows AI reasoning while keeping execution fast and deterministic, making it suitable for real-time interactive use.