Top Builders

Explore the top contributors showcasing the highest number of app submissions within our community.

Gemma

Gemma is a lightweight, open large language model (LLM) from Google, optimized for efficient AI applications. As part of the Google Gemma family, it uses a transformer-based architecture tailored for responsible and accessible AI usage. Developed as a foundational model, Gemma serves various basic language processing needs, including chatbots, content summarization, and multilingual support.

General
Relese dateFebruary 2024​
AuthorGoogle DeepMind in collaboration with Google AI teams
Website[Google AI Gemma]https://ai.google.dev/gemma
RepositoryGoogle AI Developer Resources​
TypeOpen-source AI, transformer-based LLM

Key Features

  • Efficient Deployment: Available in parameter sizes like 2.5B and 7B, Gemma balances capability with efficiency, enabling deployments on both edge devices and cloud infrastructure​.

  • Flexible Tuning Options: Offers pre-trained and instruction-tuned variants, allowing developers to optimize for specific use cases or deploy as-is.

  • Decoder-Only Transformer Architecture: Uses a streamlined decoder-only design, enabling Gemma 1 to process up to 8192 tokens in one pass for better handling of long-form text​.

  • Safety and Accessibility Tools: Integrates responsible AI features, promoting transparency and safety in AI outputs​.

Applications:

  • Chatbot Development: Optimized for conversational tasks, Gemma provides foundational capabilities for chatbot applications.

  • Summarization and Paraphrasing: Its pre-trained model structure makes it suitable for summarizing content across languages and contexts.

  • Multilingual Processing: Supports multilingual inputs, making it adaptable for global applications and translation services​.

Get started building with Gemma:

Developers can quickly integrate Gemma into applications by accessing its model weights on Google AI Studio and Kaggle. The model’s lightweight design ensures that it can run efficiently on most hardware configurations, including mobile and edge devices. For optimal performance, utilize frameworks such as Keras or JAX to customize and deploy Gemma for your specific use case. Get started today by exploring the tools and resources available on the Google AI Gemma platform​.

Google Gemma AI technology Hackathon projects

Discover innovative solutions crafted with Google Gemma AI technology, developed by our community members during our engaging hackathons.

HUMOS

HUMOS

Robot training is expensive(VLA) and hard. What if a first person view of humans hand can be used for training robots? Why is this helpful? Anyone with a mac or an iphone can then start collect training data. They can get paid for it and robotic data can be accelerated. Here's the IDEA real time camera data is used along with media pipe, sam3 , yolo and vlm so the egocentric data can be enrinched with accurate masks from sam3, reasoning from vlm, mediapipe can detect joints data. All of this is super useful for training robots in a cheap and fast way. From lack of data , abundance of data is achieved fast. This is more useful especially for specialized tasks, where only certain humans can do and they are in remote places Object tracking with persistent IDs across frames Zero-shot state classification via SigLIP — 200x faster than VLM for open/closed/ajar labels Navigation state classification via VLM (doors, drawers, handles → open/closed/ajar/blocked) Temporal diff — VLM compares consecutive frames to detect state transitions Navigation timeline — per-object state timeline with colored bars and transition events Hand-object interactions via MediaPipe Ground truth export — structured JSON with per-frame annotations Accuracy evaluation — compare predictions against manual labels Live perception — real-time webcam inference with auto-recording and post-analysis H.264 video export — browser-playable annotated videos with in-app preview Per-frame timing — inference latency breakdown per model stage Tried Gemini 3 Flash, Cosmos and gemma for vlm

RoboGripAI

RoboGripAI

This project presents a simulation-first robotic system designed to perform structured physical tasks through reliable interaction with objects and its environment. The system focuses on practical task execution rather than complex physics modeling, ensuring repeatability, robustness, and measurable performance across varied simulated conditions. Simulation-first robotic system performing structured physical tasks such as pick-and-place, sorting, and simple assembly. Designed for repeatable execution under varied conditions, with basic failure handling, environmental interaction, and measurable performance metrics. A key emphasis of the system is reliability under dynamic conditions. The simulation introduces variations such as object position changes, minor environmental disturbances, and task sequence modifications. The robot is designed to adapt to these variations while maintaining consistent task success rates. Basic failure handling mechanisms are implemented, including reattempt strategies for failed grasps, collision avoidance corrections, and task state recovery protocols. The framework incorporates structured task sequencing and state-based control logic to ensure deterministic and repeatable behavior. Performance is evaluated using clear metrics such as task completion rate, execution time, grasp accuracy, recovery success rate, and system stability across multiple trials. The modular system design allows scalability for additional tasks or integration with advanced planning algorithms. By prioritizing repeatability, robustness, and measurable outcomes, this solution demonstrates practical robotic task automation in a controlled simulated environment, aligning with real-world industrial and research use cases. Overall, the project showcases a dependable robotic manipulation framework that bridges perception, decision-making, and action in a simulation-first setting, delivering consistent and benchmark-driven task execution.