A Detective Quest

Streamlit
application badge
Created by team FlyingKittens on May 27, 2023

We create an interactive role-playing novel game. You will play the game by entering statements. The assistant played by Claude will respond you information from script. You will find clues and submit your reasoning at last. We are a team dedicated to creating immersive and interactive gaming experiences. Our game aims to offer a rich interactive experience, and we believe that the Claude-100K model API can be a valuable tool in achieving this goal. We plan to use the model API to generate dynamic and engaging text-based interactions and narratives. This includes but is not limited to: Character Interactions: Utilizing the model API, we aim to create non-player characters (NPCs) that can carry out complex and engaging conversations with the player. These interactions will be context-aware and responsive to the player's inputs. The first challenge is that Claude may respond with hallucination. We designed the prompt to weaken the hallucination. The second challenge is how to force Claude not tell the player the truth directly. We know, Claude is a very safe AI. So, it may not tell a lie even in a role-playing game. We designed the information input to make sure Claude won't Reveal the follow-up plot. We will take you more immersive game experience in the future!

Category tags:

Games, Game Developement, Game Engine Extension

Explore more applications

ELIZA EVOL INSTRUCT - Fine-Tuning

We attempted to instill the deterministic, rule-based reasoning found in ELIZA into a more advanced, probabilistic model like an LLM. This serves a dual purpose: To introduce a controlled variable in the form of ELIZA's deterministic logic into the more "fuzzy" neural network-based systems. To create a synthetic dataset that can be used for various Natural Language Processing (NLP) tasks, beyond fine-tuning the LLM. [ https://huggingface.co/datasets/MIND-INTERFACES/ELIZA-EVOL-INSTRUCT ] [ https://www.kaggle.com/code/wjburns/pippa-filter/ ] ELIZA Implementation: We implemented the script meticulously retaining its original transformational grammar and keyword matching techniques. Synthetic Data Generation: ELIZA then generated dialogues based on a seed dataset. These dialogues simulated both sides of a conversation and were structured to include the reasoning steps ELIZA took to arrive at its responses. Fine-tuning: This synthetic dataset was then used to fine-tune the LLM. The LLM learned not just the structure of human-like responses but also the deterministic logic that went into crafting those responses. Validation: We subjected the fine-tuned LLM to a series of tests to ensure it had successfully integrated ELIZA's deterministic logic while retaining its ability to generate human-like text. Challenges Dataset Imbalance: During the process, we encountered issues related to data imbalance. Certain ELIZA responses occurred more frequently in the synthetic dataset, risking undue bias. We managed this through rigorous data preprocessing. Complexity Management: Handling two very different types of language models—rule-based and neural network-based—posed its unique set of challenges. Significance This project offers insights into how the strength of classic models like ELIZA can be combined with modern neural network-based systems to produce a model that is both logically rigorous and contextually aware.

MIND INTERFACES

LlamaIndexLlama 2
Streamlit
application badge

Auto Recruit

Our platform revolutionizes recruitment with personalized experiences for candidates and streamlined processes for employers. Challenges with traditional recruitment system are: 1. TIme consuming 2. Screening Hassles 3. Inconsistent result 4. In effective methods Solutions - AutoRecruit AI is a comprehensive and cutting-edge solution to these problems. It puts the candidate at the center of its process and applies a breakthrough llama-based algorithmic approach to achieve unprecedented accuracy at speeds that haven't been seen before! Features 1. Candidate sourcing 2. Resume parsing 3.Candidate scoring and summary 4. Personalized Engagement Benefits 1. Time Saving 2. Cost-Effective 3. Efficiency Optimization 4. Better Candidate Fit

AutoHire AI

LangChainLlama 2

Visionary Plates

Visionary Plates: Advancing License Plate Detection Models is a project driven by the ambition to revolutionize license plate recognition using cutting-edge object detection techniques. Our objective is to significantly enhance the accuracy and robustness of license plate detection systems, making them proficient in various real-world scenarios. By meticulously curating and labeling a diverse dataset, encompassing different lighting conditions, vehicle orientations, and environmental backgrounds, we have laid a strong foundation. Leveraging this dataset, we fine-tune the YOLOv8 model, an architecture renowned for its efficiency and accuracy. The model is trained on a carefully chosen set of parameters, optimizing it for a single class—license plates. Through iterative experimentation and meticulous fine-tuning, we address critical challenges encountered during this process. Our journey involves overcoming obstacles related to night vision scenarios and initial model performance, with innovative solutions like Sharpening and Gamma Control methods. We compare and analyze the performance of different models, including YOLOv5 and traditional computer vision methods, ultimately identifying YOLOv8 as the most effective choice for our specific use case. The entire training process, from dataset curation to model fine-tuning, is efficiently facilitated through the use of Lambda Cloud's powerful infrastructure, optimizing resources and time. The project's outcome, a well-trained model, is encapsulated for easy access and distribution in the 'run.zip' file. Visionary Plates strives to provide a reliable and accurate license plate detection system, with the potential to significantly impact areas such as traffic monitoring, parking management, and law enforcement. The project signifies our commitment to innovation, pushing the boundaries of object detection technology to create practical solutions that make a difference in the real world.

AI Avengers

YOLOv5YOLOv8

Business Llama

📣 Exciting News from Business Llama! 📈 🚀 We're thrilled to introduce "Business Llama: Optimized for Social Engagement," our latest project that's set to transform the way you approach business planning and go-to-market (GTM) strategies. 🌟 🤖 With the power of advanced, fine-tuned models, driven by the renowned Clarifai platform, we're taking your business strategies to the next level. Here's what you can expect: 🎯 Enhanced Decision-Making: Make smarter, data-driven decisions that lead to business success. 📊 Improved Business Plans: Develop robust and realistic plans backed by deep insights. 🌐 Optimized Go-to-Market Strategies: Reach your target audience more effectively than ever before. 🏆 Competitive Advantage: Stay ahead in the market by adapting quickly to changing conditions. 💰 Resource Efficiency: Maximize resource allocation and reduce costs. 🤝 Personalization: Tailor your offerings to individual customer preferences. ⚙️ Scalability: Apply successful strategies across various products and markets. 🛡️ Risk Mitigation: Identify and address potential risks proactively. 🔄 Continuous Improvement: Keep your strategies aligned with evolving market conditions. Join us on this journey to elevate your business game! 🚀 Stay tuned for updates and exciting insights. The future of business planning and GTM strategies is here, and it's more engaging than ever. 🌐💼 #BusinessLlama #SocialEngagement #DataDrivenDecisions #Clarifai #GTMStrategies

Team Tonic

ClarifaiLlama 2OpenAIVercelCohere
Vercel
application badge

Lec2Learn - Finetuning AI Models

We present our solution Lec2Learn that works on Finetuning open source learning data for providing learning objectives. We start by obtaining all textbooks from opentextbookbc, we Process HTML to obtain the lecture and learning objectives, We then have pairs of lectures with their corresponding question groups, On the server we use Microsoft Phi 1.5 model and we fine tune it, We fine tune on the opentext data which is used so that model gets better at generating learning objectives, For the Prompt we give the lecture and learning objectives, we always start with Describe so model does not generate random data.

FineTuners

OpenAIFineTuner.aiGPT-3.5

"Really good use of Claude it is offering an immersive and interactive gaming experience "

avatar

Theodoros Ampas

Co-Founder of Content-Hive

"I'm excited to go on more quests like these! Would be really fun for multiple people to play this together. "

avatar

Mihir Chouhan

Project Manager

"Detective Quest is such a promising project! It's a fun and interactive way to play role games. Wishing you the best of luck!"

avatar

Paulo Almeida

co-founder of Stunning Green