25
8
Pakistan
1 year of experience
Hi, I am Muhammad Asad Ishfaq. I'm Pursuing the degree of BS Artificial Intelligence. I'm a beginner in the field of AI and machine learning, but I'm eager to learn and grow. I have some basic knowledge of Python, data analysis, Python libraries, machine learning, and Microsoft Azure. I'm particularly interested in using large language models (LLMs) to solve real-world problems. I'm looking forward to collaborating with other participants and mentors to develop innovative solutions. I'm also passionate about: Solving real-world problems with AI, Collaborating with others, Learning new things I'm looking forward to learning and growing with you all!
Long Description: At the heart of our vision is the belief that generative AI technology has revolutionized the film industry, opening doors for virtually anyone to step into the role of a movie producer. Our project in this Hackathon aims to validate this concept by crafting a concise, 35-second film "composed of shorts," leveraging cutting-edge generative AI tools, namely Custom GPT and Stable Video Diffusion, in line with the competition's guidelines. Our methodology encompassed four distinct phases: Utilizing our Script Writing Custom GPT to generate the film script and accompanying images. Employing Eleven Labs for the audio narration of our script. Transforming the generated images into brief, 5-second video segments using a Stable Diffusion Image 2 Video XT tool from HuggingFace. Assembling these segments in Filmora to produce a cohesive 35-second film, complete with narration and sound effects. Overall, the execution was remarkably smooth. We navigated technical hurdles with ease, and our strategy of utilizing autonomous agents for script and image creation proved to be a groundbreaking approach for swiftly developing scripts for short films.
Quantum Blend is a groundbreaking project that explores the fusion of text and images using the cutting-edge Gemini AI model. Our vision is to create a platform where language transcends traditional boundaries, intertwining with visually stunning elements to evoke a new level of engagement and understanding. Quantum Blend's innovative approach aims to revolutionize communication, storytelling, and artistic expression, setting the stage for a future where AI becomes a true creative collaborator. Join us on this transformative journey, where Quantum Blend opens doors to unparalleled possibilities in the world of multimodal AI.
The ongoing dialogue between humans and AI not only showcases the remarkable capabilities of current technologies but also illuminates the future possibilities of AI-human synergy, promising an era where AI enhances human creativity, decision-making, and problem-solving in unprecedented ways. Our hackathon project explored the interaction between humans and Large Language Models (LLMs) over time, developing a novel metric, the Human Interpretive Number (HIN Number), to quantify this dynamic. Leveraging tools like Trulens for groundedness analysis and HHEM for hallucination evaluation, we integrated features like a custom GPT-5 scene writer, the CrewAI model translator, and interactive Dall-E images with text-to-audio conversion to enhance understanding. The HIN Number, defined as the product of Groundedness and Hallucination scores, serves as a new benchmark for assessing LLM interpretive accuracy and adaptability. Our findings revealed a critical inflection point: LLMs without guardrails showed improved interaction quality and higher HIN Numbers over time, while those with guardrails experienced a decline. This suggests that unrestricted models adapt better to human communication, highlighting the importance of designing LLMs that can evolve with their users. Our project underscores the need for balanced LLM development, focusing on flexibility and user engagement to foster more meaningful human-AI interactions.