AI71 Falcon 2 11B VLM AI technology Top Builders
Explore the top contributors showcasing the highest number of AI71 Falcon 2 11B VLM AI technology app submissions within our community.
About Falcon 2 11B VLM
Falcon2-11B-VLM is an 11B parameters causal decoder-only model built by TII and trained on over 5,000B tokens of RefinedWeb enhanced with curated corpora. To bring vision capabilities, we integrate the pretrained CLIP ViT-L/14 vision encoder with our Falcon2-11B chat-finetuned model and train with image-text data. For enhancing the VLM's perception of fine-grained details w.r.t small objects in images, we employ a dynamic encoding mechanism at high-resolution for image inputs. The model is built on the same robust foundation as Falcon 2 11B, featuring 11 billion parameters. It matches or exceeds the performance of other leading models, such as Meta’s Llama 3 8B and Google’s Gemma 7B, particularly in tasks that require vision-language integration.
General | |
---|---|
Relese date | May 13, 2024 |
Author | AI71 |
Dataset | https://huggingface.co/tiiuae/falcon-11B |
Type | Vision Language Model |
Falcon 2 Tutorials
👉 Discover more Falcon 2 11B VLM Tutorials on lablab.ai
AI71 Falcon 2 11B VLM AI technology Hackathon projects
Discover innovative solutions crafted with AI71 Falcon 2 11B VLM AI technology, developed by our community members during our engaging hackathons.