Alibaba releases new AI model Qwen2-VL that can analyze videos more than 20 minutes long (5 minute read)
Alibaba Cloud has released Qwen2-VL, a new vision-language model with enhanced visual understanding, video comprehension, and multilingual text-image processing. Qwen2-VL shows superior performance against models like Meta's Llama 3.1 and OpenAI's GPT-4o and supports various applications, including real-time video analysis and tech support. The models, available in three sizes (7B, 2B, and soon 72B), are open-source under Apache 2.0 for the smaller variants.
|
|
SGLang 0.3 (12 minute read)
SGLang inference improvements are here in the latest release, with 7x Faster DeepSeek MLA, 1.5x Faster torch.compile, Multi-Image/Video LLaVA-OneVision, and more.
|
|
💪 Small models, massive performance with OctoAI (Sponsor)
OctoAI empowers engineers to leverage small open-source models like Llama 3.1-8B, outperforming GPT-4o and dramatically reducing costs for enterprise tasks.The team just released a new tutorial to teach you how to:
- Apply advanced prompt engineering to slash expenses.
- Use parameter-efficient fine-tuning for optimal performance.
- Achieve GPT-4o quality with Llama 3.1-8B at a fraction of the cost.
Read the tutorial and get $10 in free credits on OctoAI super-performant endpoints right now.
|
Laminar (GitHub Repo)
Open-source observability, analytics, evals, and prompt chains for complex LLM apps.
|
|
Icon (Product Launch)
Icon helps brands partner with creators, turn 1 video into 20 videos with AI, and A/B test messaging to find winning ads.
|
|
Love TLDR? Tell your friends and get rewards!
|
Share your referral link below with friends to get free TLDR swag!
|
|
Track your referrals here. |
Want to advertise in TLDR? 📰
|
If your company is interested in reaching an audience of AI professionals and decision makers, you may want to advertise with us.
If you have any comments or feedback, just respond to this email!
Thanks for reading,
Andrew Tan & Andrew Carr
|
|
|
|