Рет қаралды 10,480
In this video, we dive into the world of large language models (LLMs) and discover the optimal techniques for your specific tasks! Learn when to choose between training from scratch, fine-tuning, (advanced) prompt engineering and Retrieval Augmented Generation (RAG) with Activeloop’s Deep Memory. Equip yourself with the knowledge to enhance LLM performance, balancing quality, costs, and ease of use. ✨🚀
► Jump on our free LLM course from the Gen AI 360 Foundational Model Certification (Built in collaboration with Activeloop, Towards AI, and the Intel Disruptor Initiative): learn.activeloop.ai/courses/l...
With the great support of Cohere & Lambda.
► Course Official Discord: / discord
► Activeloop Slack: slack.activeloop.ai/
► Activeloop KZbin: / @activeloop
►Follow me on Twitter: / whats_ai
►My Newsletter (A new AI application explained weekly to your emails!): www.louisbouchard.ai/newsletter/
►Support me on Patreon: / whatsai
How to start in AI/ML - A Complete Guide:
►www.louisbouchard.ai/learnai/
Become a member of the KZbin community, support my work and get a cool Discord role :
/ @whatsai
Chapters:
0:00 When do what with LLMs?
0:20 What are the techniques available?
1:24 Improve your model with prompt engineering!
2:12 RAG and Deep Memory!
3:04 Fine-tuning LLMs (LoRa and QLoRa).
5:41 Training from scratch.
8:20 Conclusion.
#ai #languagemodels #llm