The best presenter I have ever listened to. Kudos! Perfect and purposeful use of tone, voice, and body language. Effective delivery. I learned more about presentation skills from this VDO than about prompt engineering lol. The team also did very well with filming, editing and directing. Well done!
@teenytinytoonsАй бұрын
thought i was watching an apple keynote. he's incredible.
@djarshad0094 ай бұрын
One of the best and most informative videos about Prompt Engineering. The presenter is extremely clear and simple to understand, his passion for AI is very prominent through his speech. Good Job
@teenytinytoonsАй бұрын
100000% agree.
@AIRecruiterTraining4 ай бұрын
Table of Contents - [00:00] Introduction to Prompt Engineering - Prompt engineering is designing prompts to get the desired output from conversational AI tools. - Prompt phrasing can influence the AI model's response, similar to how we use language in daily interactions. - [01:09] Motivation for Prompt Engineering - Yufeng, a Google engineer, was motivated by the inefficiency of getting useful responses from language models. - The goal is to improve the efficiency of conversational AI tools. - [03:00] Understanding LLMs for Effective Prompt Engineering - Understanding how LLMs work and their limitations is crucial for effective prompt engineering. - LLMs are AI models trained on massive amounts of text to identify patterns and generate responses to prompts. - [04:01] How LLMs Predict Words and Generate Text - LLMs are trained on vast amounts of text data to learn language patterns and predict the next word in a sequence. - They analyze probabilities to choose the most likely word based on the context, but may vary in their responses. - [04:57] ⚠ Limitations of LLMs: Bias and Unexpected Outputs - LLM outputs can be biased due to biases present in their training data. - You may not always get the desired output because LLMs rely on statistical analysis of training data. - [05:26] ⚠ Limitations of LLMs: Bias and Unexpected Outputs (continued) - LLM outputs can be factually inaccurate due to limitations in training data or its analysis methods (hallucinations). - Examples of hallucinations include generating incorrect historical data or company information. - [06:54] Critical Evaluation of LLM Outputs - It's crucial to critically evaluate LLM outputs for factual accuracy, bias, relevance, and information sufficiency. - This applies to all LLM outputs, regardless of the task (summarization, marketing ideas, project plans). - [08:22] Prompt Engineering for Effective LLM Use - The quality of the prompt affects the quality of the LLM output, similar to how high-quality ingredients affect a dish. - Prompt engineering involves designing clear, specific prompts with relevant context to get the desired output from an LLM. - [10:23] The Importance of Context in Prompt Engineering - Providing clear and specific context in prompts is crucial for getting the desired output from LLMs. - Example: A prompt about a professional conference will lead to different results than a prompt about a party. - [11:25] Prompt Engineering as an Iterative Process - Even well-designed prompts may require iteration to achieve the best results from LLMs. - The first attempt may not be perfect, so revising the prompt based on the initial output can be necessary. - [12:21] LLMs for Content Creation and Summarization Tasks - LLMs can be used to create content (articles, outlines, emails) and summarize lengthy documents. - Provide clear instructions and specify the desired task (create, summarize) in the prompt. - Example prompts are given for creating an article outline and summarizing a paragraph. - [14:24] LLMs for Text Classification and Data Extraction - LLMs can be used to classify text data (e.g., sentiment analysis of reviews) and extract information from text to a structured format (e.g., creating tables from reports). - Examples are provided for classifying customer reviews and extracting city and revenue data from a report. - [15:25] LLMs for Translation and Text Editing - LLMs can translate text between languages and edit text for tone, style, and grammar. - Examples are given for translating a training session title and changing the tone of a technical analysis. - [18:22] Iterative Prompt Engineering for Optimal Output - Prompt engineering is an iterative process, similar to developing a product or presentation. - It often requires multiple tries and revisions to the prompt before achieving the desired LLM output. - Don't be discouraged by initial failures; instead, evaluate the output and revise the prompt accordingly. - [19:19] Iterative Prompt Improvement for Better LLM Output - Getting the best output from an LLM often requires iteration on the prompt. - Evaluate the output after each attempt and revise the prompt to address shortcomings. - Reasons to revise the prompt include missing context, inaccurate information, or unsuitable formatting. - [21:42] Refining Prompts for Well-Structured Output - LLMs can be instructed to deliver output in a specific format (e.g., table). - Use clear and specific language to indicate the desired format in the prompt. - The example focuses on creating a well-organized table with relevant college information. - [24:11] Using Examples to Improve Prompt Efficiency - Including examples (or "shots") in prompts can significantly improve the quality of LLM output. - There are different prompting techniques based on the number of examples provided: zero-shot, one-shot, and few-shot. - Zero-shot prompts give no examples, while few-shot prompts provide two or more examples to guide the LLM. - [26:06] Using Few-Shot Prompting to Achieve Specific Style or Format - Few-shot prompting provides multiple examples (2+) to guide the LLM towards a desired output style or format. - The prompt should specify the number of adjectives, sentence length, and overall style desired. - The example focuses on creating a product description that matches the style of previous examples.
@abhijeetcreates4 ай бұрын
What an amazingly crafted learning video, this man has become my most favourite teacher (who has actually added to my skills). although i knew most of the concepts already , | "This lecture enhanced those skills like and Upgrade"... These 30 minutes are gonna help me out many times in the future . A Big Thanks to you. |
@teenytinytoonsАй бұрын
saved it my AI playlist as well
@capt20266 ай бұрын
Crystal clear presentation. Good job!
@Anthony-dj4nd6 ай бұрын
This guy is a great presenter
@hmnemonic6 ай бұрын
I love his voice…❤ Thanks for your presentation!
@Berserq6 ай бұрын
Seems like an AI
@looitaiyew61424 ай бұрын
Hi Yufeng, Many Thanks for the presentation.Well done.
@mahmoudabdelazizabdelaziz27937 ай бұрын
I am against using the word engineering in “prompt engineering”. I understand that coding is undergoing significant changes with the introduction of LLMs such as GPT, however, in my opinion we need to be more precise in choosing words that match the actual task being done by the software developer. Engineering design has a very precise meaning and it incorporates a lot of stages starting from setting quantitative measurable specifications, proposing multiple design alternatives, evaluating them against the specifications, iterative analysis and design usually involving mathematics and physics, selecting a design after examining trade offs among conflicting metrics, implementing the selected design, evaluating the implementation on different platforms (if feasible), module testing, integration testing, and more. Those engineering design stages constitute a cycle that can be revisited with each iteration including even the specifications stage which can be modified if necessary, but with caution and after agreement with the project stakeholders. In that context, software engineering for example might not satisfy all the above while being considered engineering nevertheless, but with some reservation in my opinion. However, prompt engineering barely includes a very small fraction of engineering design stages, and thus in my opinion should be given another name.
@zealousprogrammer45397 ай бұрын
I agree I am also against data "scientists" term in data science.
@andremaxville49366 ай бұрын
I always refer to it as prompt writing instead. It's not a job for a coder, but for a skilled writer.
@ErwinSalasErwin6 ай бұрын
Engineering is meant to solve problems with technology, since prompt engineering do it, it’s ok to say engineering
@Balloonbot6 ай бұрын
"Prompt pooping"
@otenyop5 ай бұрын
I agree 100%
@LAG4556 ай бұрын
Thank you for your teaching ☺️
@bahlechonco2117 ай бұрын
i love this this video but why are you crouching? the furniture looks small
@coolfrisbee7 ай бұрын
I created a prompt for a tiny chair
@bahlechonco2117 ай бұрын
@@OfficialUIUX 😂😂😂😂
@bahlechonco2117 ай бұрын
@@coolfrisbee 😂😂😂
@dasalsakid7 ай бұрын
😂😂😂😂 so true
@ailearnershub5 ай бұрын
Super'o'Super presentation and explanation and emphasis - Thanks Gentleman...
@krishangopal48087 ай бұрын
Can Gemini and ChatGPT help in writing research paper which can minimise Palagrism?
@richpoorworstbest48127 ай бұрын
I did exactly that
@PeterPan-hs5tu7 ай бұрын
my understanding of plagiarism is carbon copy of the exact combination of a collection of words from a publication made by an author. and what chatGPT capable of doing is to re-orienting the placement of the words. It directly challenges the very meaning of plagiarism as we know. it’s interesting to reflect on the fact how we define plagiarism, perhaps we have to move it’s goalpost in a post AI era 🤔 … or maybe it becomes the a concept of the past
@teenytinytoonsАй бұрын
you're only cheating yourself
@saurabhbanerjee299828 күн бұрын
Loved this presentation!❤
@Psrk42876 ай бұрын
What do you do when you don't know what the output should be? How will you use critical reasoning to evaluate the output then?
@bravosierra24475 ай бұрын
Good question . In my case I start with a general prompt that involves at least three values. From the output I get I refine my prompts until the output satisfies what I am looking for even when I start off unsure. Hopes this helps.
@Footballfantastic076 ай бұрын
thanks for this type of knowledge pls update us with type of videos in future
@GoogleCareerCertificates6 ай бұрын
Subscribe to our channel for updates and future videos!
@yogasetiawan208923 күн бұрын
does he use script for the presentation?
@vimalkumarv2 ай бұрын
Excellent teacher.
@AlliyuAbdullihiАй бұрын
I love this video and I happy ❤🎉
@begeshgangadharan6666 ай бұрын
There is some glitch in the Google map can you guys fix that?
@565856565876 ай бұрын
Your prompt did not include sufficient detail and requires multimodal approach.
@prashantsawant45876 ай бұрын
Good i description & details on how LLlM can be.. Thank You...
@manuelherrera56154 ай бұрын
Great job!
@PeterPan-hs5tu7 ай бұрын
amazing 🎉 thanks buddy
@chrisbrown45394 ай бұрын
"Shot" is synonymous w/ "example." 0-shot prompts provide no examples for the LLM. Non-zero shot prompts are a basic method to fine-tune LLM output, getting less generalized results.