- [00:04] Machine learning expectations - [01:42] Scale improves results - [04:25] Reversing ML capabilities - [07:24] Accuracy advancements in vision - [08:18] Speech recognition strides - [10:20] Hardware for ML efficiency - [13:57] TPU evolution for ML - [18:31] Distributed word representations - [20:56] Transformer model revolutionizes - [44:04] TPUs accelerate ML. - [48:08] General models specialized. - [49:17] Generative models for images. - [54:05] ML enhancing smartphones. - [57:06] ML revolutionizes science. - [59:06] ML aids medical diagnosis. - [01:02:08] Principles for ethical ML. - [01:06:03] More data improves. - [01:07:27] Plenty of data. - [01:08:07] Multimodal models benefit. - [01:09:27] Challenges for startups. - [01:10:51] Diversity in models.
@megairrational8 ай бұрын
Amazing! Thanks
@joeylantis226 ай бұрын
no need to thank people who post these anymore, they’re all generated with AI, not by hand like the early days of KZbin.
@AlexSuslin3 ай бұрын
@@joeylantis22 why not to thank people if they used AI to speed up their work?
@ericmcnally51289 ай бұрын
I really appreciate that lectures like these are made public for interested parties to consume. As an armchair enthusiast without access to quality institutional education on the topic, I have only been able to learn about speech and vision technology from publicly available talks like this. So thank you so much!
9 ай бұрын
There are free courses if you have time/are interested. But the math will definitely cross dimensions.
@ericmcnally51289 ай бұрын
This is actually where I am sort of stuck :not knowing which maths to focus on brushing up on.
@armin30579 ай бұрын
you gotta start somewhere, you could for instance download the math for machine learning book, called mml, free to download. dont worry if you dont get everything right, you find the solutions online too.@@ericmcnally5128
@sayo93949 ай бұрын
much appreciated for uploading this educational video onto KZbin and making it freely accessible to all.
@alexeikorol63129 ай бұрын
🎯 Key Takeaways for quick navigation: 00:04 *💡 Introduction to Machine Learning Trends * - Introduction to trends in machine learning that includes opportunities and challenges. - This talk presents the work of many people at Google, including some co-authored by Dean. 01:26 *👁️ The Evolution of Computer Perception* - The evolution of computer perception from rudimentary speech recognition and image processing to comprehending language, multi-lingual data, and enhanced perception. - Opportunities opened up by advanced machine senses in a variety of fields. 02:35 *💾 The Change in Computer Processing Needs Due to Machine Learning* - Transition from traditional code to machine learning analogies triggering a need for different kinds of hardware. - Discussion on how better hardware can lead to lower economic and energy costs while improving model quality. 06:14 *📈 Progress in Image Recognition and Speech Recognition* - Detailed on the history and advancements in image and speech recognition, with a focus on neural networks. - Importance of continued research and scaling for improving accuracy and usability of systems. 10:20 *🖥️ Google's Development of Tensor Processing Units* - Google’s development of Tensor Processing Units (TPUs) to serve machine learning models effectively. - TPU's ability to carry out reduced precision computations and assemble different linear algebra operations. 14:11 *📚 Language Processing with Machine Learning* - The radical changes and opportunities in language processing using machine learning. - The use of distributed representations to represent words in high dimensional vectors which can push similar words closer in space and separate different words. 18:31 *⚙️ Sequence to Sequence Learning Model* - Introduction to the Sequence to Sequence learning model which can translate input sequences into different languages. - The model's ability to absorb input sentences and decode the corresponding translated sentence iteratively. 20:27 *🔃 Multi-Turn Conversations with Neural Language Models* - Discusses the use of sequence-to-sequence models for creating meaningful multi-turn conversations, where the system takes context of previous interactions into account. - The model effectively uses input from the previous turns of conversations to generate appropriate responses. 21:06 *🔂 Transformer Model for Parallel Processing* - Explanation of the Transformer model, shifting from a sequential processing approach to a parallel one. - The Transformer model allows processing of each word in input independently and uses "attention" to focus on relevant parts during translation. - This shift led to a significant improvement in accuracy with more computational efficiency. 23:17 *🚀 Scaling Up Models and Obtaining Better Results* - Talks about the trend of increasing the scale of models and using Transformer models for training on conversational style data. - Describes how model evaluation ensures that generated responses are both sensible and specific. - Overview of the progression of neural language models and their improvements over time. 25:31 *🧩 Multimodal Models for Various Inputs* - Moving towards multimodal models that can process different types of inputs such as images, text, and audio. - Establishes the goal to train the world's best multimodal models and use them across Google products. 28:31 *💡 Scaling Up the Training Infrastructure and Data Handling* - Explanation of how Google's training infrastructure operates at a large scale to map computations onto available hardware. - Discusses the importance of having a fast recovery system to mitigate the impact of system failures. - Describes the critical role of high-quality training data and proposes automated learning curriculums as a future research area. 33:07 *🧠 Better Model Elicitation and Multimodal Reasoning* - Proposes techniques to elicit better responses from models by asking them to "show their work". - Details examples of multimodal reasoning through the Gemini model, where complex problems are solved using various data types (texts, images, etc.). - The potential educational implications of these advancements are highlighted, such as individualized tutoring. 38:07 *📊 Evaluation of Models and Performance Comparisons * - Emphasizes the importance of model evaluation in identifying strengths/weaknesses, ensuring a well-trained model, and comparing its capabilities with other models/systems. - Presents performance statistics of the Gemini Ultra model, indicating its state-of-the-art performance in multiple areas of evaluation. 40:35 *🏆 State-of-the-Art Performance Across Benchmarks* - The Gemini model demonstrates state-of-the-art performance across multiple benchmarks. - Achieved top results in text, image, video, and audio understanding, even on data it had not previously encountered. 42:25 *💬 Coherent Conversations and Advanced Capabilities* - The model can generate surprisingly coherent conversations and offer domain-specific knowledge. - Showcases various input requests and the model’s impressive capacity to carry out meaningful, accurate responses. 44:45 *🤖 Chatbot Integration and Performance Measures* - Integration of the Gemini Model into the Bard chatbot. - Evaluation of the model's performance using the ELO scoring method. 49:17 *🖼️ Generative Models for Image Production* - Overview of how generative models are used to create images based on detailed prompts. - Emphasizes the influence of model scale on the quality of generation. 54:05 *📱 Machine Learning Applications in Everyday Tech* - Discussion of various machine learning features that have improved functionalities in smartphones, particularly in camera features. - Stresses the broad implications of machine learning technology for areas like language translation or literacy support. 57:06 *🧪 Machine Learning Potential in Material Science and Healthcare* - Discusses how machine learning can aid in exploring scientific hypothesis spaces and searches for new materials. - Highlights the potential of machine learning in healthcare, particularly in medical imaging and diagnostics. 01:01:27 *🩺 Machine Learning in Dermatological Diagnostics* - Utilizes machine learning to assist in diagnosing dermatological conditions. - Users can take a photo of their skin concern, and the system can provide potential diagnoses based on similar images in dermatological databases. 01:02:08 *📚 Guidance Principles for Applying Machine Learning* - Google's published a set of principles in 2018 to guide internal teams on considerations when applying machine learning to problems. - Highlights important aspects such as avoiding creating or reinforcing unfair bias and being accountable and sensitive to privacy. - Active areas of research include fairness, bias, privacy, and safety. 01:04:29 *🌟 Future Prospects and Responsibilities in Machine Learning* - The capacity for computers to understand various modalities and react accordingly is constantly expanding, paving the way for more intuitive and seamless user experiences. - The responsibility to harness machine learning for social benefit is proportional to the opportunities it presents. 01:05:49 *💡 Q&A Session* - Discussion about the impact of more data on the performance of machine learning models. - Views about the evolution of LLMs and the growth of multimodal models. - Reflections on the accessibility of machine learning research for small startups and individuals. Made with HARPA AI
@brookshamilton18 ай бұрын
Very helpful! Thank you
@jasonsong40359 ай бұрын
🎯 Key Takeaways for quick navigation: 00:04 *🤖 Introduction to Machine Learning Trends* - Overview of exciting trends in machine learning, - Jeff Dean highlights the broad impacts and opportunities in AI, - Importance of awareness in technological development. 00:43 *🌐 Evolution of Machine Learning Capabilities* - Machine learning has transformed our expectations of computers, - Significant improvements in speech recognition, image understanding, and language processing, - Transition from basic functionalities to advanced perception and interaction. 02:07 *⚙️ Scaling and Hardware Innovations* - Scaling up computing resources leads to better machine learning performance, - The shift towards specialized hardware for more efficient computations, - Larger datasets and models contribute to advancements in AI capabilities. 03:57 *🔄 Reversibility in Machine Learning Models* - Recent progress in reversing traditional input-output relationships in AI models, - Examples include generating images from descriptions and converting text to speech, - These advancements open up new possibilities for creative and practical applications. 05:06 *📈 Benchmark Improvements Over the Decade* - Significant improvements in image recognition and speech recognition benchmarks, - The evolution of machine learning models has led to surpassing human accuracy in certain tasks, - Continuous advancements underscore the rapid development in the field of AI. 08:33 *🖥️ Specialized Machine Learning Hardware* - The development of hardware optimized for machine learning, like Google's TPU, - Improvements in computational efficiency and energy consumption, - The role of reduced precision and linear algebra in machine learning computations. 13:57 *🗣️ Advances in Language Understanding* - Significant progress in language models and translation, - From basic n-gram models to advanced neural network-based approaches, - The importance of distributed representations and sequence-to-sequence learning in improving language understanding. 20:27 *💬 Advancements in Conversational AI* - Introduction to effective multi-turn conversations using neural language models, - The progression from sequence-to-sequence models to more advanced Transformer models enabling parallel processing for efficiency and accuracy. 23:57 *🗨️ Evolution of Neural Language and Chat Models* - Overview of the development in neural language models and chatbots, including GPT and BERT variations, - Emphasis on the transformative impact of the Transformer architecture on model efficiency and capability. 26:00 *🌐 Introduction to Gemini Multimodal Models* - The goal of creating multimodal models capable of understanding and generating content across various data types, including text, images, and audio, - The introduction of Gemini models by Google for enhanced AI capabilities in handling multiple modalities simultaneously. 28:16 *⚙️ Scalable Training Infrastructure and Data Quality* - Discussion on the scalable training infrastructure designed to efficiently map computations onto available hardware, - Emphasis on data quality and its critical role in model performance, including strategies for enhancing training data relevance and richness. 33:07 *🧠 Techniques for Eliciting Better Responses from Models* - Introduction of techniques like Chain of Thought prompting to improve model accuracy and interpretability, - Examples demonstrating how guiding models to "show their work" can significantly enhance performance on complex tasks. 35:50 *🤖 Multimodal Reasoning in Gemini Models* - Presentation of Gemini's capabilities in multimodal reasoning with an example of solving a physics problem, - Discussion on the potential of multimodal AI models like Gemini for personalized educational tools and tutoring. 38:21 *📊 Evaluation and Performance Benchmarking of Gemini Models* - Overview of Gemini's evaluation process and its performance across various academic benchmarks, - Comparison of Gemini Ultra with other state-of-the-art models, highlighting its superior performance in a majority of evaluated tasks. 40:35 *🏅 State-of-the-Art Benchmarks in Image, Video, and Audio Understanding* - Gemini's exceptional performance on various benchmarks, including image, video, and audio understanding, - Achievements in multimodal capabilities with state-of-the-art results across multiple domains, - Importance of unbiased benchmark testing to validate model capabilities. 42:25 *💡 Conversational AI and Practical Applications* - The evolution of conversational AI models leading to coherent and helpful interactions, - Examples of Gemini's capabilities in providing detailed, context-aware responses in a conversational setting, - Introduction of programming concepts and detailed explanations as part of AI-generated responses. 48:08 *🏥 Domain-Specific Model Refinements for Medical Applications* - Refining general models for domain-specific applications, particularly in the medical field, - Achievements of the Med-PaLM model in exceeding medical board exam benchmarks, - Potential of domain-enriched training to achieve expert-level performance in specialized areas. 49:17 *🎨 Advances in Generative Models for Images and Video* - Development of generative models capable of creating detailed and contextually accurate images from textual descriptions, - Impact of model scaling on the fidelity and accuracy of generated images, - Integration of generative models into practical applications for creative and educational purposes. 54:05 *📱 Machine Learning in Everyday Devices* - The invisible role of machine learning in enhancing smartphone features and user experiences, - Examples of computational photography, live captioning, and language translation powered by AI on mobile devices, - The potential of AI to assist users in a variety of practical and accessibility-oriented tasks. 57:06 *🔬 Machine Learning in Material Science and Healthcare* - The influence of machine learning on scientific research, particularly in material science and healthcare, - Automated discovery of new materials with desirable properties using AI-driven simulations and structural pipelines, - The application of machine learning in medical diagnostics, with a focus on diabetic retinopathy and dermatology screening. 01:01:27 *📸 AI in Dermatology* - Deployment of AI systems for dermatological assessments through smartphone photography, - The system's capability to match user-uploaded images with dermatological databases for condition identification, - Emphasis on the potential for AI to distinguish between serious and benign skin conditions. 01:02:08 *🤖 Ethical Principles in Machine Learning* - The importance of ethical considerations and principles in the application of machine learning technologies, - Google's publication of AI principles to guide responsible development and usage, - Focus on avoiding bias, ensuring accountability, and enhancing social benefits through AI applications. 01:04:29 *🚀 Future of Computing with Learned Systems* - The shift from encoded software systems to learned models that interact more naturally with humans and the world, - The expanding capabilities of computers to understand and generate various modalities like speech, text, and images, - Discussion on the opportunities and responsibilities in advancing AI to ensure social benefits. 01:07:13 *💡 Data Quality and Model Performance* - The relationship between data quality, model capacity, and performance, - The importance of high-quality data and appropriate model scaling for improved AI effectiveness, - Mention of potential adverse effects of low-quality data on model capabilities. 01:08:07 *🧠 The Future of Large Language Models (LLMs)* - Discussion on the future of LLMs and the availability of high-quality training data, - Exploration of untapped data sources like video for further training and development of LLMs, - The ongoing potential for significant advancements in AI through diverse data utilization. 01:09:13 *🌐 Multimodal Models and Specialized Applications* - The impact of multimodal models on performance across different domains, - Considerations on whether multimodal models outperform domain-specific models in their respective areas, - The potential of base models enriched with domain-specific data for targeted applications. 01:10:09 *🚀 Opportunities in AI Research for Individuals and Startups* - Encouragement for individuals and startups with limited resources to engage in innovative AI research, - Highlighting the potential for significant contributions to AI through clever ideas and efficient use of available computational resources, - The importance of diversity in research topics within the AI field, beyond large-scale model training.
@varshneydevansh9 ай бұрын
Finally, the word Machine Learning rather than AI, AGI, and LLM
@nicohornswag9 ай бұрын
A breath of fresh air
@Prometheus707479 ай бұрын
All categories of machine learning. Ain't rocket science. There are different categories of any field of science or engineering. What do you expect?
@Prometheus707479 ай бұрын
All categories of machine learning. There other types of machine learning categories so it's sort of necessary to categorise subsets of machine learning, which is an umbrella term for different categories.
@vsenderov9 ай бұрын
I think AI is a superset including ML and GOFAI and oddballs like the Semantic Web.
@idrisssarmouk80449 ай бұрын
We dont even have a definition of intelligence yet
@Dima-rj7bv9 ай бұрын
Surprisingly amazing questions in the end. Right to the point.
@Noobsitogamer109 ай бұрын
Love the passion Jeff has for machine learning.
@collins43599 ай бұрын
love jeff. great speaker
@dr.mikeybee9 ай бұрын
One thing that doesn't ever seem to get mentioned when discussing the context signatures, a.k.a. dense representations, in sequence-to-sequence modeling for translation is that different languages' semantic spaces have much of the same shape. This is a function of the fact that different languages all model our shared experiences. And despite misalignments in the semantic spaces, there's still enough similarity (i.e., the shapes are close enough) to make translation possible.
@quasimot89 ай бұрын
Huh, that's fascinating
@stevo-dx5rr9 ай бұрын
It’d be interesting to look at the differences.
@planetsharks47467 ай бұрын
Underrated talk from the GOAT
@sombh19719 ай бұрын
37:16 Wow! There goes education! 44:03 There goes coding!
@wolpumba40998 ай бұрын
*Abstract* This comprehensive video presentation delves into the current state and future prospects of machine learning (ML), underlining significant advancements and the technological evolution that has shaped the field. The talk begins with an overview of machine learning trends, emphasizing the dramatic improvements in speech recognition, image understanding, and natural language processing over the last decade. It attributes these advancements to increased computing resources, specialized hardware, and larger datasets. A notable highlight is the development of Google's Tensor Processing Units (TPUs), designed to optimize ML computations efficiently, showcasing the importance of scalable and efficient hardware in pushing the boundaries of ML capabilities. The discussion progresses to the hardware evolution, with the latest TPUs achieving 1.1 exaFLOPS of computational power, and introduces the V5 series, enhancing performance for both inference and training. Attention is given to the strides in language models and translation, detailing the shift from traditional algorithms to neural networks and the transformative impact of models like Transformer, which allows parallel data processing for improved accuracy and efficiency. Central to the presentation is the unveiling of Gemini, Google's ambitious multimodal model, aimed at mastering the integration of text, image, video, and audio data. Gemini's varying sizes cater to different applications, from powerful cloud-based solutions to on-device implementations. The model's training, data filtering, and quality assurance processes are discussed, alongside innovative techniques like "Chain of Thought" prompting for eliciting more accurate and interpretable responses from the model. Performance evaluations reveal Gemini's superior capabilities across a wide range of benchmarks, outperforming state-of-the-art models in text, image, video, and audio understanding, as well as in conversational AI. The talk further explores the application of machine learning in enhancing smartphone features, material science, healthcare, and raises ethical considerations vital for responsible ML deployment. The session concludes with a Q&A segment addressing the audience's inquiries on model performance improvement with high-quality data, the future of large language models, the comparison between multimodal and domain-specific models, accessibility of AI research for individuals and startups, and concerns regarding the diversity of machine learning models. This presentation underscores the remarkable journey of machine learning, highlighting Google's leading role in advancing the field, and points towards a future where ML's potential to benefit society is fully realized, provided it is used responsibly. *Summary* *Introduction and Observations on Machine Learning* - *0:04* Introduction to trends in machine learning, its significance, opportunities, and considerations. - *0:22* Acknowledgment of Google's collective work in machine learning. - *0:48* Initial observations on machine learning improvements in speech recognition, image understanding, and natural language processing. - *1:59* Mention of the role of computing scale, specialized hardware, and large datasets in enhancing machine learning results. *Progress and Developments in Machine Learning* - *3:11* Examples of progress in image classification, speech recognition, and translation. - *4:17* Discussion on reversing machine learning processes for image generation from descriptions. - *5:13* Progress in image recognition accuracy, highlighted by ImageNet benchmark. - *7:42* Significant improvements in speech recognition accuracy. - *8:37* The importance of scalable and efficient hardware for machine learning. - *9:17* Benefits of reduced precision and focus on linear algebra in neural networks. *Hardware Innovations and Computing Power* - *10:27* Introduction to Google's Tensor Processing Units (TPUs) for efficient machine learning computation. - *12:02* Scaling with TPU pods for enhanced machine learning capabilities. - *12:58* Describes computing power in data centers with 1.1 exaFLOPS of computation. - *13:15* Introduction of the V5 series TPUs with enhanced memory and bandwidth. *Advances in Models and Translation* - *14:00* Advances in language models beyond traditional areas. - *18:31* Introduction to sequence learning and neural networks for translation. - *21:13* Explanation of the Transformer model allowing for parallel data processing. - *23:52* Evolution of neural language models and conversational AI, including developments in GPT and Transformer models. *Gemini: A Multimodal Model by Google* - *25:54* Introduction to Gemini models aiming to lead in multimodal machine learning. - *28:16* Training infrastructure and focus on maximizing "goodput." - *31:34* Importance of data quality and filtering for Gemini's training. - *33:19* "Chain of Thought" prompting technique for improved model performance. - *35:53* Multimodal reasoning capabilities of Gemini, with applications in education. *Performance and Applications of Gemini* - *39:14* Performance of Gemini Ultra in benchmarks. - *42:27* Conversational capabilities and development of domain-specific models. - *49:17* Generative models for creative image and video generation. - *53:01* Machine learning advancements in visual recognition and its applications in various fields. *Ethical Considerations, Conclusion, and Q&A* - *1:02:02* Emphasis on ethical considerations and responsible use of machine learning. - *1:04:18* Conclusion highlighting the shift to learned systems and their societal potential. - *1:05:38* Speaker's decline of further questions due to overwhelming response. - *1:06:13* Audience questions on model performance, future of LLMs, multimodal models, accessibility of AI research, and diversity in machine learning models. Disclaimer: I used gpt4-0125 to summarize the video transcript. This method may make mistakes in recognizing words and it can't distinguish between speakers.
@George-nd7ot9 ай бұрын
This man made Google. LEGEND
@CharlesVanNoland9 ай бұрын
Well that was Larry Page and Sergey Brin but Jeff Dean did join early on in 1999 about a year after Google was founded - which was definitely before I'd even started using Google. I was still using excite and altavista until at least 2000 or 2001. Google was just way better at finding whitepapers and scientific content than any other search engine I'd ever used before.
@edgedg9 ай бұрын
@@CharlesVanNoland He means "technologically", not the organization. Dean was a true programming monster for making things run fast, and almost every Google project he touched was a game changing success.
@doa_form8 ай бұрын
this is more of a historical look rather than a focust toward the future
@TopicalAuthority19 күн бұрын
Thanks for the sharing.
@ArooPC9 ай бұрын
Wish he was asked on the reason for why they compared CoT 32-shot with 10-shot. And if they're developing a Tranformer successor or tried out Mamba.
@BennduR9 ай бұрын
36:56 the length of the slope is not the hypotenuse, it's simply the length L in the diagram aka 80
@BrianMosleyUK9 ай бұрын
This was posted 3 days ago, shame he wasn't able to discuss Gemini 1.5
@alph49669 ай бұрын
Jeff Dean's talent is real. but there is a CEO at Google who is crushing that talent.
@FoxnewsFan9 ай бұрын
Also their horrible AI product and marketing team
@ai_outline9 ай бұрын
Computer Science is awesome!!
@Fordtruck4sale9 ай бұрын
Thanks Jeff (If you're reading this) it clarified how embeddings work for me, on top of a lot of other things. Great presentation!
@scign9 ай бұрын
Love how he completely skips over GPUs and the link to gaming and video! Goes straight from CPU to TPU.
@oscar12.349 ай бұрын
Way better than most online services I've used for improving quality.
@AravindanUmashankar9 ай бұрын
Very insightful
@raxcoins9 ай бұрын
i would like to ask for a better example. jeff is a great person, and he has more to say, with prowess to back it. i just dont see him putting much time into these as he should be.
@mikaelfiil37338 ай бұрын
Well, at 47 min. Genini gets the ordering wrong, when it states the countries with the most companies per 1 million residents. Either the table is wrong or the US should go last with 44.16. He didn't even notice it when reading the text below the table. So how can you trust it?
@liubianxing9 ай бұрын
thx and appreciate for this sharing, Gemini and multi-model could bring along new trend in 2024
@deniz.72009 ай бұрын
Gemini is the worst
@mnchester9 ай бұрын
Amazing talk! 😊
@denmark8th9 ай бұрын
super awesome talk!!
@iker649 ай бұрын
Wonder what's next for AI?
@Elpapucho179 ай бұрын
I second that, their upscaling is really top notch.
@istvandarvas33729 ай бұрын
this was amazing! Thanks!
@fearlesspony9 ай бұрын
@ [18:04] - they got the vector "king - queen" pointing in the wrong direction.... same for "man - woman"
@paragmahajani54503 ай бұрын
P7😂
@정창욱-z6v9 ай бұрын
멋있는 분이시네.
@skavihekkora50399 ай бұрын
Timestamps please!
@skypickle299 ай бұрын
who is the host?
@RiceKenKennedyInstitute9 ай бұрын
This lecture was hosted by the Ken Kennedy Institute, an interdisciplinary group at Rice University that works collaboratively on groundbreaking research in artificial intelligence, data, and computing. Visit kenkennedy.rice.edu/ to learn more about our events and activities!
@ps33019 ай бұрын
Jeff should start to build a video transformer to compete with sora.
@bartoszkunat11349 ай бұрын
46:00
@Papu192719 ай бұрын
This guy dreams in Python code for sure.
@Epistemophilos8 ай бұрын
"Avoid creating or reinforcing unfair bias." Sure.
@nbansal9 ай бұрын
Promised - Self-Driving Cars Result - Here is a cute video of dog 😅😅
@sampirat28 күн бұрын
This look like a google portfolio sales exposee rather than an overview of curent trades in the field. Would the title had reflected what this lecture realy is me and i'm sure many otehr people genuinly interested in having a quick tour of exciting trends in machine learning would not have waste their time. But i guess that should have been expected from google as their view of AI is how to build the ultimate bulshiter (yes that's what they pollitically correcltly call "conversational agents"), so they very logically sent someone to sell their bulshit. However i would have expected more from a respectable academic institute.
@chromosome249 ай бұрын
Its interesting to see how poorly these transformer models perform at math.
@pieter54669 ай бұрын
1:03:04 "the world as we'd like it to be" - who's "we"?
@Prometheus707479 ай бұрын
Other people
@prashantmorgaonkar30959 ай бұрын
TLDR; please 😢
@bartoszkunat11349 ай бұрын
20:09
@RobertMaximVerdes9 ай бұрын
17:10 king minus queen roughly equals man minus woman
@mikhailbaalberith9 ай бұрын
My boy lost a good opportunity to talk less about Gemini and more about PIML
@jack_galt9 ай бұрын
1:11:58 Mhm mhm mhm
@darylltempesta9 ай бұрын
Why hide the wisdom of the audience?
@hanchisun61649 ай бұрын
Seeing Jeff himself lying about the numbers of Gemini, I am starting to think it is not necessarily Pichai's fault for corrupting the company's culture.
@pillescasdies8 ай бұрын
Sincere Q: what part was a lie? The elo results?
@theyruinedyoutubeagain9 ай бұрын
I love Jeff, but OpenAI is kicking their ass with their crazy iteration speed, while Google is hamstrung by their incompetent leadership and can't focus or even deliver something reasonable. I keep saying that Sundar should've been dishonourably ejected years ago.
@deniz.72009 ай бұрын
Google Gemini is the worst AI I have ever used, its a waste of time and money sadly
@ellielikesmath8 ай бұрын
ugh so boring
@lukeb1118 ай бұрын
About as useful as bard
@JoseJimenez-il5vs9 ай бұрын
Still waiting for Jeff to apologise to Timnit Gebru. Absolute clown.