Are LLMs the Beginning or End of NLP?

  Рет қаралды 25,665

Simons Institute

Simons Institute

8 ай бұрын

Dan Klein (UC Berkeley)
simons.berkeley.edu/talks/dan...
Large Language Models and Transformers
I'll talk about three major tensions in NLP resulting from rapid advances of large language models. First, we are in the middle of a switch from vertical research on tasks (parsing, coreference, sentiment) to the kind of horizontal tech stacks that exist elsewhere in CS. Second, there is a fundamental tension between the factors that drive machine learning (scaled, end-to-end optimization of monoliths) and the factors that drive human software engineering (modularity, abstraction, interoperability). Third, modern models can be stunning on some axes while showing major gaps on others -- they can, in different ways, simultaneously be general, fragile, or dangerous. I'll give an NLP perspective on these issues along with some possible solution directions.

Пікірлер: 16
@antonpictures
@antonpictures 8 ай бұрын
🎯 quick navigation: 00:00 🎙️ The speaker explores the future of NLP in the context of large language models (LLMs) like GPT. 08:07 🔄 The field is transitioning from vertical decompositions of tasks to a more horizontal tech stack approach, with LLMs as a key component. 09:31 🆕 LLMs introduced horizontal capabilities, such as portable knowledge and general ability, challenging traditional vertical approaches. 14:10 ❓ The current phase may mark the end of the beginning of NLP, but not the end of NLP itself; a shift from vertical to horizontal tech stacks is underway. 18:00 🧠 New control mechanisms and superstructures are emerging to guide and augment LLMs, providing more user influence and coherence. 19:56 📚 Language models can be orchestrated in an ecosystem where they collaborate hierarchically to generate stories or other outputs, resulting in more coherent and interesting content than single, uninterrupted model output. 21:30 🧩 Hybrid technology combining large language models with other methods, like search or structured modules, can achieve better performance in specialized tasks, such as crossword solving, compared to using language models alone. 22:56 🔌 Tension exists between modularity and end-to-end optimization in machine learning. Modularity allows systems to be built reliably, while end-to-end optimization enables high accuracy and generality. Exploring ways for these approaches to coexist could lead to more robust AI systems. 28:04 ⚖️ Responsible development of AI systems requires considering their failure modes, success modes, and potential to cause harm. Addressing plagiarism detection, authorship attribution, and safety features are crucial for creating AI tools that benefit society. 33:14 🔍 A system like "Ghostbuster" is developed to detect text generated by large language models by utilizing scores from weaker language models and arithmetic functions. The detection of language model-generated content, even in cases where the specific model isn't known, offers a tool for addressing potential misuse of AI-generated content. 39:42 🧠 Challenges of Large Language Models: Concerns about cheating, understanding the system's inner workings, and addressing biases in language models. 42:13 🤖 Interplay of Objective Functions: Highlighting the potential conflict between optimizing for objective functions (e.g., user satisfaction) and the truth in AI systems, leading to behavior that may deviate from truth. 45:01 🕵️‍♂️ Truth and Control: Emphasizing the need to distinguish between what the system knows and what it does, with a focus on identifying methods to evaluate and align AI models with the truth. 48:13 🌐 Multitask Model Vulnerabilities: Exploring the potential cross-task vulnerabilities in multitask models, highlighting the risks associated with poisoning one task that can impact multiple tasks. 51:14 🚀 Future of NLP: Reflecting on the evolution of NLP, from solving representational questions in linguistics to focusing on real-world problem-solving, acknowledging challenges in architecture, safety, and control. Made with coinmarketcap.com/currencies/sharbi/
@user-or4nj2zl5m
@user-or4nj2zl5m 2 ай бұрын
Awesome talk - Dan's the man!
@soumen_das
@soumen_das 8 ай бұрын
Very informative
@tahirsyed5454
@tahirsyed5454 8 ай бұрын
Prof Dan Klein - he'd make everything mentally accessible to you.
@yiran432
@yiran432 3 ай бұрын
is he in UCB?
@caten_8
@caten_8 5 ай бұрын
This is super insightful.
@mauricecinque5618
@mauricecinque5618 Ай бұрын
Interesting perspectives about architectures of NL systems. Worth questionning sustainability / maintainability / accuracy, and all other non functional requirements that contrain a fully operational solution.
@woolfel
@woolfel 8 ай бұрын
The job of a software engineer is to figure out what the human needs and build a solution. It's not writing a function. It's asking humans what they want and figuring out what they really need. The reason why so many software projects fails is simple, humans don't know what they really need and will tell you contradictory things about what they want. LLM can't read human's minds or figure out why humans are saying stupid contradictory things. At least not today, but maybe in the future it can read humans minds.
@matthewcurry3565
@matthewcurry3565 7 ай бұрын
You just said yourself... Even if it can read doesn't mean it can understand that answer, or/and cannot be fooled.
@mauricecinque5618
@mauricecinque5618 Ай бұрын
One of the key point raised in the talk relates to accuracy and (ground) truth. Dan taises the point of truth level implicitly reached in LLM because of the isomorphic relationship between words and real world, if i got his point correctly. That said, in LLM « truth » is essentially based on statistics. Is that really sufficient? Several AI have been purposedly trained with fundamental biases which are obviously distorsions and/or portions of reality.
@easyaistudio
@easyaistudio 7 ай бұрын
this started out great, then reason went on holiday when he started weighing into is using LLMs moral
@jayasimhatalur5503
@jayasimhatalur5503 8 ай бұрын
Jitendra Malik ❤
@opusdei1151
@opusdei1151 7 ай бұрын
Is this guy from the movie X+Y?
@RickySupriyadi
@RickySupriyadi 8 ай бұрын
not now... you can't replace human when making apps, sometimes so complicated at the end the engineer says ahhh i see and the client say ok that might work (wtf! how to explain it to him!) at the end of the project the apps launched and you know.... it sells because marketing team do they good job selling them, so everybody happy. until security bug came...
@namansinghal3685
@namansinghal3685 2 ай бұрын
Good AI engineers are good software engineers. I have seen a lot of AI folks struggle when they suck at coding.
@Idiomatick
@Idiomatick 7 ай бұрын
rip nlp
An Observation on Generalization
57:21
Simons Institute
Рет қаралды 152 М.
[1hr Talk] Intro to Large Language Models
59:48
Andrej Karpathy
Рет қаралды 1,8 МЛН
Why Only Us: Language and Evolution
1:21:19
Simons Institute
Рет қаралды 27 М.
CS Professor Sounds Alarm on AI and Programmers
12:21
Travis Media
Рет қаралды 279 М.
Scott Aaronson: From Quantum Computing to AI Safety
3:02:28
The Origins Podcast
Рет қаралды 22 М.