ACL is not an AI Conference (?)
Yoav Goldberg, August 2024
In her âPresidential Addressâ at the ACL 2024, Emily Bender gave a talk called âACL is not an AI Conferenceâ. For those who did not attend (or were not paying close attention), you can find the slides in the following link: https://faculty.washington.edu/ebender/papers/ACL_2024_Presidential_Address.pdf
Somewhat surprisingly, I found myself agreeing with some core aspects of her argument. Perhaps less surprisingly, there is also a substantial part which I strongly disagree with. This text is a response to this address, and, beyond just responding, may also shed some light on what is ACL, and what is NLP. I of course welcome discussion on these topics, either on the comments section here (unfortunately not very convenient) or on Twitter (not convenient in a different way). Ok, Letâs go.
ACL is not a Computational Linguistics Conference
First, before we go into the core of the issue, I will touch on something a bit tangential. Throughout the talk, Emily referred to ACL as a âComputational Linguisticsâ or âComputational Linguistics or NLP Conferenceâ. This was very intentional, perhaps in a Sapir-Whorfian attempt to change reality.
But letâs get this out of the way: unless we redefine what Computational Linguistics means to mean âNLPâ (and I really donât think Emily intended to make such redefinition), ACL is in no way or form a Computational Linguistics conference. Yes, it has CL in its name, that is an unfortunate historical artifact.
But ACL does not deal with computational linguistics (a sub-field of linguistics that attempts to answer linguistic questions using computational methods), but with NLP (computational methods of transforming, âunderstandingâ or generating natural language).
These are not the same thing. In fact they are very different things. While an argument can be made regarding having both topics in the same conference (an argument can also be made against it), the fact is that computational linguistics works are not presented at ACL, will have a very hard time getting accepted to ACL, and if such a work happens to be accepted to ACL, the authors will fill very lonely at the conference. There in fact is a venue for actual computational linguistic works, it is called SCiL (âSociety for Computation in Linguisticsâ), and there is a community that publishes there.
Pretending that ACL is also a form of computational linguistics venue is just silly, and beyond silly, I believe it is also harmful to the real computational linguistics community, who really should have their own venue where they can publish, be read, and attend conferences with people who care about and are excited about their work, and not have to occasionally submit to ACL by mistake just to be confused by the reviews.
Having that out of the way, letâs go on to the main issue.
The Good, the Straw-man, and the Ugly
The Good
Emily titled her talk âACL is not an AI Conferenceâ, I strongly agree with that. ACL is an NLP conference, and NLP is not AI.1
As Emily said, âLanguage processing is a prerequisite for âAIâ, but that doesnât mean that âAIâ is the only goal of CL/NLPâ.
I strongly agree with this as well (if we ignore that âCLâ that got inserted there): I certainly did not enter the field because I cared about AI. I entered the field to do NLP: to understand how to build algorithms that work with human language, and to enable applications that process and/or interact with human language.
So, yeah, ACL is not an AI conference. ACL is an NLP conference. NLP is great.
Emily also gives a âbest practices in CL2/NLP researchâ (slides 10-15). I agree with these as well, of course (though once we recast âKnowledge of How Languages Workâ as âKnowledge of the domain one is working inâ, we are left with just a list of good overall scientific and engineering practices. Which is great, but not really about NLP).
But the full title of this slide is âContrast with best practices in CL/NLP researchâ, where the contrast is âAI researchâ which leads me toâŠ
The Straw-man
A major issue with the piece is that it attacks a straw-man. In slides 5-7 Emily explains what she means by âAIâ in âACL is not an AI conferenceâ.
AI as a research & commercial field Asks questions like:
- How do we build âthinking machinesâ that can do âhuman-likeâ reasoning?
- How do we build âthinking machinesâ that can âsurpassâ humans in cognitive work?
- (and cure cancer, solve the climate crisis, make end-of-life decisions, etc)
- How do we automate the scientific method?
- How do we automate away such creative work as painting and writing?
- Or: How do we steal artwork at scale and try to convince people this is âfor the common goodâ?
Slide 5 is reasonable, though somewhat reductive of what AI is or aims to do as a research field (well, very reductive).
The last bullet (and especially its sub-bullet) then ventures into âcommercial AIâ, which leads us to slides 6 and 7:
AI as a research & commercial field
And makes assertions like:
- HumanityÊŒs destiny is to merge with machines and become âtranshumanâ
- The singularity is coming: âAGIâ is inevitable and will outstrip people in all ways that matter
- âAIâ (really synthetic text extruding machines) is a suitable replacement for the services we owe each other (education, healthcare, legal representation)
- All of this is inevitable and refusal is futile
AI as a research & commercial field
Suffers from multiple scourges:
- Intense (though maybe waning?) interest from venture capital
- Intense (and not waning) interest from billionaires
- The racist history and present of the notion of âintelligenceâ/IQ
- Intense interest from proponents of TESCREAL ideologies (Gebru & Torres 2024)
Here it becomes clear that Emily has in mind a very particular notion of âAIâ, the one that is peddled by doomers and/or billionaires and which aligns with âTESCREAL ideologiesâ. Well, sure, there are some commercial parties, and some individuals, who subscribe to this version of âAIâ. But these are, by and large, not what the academic field of âAIâ is concerned with. By this definition of âAIâ, ICLR, NeurIPS, ICML, and also IJCAI and AAAI are not AI conferences. Indeed, by this definition there are no academic AI conferences (well maybe NeurIPS is. But then, I did say academic conferences ;) ). And since all the argumentation is built around contrasting with this very specific---and somewhat irrelevant to our context---definition of AI, things fall apart very quickly for me.
I get it, the commercial AI trends are bad, TESCREAL ideologies may be problematic. Certainly Emily is part of a campaign against them, and this campaign may be justified (I am not judging). But what does it have to do with ACL, with NLP, or with Academic AI and ML more generally? And why mix the ideological campaign with what NLP is or is not? (and is Emily conflating this wacko definition of AI with âLLMsâ? hmmm)
This is just a bad argument.
The second-to-last slide lists âACL 2024 papers that have nothing to do with AIâ. I think she meant âHave nothing to do with LLMsâ or âHave nothing to do with MLâ. Which is (a) telling; and (b) departs from her previous---straw-man---definition of âAIâ in a somewhat sneaky way.
The ugly
Putting the straw-man issue aside, and maybe the most important, I think the talk gives a very conservative, and very restrictive, view of what NLP is, what language is, what good research practices are, and NLP ought to be.
Yes, NLP is broader now. This is a good thing.
Letâs get back the the first bullets of slide 5:
- How do we build âthinking machinesâ that can do âhuman-likeâ reasoning?
- How do we build âthinking machinesâ that can âsurpassâ humans in cognitive work.
I already wrote I think this is a very reductive view of what âAI Researchâ is. But by contrasting it to NLP, as Emily does, it also a highlights a very reductive view of what language is, and of what language processing is.
Yes, I also observed a few months ago that the fields shifted from what I considered to be âclassicâ NLP work to notions like âreasoningâ, and my immediate reaction was âhey this is not about language anymoreâ. But after a bit of reflection, I realized that, hey, it is very much about language, because reasoning process is part of language. You cannot decouple (certain notions of) reasoning abilities from language use, and you cannot even perform certain kinds of reasoning without natural language (just look at how much trouble semanticists are taking to try and formalize language with logic, without coming with a complete, let alone elegant, solution). People reason in language. Communication involves reasoning. Forming, perceiving and responding to âcommunicative intentsâ, a term Emily likes to use, involves a hefty bit of reasoning. Sure, not all reasoning is linguistic, but to contrast reasoning with NLP is just bananas.
Similarly for âthinkingâ, yes, language and thinking are intimately correlated. Especially if we consider that we want to emulate âhuman-likeâ behavior. This is all hard-core natural language stuff, it does not by any means contrast with it.
And finally, of course we want to âsurpassâ humans (in cognitive work or otherwise), this is what computers are for. They are faster and more accurate in certain things, and this is good!
Emily also lists (same slide 5) âcure cancer, solve the climate crisis, make end-of-life decisions, etcâ and âHow do we automate the scientific method?â as (problematic) aspects of AI. First, these are all good things that we should be doing. Second, some people in NLP work on these and similar problems, and this is a good thing!! Yes, we want to cure cancer, we want to solve climate crisis, we want science to be more automated, faster, more efficient, and yes, NLP can be of great help with that.
Slide 8 lists a bunch of things NLP research is asking. And true, it does. But it is not ALL that it does, and I am truly surprised a linguist would miss that and dismiss it as â(the bad part of) AIâ.
These are not, in fact, tropes of bad NLP research.
Finally, slide 10 lists âBad research practicesâ that are âdue to AIâ.
I already covered the âdue to AIâ part in the straw-man section above. Letâs now focus on the content.
- Demands to evaluate against âSOTAâ closed models
- Unmanageably large datasets (â lack of held out data)
- Exploitative research & development practices
The last one is not a research practice at all, so I am not sure what it does in this list rather than for bashing AI. But letâs focus on the first two: they are research practices, but I strongly disagree that they are bad ones. In fact, I believe that not doing them is in many cases a bad research practice.
Eval against closed models: yes, it is annoying that we do not know how the models work. Yes, this limits the conclusions we can reach by studying them. But this does not mean we should pretend that they donât exist, and in many (though not all!) cases one really should compare to closed models if they want to convince that their point holds. Not everything is observable, this is a fact of nature. Science deals with that with finding more robust ways of forming valid conclusions despite the non-observability, not by ignoring reality. The bad practice is not to demand evaluating against closed models, this is actually a good practice. The bad practice is to demand an evaluation against closed models in situations in which this is irrelevant to the argument the authors are trying to make.
Unmanageably large datasets: it is very clear to anyone who has been paying even slight attention that scale helps. A lot. For better or worse, we need these large datasets. And already tools appear to aid in querying and inspecting them. Rather than labeling the large corpora as âbad practiceâ, âdangerâ, âtoo largeâ, we should promote work on improving the inspection tools, and promote looking into ways that allow to make convincing claims and form valid conclusions despite not having access to the training data. Big-data seems to be a major part of LLMs. The good practice is to study it, not to ignore its existence.
Fin
Let me summarize my main points:
- ACL is not an AI conference.
- ACL is not a CL conference.
- ACL is an NLP conference.
- Academic AI is not TESCERAL nor doomerism.
- NLP intersects with AI, but is neither subsumed by nor subsumes AI.
- NLP conferences should focus on language processing.
- Language processing is expanding its boundaries, and this is a good thing.
- Closing your eyes and ignoring reality is not good science.
- Embracing reality and finding way around limitations is good science.
- Activism leads to bad science.