You seem to not make the difference between maths and, say, literature or history.
Do you actually think that an LLM can take, say, a Harry Potter book as an input, and give it a grade in such a way that everybody will always agree on?
And to go further, do you actually use LLMs to generate graphs and statistics from spreadsheet? Because that is probably a bad idea given that there are tools that actually do it right.
> Do you actually think that an LLM can take, say, a Harry Potter book as an input, and give it a grade in such a way that everybody will always agree on?
No, but I also don't think a human can do that either. Subjective things are subjective. I'm not sure I understand how this connects to the idea you expressed that doing various tasks with automation tools like LLMs prevent you from "internalizing" the data, or why not "internalizing" data is necessarily a bad thing. Am I just misunderstanding your concern?
Many of the posts I find here defending the use of LLMs focus on "profitability". "You ask me to give you 3 pages about X? I'll give you 3 pages about X and you may not even realise that I did not write them". I completely agree that it can happen and that LLMs, right now, are useful to hack the system. But if you specialise in being efficient at getting an LLM to generate 3 pages, you may become useless faster than you think. Still, I don't think that this is the point of the article, and it is most definitely not my point.
My point is that while you specialise in hacking the system with an LLM, you don't learn about the material that goes into those 3 pages.
* If you are a student, it means that you are losing your time. Your role as a student is to learn, not to hack.
* More generally as a person, "I am a professional in summarising stuff I don't understand in a way that convinces me and other people who don't understand it either" is not exactly very sexy to me.
If you want to get actual knowledge about something, you have to actually work on getting that knowledge. Moving it from an LLM to a word document is not it. Being knowledgeable requires "internalising" it. Such that you can talk about it at dinner. And have an opinion about it that is worth something to others. If your opinion is "ChatGPT says this, but with my expertise in prompting I can get it to say that", it's pretty much worthless IMHO. Except for tricking the system, in a way similar to "oh my salary depends on the number of bugs I fix? Let me introduce tons of easy-to-fix bugs then".
Do you actually think that an LLM can take, say, a Harry Potter book as an input, and give it a grade in such a way that everybody will always agree on?
And to go further, do you actually use LLMs to generate graphs and statistics from spreadsheet? Because that is probably a bad idea given that there are tools that actually do it right.