Hacker News new | past | comments | ask | show | jobs | submit | more qt31415926's comments login

I don't think the statement reads that the 44% and the 26% should be additive. Especially given the zombie graphic where it looks like they overlap the 26 on top of the 44, where the orange bar is the 26% and the remaining yellow bar is the 44%


Thanks, didn't know. This article seems to have been written by a reader though so not the same authors of the sketchy lithium work


I'm impressed you were able to not get tangled up by the youtube/captcha/color hex/roman numeral mess! The youtube one is what screwed me over and over out of all my attempts


"00:00 / <minutes>:<seconds>" youtube

Google that, including the quotes, replace the minutes and seconds with your given time, then look through the results and find a URL which fits the password criteria.

Every result should be a video with that duration. If it isn't, check your useragent. I noticed some weirdness with that.


Hmm, I tried googling "00:00 / 17:02" youtube and none of the results were the right duration on both Chrome and Firefox. Any ideas?


Not sure. It's working for me


The Chrome Inspector Color Picker does not return the right hex code! You have to go into the properties to get the right value.


I don't think it's easy. Verification is much easier than generating correct solutions for this.

Looking at the JS, these rules use RNG such that you can have an inconsistent or impossible password. E.g. if the only youtube video URLs that work with your duration have roman numerals that multiply above 35 in it you are hard stuck. Your youtube URL can also hard stuck your atomic number summation to 200 if it happens to contain enough elements that adds above 200. Your color hex can hard stuck your 25 sum, etc. The code does not try to generate working passwords given all the rules, it simply adds checks and randomly generates the requirement per rule.

You'd have to have the RNG rules to align well in order to win i.e. youtube video with no roman numerals or numbers or elements, captcha with no numbers or roman numerals or elements, to minimize conflict.


For a given video length, there will be some youtube video urls without roman numerals, and with low digit sum, and atomic number.

I first search this on Google

"0:00 / mm:ss" site:youtube.com

Where mm:ss is the desired length. Then I used some Javascript to scrape the results, finding only youtube urls without roman numbers, and print them out sorted by digit sum and atomic number

I've done it a few times, never had a situation where there was no suitable url

As for the color hex, if it's not suitable, you can regenerate it


I wonder if something like quickcheck could be used to randomly generate characters which pass the criteria. I don't know how it would handle Paul though...


Well, you could always upload your own YouTube video...

Make the generator use Pupetter and your YT credentials to upload an empty mp4 video of the given length.


I uploaded a video quick just to pass that step since the length was hard to find. The generated video ID was a bunch of Roman numerals so I was totally fucked.


Re-upload until you get a good ID


Her field has also taken the largest hit from the success of LLMs and her research topics and her department are probably no longer prioritized by research grants. Given how many articles she's written that have criticized LLMs it's not surprising she has incentives.


LLMs are in her field; they are one of her research topics and they're definitely getting funding.

We absolutely should not be ignoring research that doesn't support popular narratives; dismissing her work because it is critical of LLMs is not reasonable.


It is not that she is critical of LLMs that is the issue.

Instead, it is that she has strong ideological motivations to make certain arguments.

Those motivations being that her research is now worthless, because of LLMs.

I don't believe the alignment doomsayers either, but that is for different reasons than listening to her.


In her field doesn't mean that's what she researches, LLMs are loosely in her field but the methods are completely different. Computational linguistics != deep learning. Deep learning does not directly use concepts from linguistics, semantics, grammars or grammar engineering, which is what Emily was researcing for the past decades.

It's the same thing as saying a number theorist and a set theorist are in the same field cause they both work in the Math field.


They are what she researches though. She has published research on them.

LLMs don't directly use concepts from linguistics but they do produce and model language/grammar; it's entirely valid to use techniques from those fields to evaluate them, which is what she does. In the same vein, though a self-driving car doesn't work the same way as a human driver does, we can measure their performance on similar tasks.


Hmm I looked into it, and looked at papers/pdfs in google scholar's advanced search with her as an author that mentioned LLMs or GPT in the past 3 years. Every single one was a criticism about how they couldn't actually understand anything (e.g. "they're only trained on form" and "at best they can only understand things in a limited well scoped fashion") and that linguistic fundamentals for NLP was more important.

Good to know my hunch was correct


Current topic aside, I feel like that stochastic parrots paper aged really poorly in its criticisms of LLMs, and reading it felt like political propaganda with its exaggerated rhetoric and its anemic amount of scientific substance e.g.

> Text generated by an LM is not grounded in communicative intent, any model of the world, or any model of the reader’s state of mind. It can’t have been, because the training data never included sharing thoughts with a listener, nor does the machine have the ability to do that.

I'm surprised its cited so much given how many of its claims fell flat 1.5 years later


It's extremely easy to publish in NLP right now. 20-30% acceptance rates at even the top end conferences and plenty of tricks to increase your chances. Just because someone is first author on a highly cited paper doesn't imply that they're "right"


It's trained on pre-2021 data. Looks like they tested on the most recent tests (i.e. 2022-2023) or practice exams. But yeah standardized tests are heavily weighed towards pattern matching, which is what GPT-4 is good at, as shown by its failure at the hindsight neglect inverse-scaling problem.


I believe they showed that in GPT4 reversed the trend on the hindsight neglect problem. Search for "hindsight neglect" in the website and you can see that it's accuracy on the problem shot up to 100%.


oh my bad, totally misread that


Curious since it does well on the LSAT, SAT, GRE Verbal.


For Vietnamese, ChatGPT has been much more effective for me. You can tell it very specific things to modify as well as give it additional context which you can't really do with Google Translate. Especially with pronouns, google will just translate everyone as 'uncle' or 'aunt' or 'grandma' or 'grandpa' and it'll get genders wrong all the time, which you can correct for with ChatGPT.


makes sense, so they just looped that segment. thanks!


Consider applying for YC's Summer 2025 batch! Applications are open till May 13

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: