Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

The University of Waikato, New Zealand has had a lot of research going on to use compression for named entity tagging (name, location, date, person, ...) etc.

While it's not the best-performing paradigm for text sequence tagging, it is intellectually intriguing as you say because of the parallel between the concepts "compression" and "understanding", even in the human brain. If we can't understand s.th., we need to memorize it; if we understand it, it doesn't need much space or cognitive load at all, basically a name that is well-linked to other concepts.



Yeah it’s interesting this got me thinking, lossless compression is just removing redundancy right - like it doesn’t introduce any ambiguity in the data (?)

So feeding AI compressed data might allow it to be more efficient with its limited resources … I had never considered that, it’s very interesting idea


You are correct, compression is essentially extracting latent features in the data and discarding the rest.

Auto encoder networks, or networks that have an auto encoder like structure (U-net) employ essentially compression internally in the model to extract latent features.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: