Latest from MIT : Researchers discover a shortcoming that makes LLMs less reliable
Large language models (LLMs) sometimes learn the wrong lessons, according to an MIT study. Rather than answering a query based on domain knowledge, an LLM could respond by leveraging grammatical patterns it learned during training. This can cause a model to fail unexpectedly when deployed on new tasks. The researchers found that models can mistakenly…
