Latest from MIT Tech Review – Why Meta’s latest large language model survived only three days online

On November 15 Meta unveiled a new large language model called Galactica, designed to assist scientists. But instead of landing with the big bang Meta hoped for, Galactica has died with a whimper after three days of intense criticism. Yesterday the company took down the public demo that it had encouraged everyone to try out….

Latest from Google AI – The Data Cards Playbook: A Toolkit for Transparency in Dataset Documentation

Posted by Mahima Pushkarna, Senior Interaction Designer, and Andrew Zaldivar, Senior Developer Relations Engineer, Google Research As machine learning (ML) research moves toward large-scale models capable of numerous downstream tasks, a shared understanding of a dataset’s origin, development, intent, and evolution becomes increasingly important for the responsible and informed development of ML models. However, knowledge…

Latest from Google AI – Mixture-of-Experts with Expert Choice Routing

Posted by Yanqi Zhou, Research Scientist, Google Research Brain Team The capacity of a neural network to absorb information is limited by the number of its parameters, and as a consequence, finding more effective ways to increase model parameters has become a trend in deep learning research. Mixture-of-experts (MoE), a type of conditional computation where…

Latest from MIT Tech Review – Feeding the world by AI, machine learning and the cloud

Although the world population has continued to steadily increase, farming practices have largely remained the same. Amid this growth, climate change poses great challenges to the agricultural industry and its capacity to feed the world sustainably. According to the World Bank, 70% of the world’s fresh water is used in agriculture and droughts and heat…

Latest from MIT : Solving brain dynamics gives rise to flexible machine-learning models

Last year, MIT researchers announced that they had built “liquid” neural networks, inspired by the brains of small species: a class of flexible, robust machine learning models that learn on the job and can adapt to changing conditions, for real-world safety-critical tasks, like driving and flying. The flexibility of these “liquid” neural nets meant boosting…

Latest from MIT Tech Review – Why we need to do a better job of measuring AI’s carbon footprint

This story originally appeared in The Algorithm, our weekly newsletter on AI. To get stories like this in your inbox first, sign up here. Lately I’ve lost a lot of sleep over climate change. It’s just over five weeks until Christmas, and last weekend in London, it was warm enough to have a pint outside without…

Latest from MIT Tech Review – We’re getting a better idea of AI’s true carbon footprint

Large language models (LLMs) have a dirty secret: they require vast amounts of energy to train and run. What’s more, it’s still a bit of a mystery exactly how big these models’ carbon footprints really are. AI startup Hugging Face believes it’s come up with a new, better way to calculate that more precisely, by…

Latest from MIT Tech Review – Best practices for bolstering machine learning security

Nearly 75% of the world’s largest companies have already integrated AI and machine learning (ML) into their business strategies. As more and more companies — and their customers — gain increasing value from ML applications, organizations should be considering new security best practices to keep pace with the evolving technology landscape.  Companies that utilize dynamic…

Latest from Google AI – Robots That Write Their Own Code

Posted by Jacky Liang, Research Intern, and Andy Zeng, Research Scientist, Robotics at Google <!—-><!—-> A common approach used to control robots is to program them with code to detect objects, sequencing commands to move actuators, and feedback loops to specify how the robot should perform a task. While these programs can be expressive, re-programming…

Latest from Google AI – ReAct: Synergizing Reasoning and Acting in Language Models

Posted by Shunyu Yao, Student Researcher, and Yuan Cao, Research Scientist, Google Research, Brain Team <!—-> Recent advances have expanded the applicability of language models (LM) to downstream tasks. On one hand, existing language models that are properly prompted, via chain-of-thought, demonstrate emergent capabilities that carry out self-conditioned reasoning traces to derive answers from questions,…