"Long-termism," the idea that the fate of humanity should be our top moral priority, has gained prominence recently in the writings of philosopher William MacAskill and UN Secretary-General António Guterres. But we need deep reflection over the tension and trade offs between long-term and nearer-term goals, especially when it comes to emerging technology like AI, write Anja Kaspersen and Wendell Wallach.