3 Sources
[1]
Future AI models could be turbocharged by brand new system of logic that researchers call 'inferentialism'
The rigid structures of language we once clung to with certainty are cracking. The rigid structures of language we once clung to with certainty are cracking. Take gender, nationality or religion: these concepts no longer sit comfortably in the stiff linguistic boxes of the last century. Simultaneously, the rise of AI presses upon us the need to understand how words relate to meaning and reasoning. A global group of philosophers, mathematicians and computer scientists have come up with a new understanding of logic that addresses these concerns, dubbed "inferentialism". One standard intuition of logic, dating back at least to Aristotle , is that a logical consequence ought to hold by virtue of the content of the propositions involved, not simply by virtue of being "true" or "false". Recently, the Swedish logician Dag Prawitz observed that, perhaps surprisingly, the traditional treatment of logic entirely fails to capture this intuition. The modern discipline of logic, the sturdy backbone of science, engineering, and technology, has a fundamental problem. For the last two millennia, the philosophical and mathematical foundation of logic has been the view that meaning derives from what words refer to. It assumes the existence of abstract categories of objects floating around the universe, such as the concept of "fox" or "female" and defines the notion "truth" in terms of facts about these categories. For example, consider the statement, "Tammy is a vixen". What does it mean? The traditional answer is that there exists a category of creatures called "vixens" and the name "Tammy" refers to one of them. The proposition is true just in the case that "Tammy" really is in the category of "vixen". If she isn't a vixen, but identifies as one, the statement would be false according to standard logic. Logical consequence is therefore obtained purely by facts of truth and not by process of reasoning. Consequently, it can't tell the difference between, say, the equations 4=4 and 4=((2 x 52 ) -10)/10 simply because they are both true, but most of us would notice a difference. If our theory of logic can't handle this, what hope do we have to teach more refined, more subtle thinking to AI? What hope do we have of figuring out what is right and what is wrong in the age of post-truth? Sign up for the Live Science daily newsletter now Get the world's most fascinating discoveries delivered straight to your inbox. Contact me with news and offers from other Future brandsReceive email from us on behalf of our trusted partners or sponsorsBy submitting your information you agree to the Terms & Conditions and Privacy Policy and are aged 16 or over. Related: Communicating with aliens one day could be much easier if we study the way AI agents speak with each other Language and meaning Our new logic better represents modern speech. The roots of it can be traced to the radical philosophy of the eccentric Austrian philosopher Ludwig Wittgenstein, who in his 1953 book, Philosophical Investigations, wrote the following: "For a large class of cases of the employment of the word 'meaning' -- though not for all -- this word can be explained in this way: the meaning of a word is its use in the language." This notion makes meaning more about context and function. In the 1990s, the US philosopher Robert Brandom refined "use" to mean "inferential behaviour", laying the groundwork for inferentialism. Suppose a friend, or a curious child, were to ask us what it means to say "Tammy is a vixen". How would you answer them? Probably not by talking about categories of objects. We would more probably say it means, "Tammy is a female fox". More precisely, we would explain that from Tammy being vixen we may infer that she is female and that she is a fox. Conversely, if we knew both those facts about her, then we may indeed assert that she is a vixen. This is the inferentialist account of meaning; rather than assuming abstract categories of objects floating around the universe, we recognize that understanding is given by a rich web of relationship between elements of our language. Consider controversial topics today, such as those around gender. We bypass those metaphysical questions blocking constructive discourse, such as about whether the categories of "male" or "female" are real in some sense. Such questions don't make sense in the new logic because many people don't believe "female" is necessarily one category with one true meaning. As an inferentialist, given a proposition such as "Tammy is female", one would only ask what one may infer from the statement: one person might draw conclusions about Tammy's biological characteristics, another about her psychological makeup, while yet another might consider a completely different facet of her identity. Inferentialism made concrete So, inferentialism is an intriguing framework, but what does it mean to put it in practice? In a lecture in Stockholm in the 1980s, the German logician Peter Schroeder-Heister baptized a field, based on inferentialism, called "proof-theoretic semantics". In short, proof-theoretic semantics is inferentialism made concrete. This has seen substantial development in the last few years. While the results remain technical, they are revolutionizing our understanding of logic and comprise a major advancement in our understanding of human and machine reasoning and discourse. Large language models (LLMs), for example, work by guessing the next word in a sentence. Their guesses are informed only by the usual patterns of speech and by a long training programme comprising trial and error with rewards. Consequently, they "hallucinate", meaning that they construct sentences that are formed by logical nonsense. By leveraging inferentialism, we may be able to give them some understanding of the words they are using. For example, an LLM may hallucinate the historical fact: "The Treaty of Versailles was signed in 1945 between Germany and France after the second world war" because it sounds reasonable. But armed with inferential understanding, it could realize that "Treaty of Versaille" was after the first world war and 1918, not the second world war and 1945. This could also come in handy when it comes to critical thinking and politics. By having a fit for purpose understanding of logical consequence, we may be able to automatically flag and catalog nonsense arguments in newspapers and debates. For example, a politician may declare: "My opponent's plan is terrible because they have a history of making bad decisions." RELATED STORIES -- Will language face a dystopian future? How 'Future of Language' author Philip Seargeant thinks AI will shape our communication -- What's the difference between deductive reasoning and inductive reasoning? -- Can we think without using language? A system equipped with a proper understanding of logical consequence would be able to flag that while it may be true that the opponent has a history of poor decisions, no actual justification has been given for what is wrong with their current plan. By removing "true" and "false" from their pedestals we open the way for discernment in dialogue. It is based on these developments that we can claim that an argument -- whether in the heated arena of political debate, during a spirited disagreement with friends, or within the world of scientific discourse -- is logically valid. This edited article is republished from The Conversation under a Creative Commons license. Read the original article.
[2]
A new system of logic could boost critical thinking and AI
The rigid structures of language we once clung to with certainty are cracking. Take gender, nationality or religion: these concepts no longer sit comfortably in the stiff linguistic boxes of the last century. Simultaneously, the rise of AI presses upon us the need to understand how words relate to meaning and reasoning. A global group of philosophers, mathematicians and computer scientists have come up with a new understanding of logic that addresses these concerns, dubbed "inferentialism". One standard intuition of logic, dating back at least to Aristotle , is that a logical consequence ought to hold by virtue of the content of the propositions involved, not simply by virtue of being "true" or "false". Recently, the Swedish logician Dag Prawitz observed that, perhaps surprisingly, the traditional treatment of logic entirely fails to capture this intuition. The modern discipline of logic, the sturdy backbone of science, engineering, and technology, has a fundamental problem. For the last two millennia, the philosophical and mathematical foundation of logic has been the view that meaning derives from what words refer to. It assumes the existence of abstract categories of objects floating around the universe, such as the concept of "fox" or "female" and defines the notion "truth" in terms of facts about these categories. For example, consider the statement, "Tammy is a vixen". What does it mean? The traditional answer is that there exists a category of creatures called "vixens" and the name "Tammy" refers to one of them. The proposition is true just in the case that "Tammy" really is in the category of "vixen". If she isn't a vixen, but identifies as one, the statement would be false according to standard logic. Logical consequence is therefore obtained purely by facts of truth and not by process of reasoning. Consequently, it can't tell the difference between, say, the equations 4=4 and 4=((2 x 5 ) -10)/100 simply because they are both true, but most of us would notice a difference. If our theory of logic can't handle this, what hope do we have to teach more refined, more subtle thinking to AI? What hope do we have of figuring our what is right and what is wrong in the age of post-truth? Language and meaning Our new logic better represents modern speech. The roots of it can be traced to the radical philosophy of the eccentric Austrian philosopher Ludwig Wittgenstein, who in his 1953 book, Philosophical Investigations, wrote the following: "For a large class of cases of the employment of the word 'meaning' -though not for all -- this word can be explained in this way: the meaning of a word is its use in the language." This notion makes meaning more about context and function. In the 1990s, the US philosopher Robert Brandom refined "use" to mean "inferential behavior", laying the groundwork for inferentialism. Suppose a friend, or a curious child, were to ask us what it means to say "Tammy is a vixen". How we would you answer them? Probably not by talking about categories of objects. We would more probably say it means, "Tammy is a female fox". More precisely, we would explain that from Tammy being vixen we may infer that she is female and that she is a fox. Conversely, if we knew both those facts about her, then we may indeed assert that she is a vixen. This is the inferentialist account of meaning; rather than assuming abstract categories of objects floating around the universe, we recognize that understanding is given by a rich web of relationship between elements of our language. Consider controversial topics today, such as those around gender. We bypass those metaphysical questions blocking constructive discourse, such as about whether the categories of "male" or "female" are real in some sense. Such questions don't make sense in the new logic because many people don't believe "female" is necessarily one category with one true meaning. As an inferentialist, given a proposition such as "Tammy is female", one would only ask what one may infer from the statement: one person might draw conclusions about Tammy's biological characteristics, another about her psychological makeup, while yet another might consider a completely different facet of her identity. Inferentialism made concrete So, inferentialism is an intriguing framework, but what does it mean to put it in practice? In a lecture in Stockholm in the 1980s, the German logician Peter Schroeder-Heister baptized a field, based on inferentialism, called "proof-theoretic semantics". In short, proof-theoretic semantics is inferentialism made concrete. This has seen substantial development in the last few years. While the results remain technical, they are revolutionizing our understanding of logic and comprise a major advancement in our understanding of human and machine reasoning and discourse. Large language models (LLMs), for example, work by guessing the next word in a sentence. Their guesses are informed only by the usual patterns of speech and by a long training program comprising trial and error with rewards. Consequently, they "hallucinate", meaning that they construct sentences that are formed by logical nonsense. By leveraging inferentialism, we may be able to give them some understanding of the words they are using. For example, an LLM may hallucinate the historical fact: "The Treaty of Versailles was signed in 1945 between Germany and France after the second world war" because it sounds reasonable. But armed with inferential understanding, it could realize that "Treaty of Versaille" was after the first world war and 1918, not the second world war and 1945. This could also come in handy when it comes to critical thinking and politics. By having a fit for purpose understanding of logical consequence, we may be able to automatically flag and catalogue nonsense arguments in newspapers and debates. For example, a politician may declare: "My opponent's plan is terrible because they have a history of making bad decisions." A system equipped with a proper understanding of logical consequence would be able to flag that while it may be true that the opponent has a history of poor decisions, no actually justification has been given for what is wrong with their current plan. By removing "true" and "false" from their pedestals we open the way for discernment in dialogue. It is based on these developments that we can claim that an argument -- whether in the heated arena of political debate, during a spirited disagreement with friends, or within the world of scientific discourse -- is logically valid.
[3]
Researchers have invented a new system of logic that could boost critical thinking and AI
University College London provides funding as a founding partner of The Conversation UK. The rigid structures of language we once clung to with certainty are cracking. Take gender, nationality or religion: these concepts no longer sit comfortably in the stiff linguistic boxes of the last century. Simultaneously, the rise of AI presses upon us the need to understand how words relate to meaning and reasoning. A global group of philosophers, mathematicians and computer scientists have come up with a new understanding of logic that addresses these concerns, dubbed "inferentialism". One standard intuition of logic, dating back at least to Aristotle , is that a logical consequence ought to hold by virtue of the content of the propositions involved, not simply by virtue of being "true" or "false". Recently, the Swedish logician Dag Prawitz observed that, perhaps surprisingly, the traditional treatment of logic entirely fails to capture this intuition. The modern discipline of logic, the sturdy backbone of science, engineering, and technology, has a fundamental problem. For the last two millennia, the philosophical and mathematical foundation of logic has been the view that meaning derives from what words refer to. It assumes the existence of abstract categories of objects floating around the universe, such as the concept of "fox" or "female" and defines the notion "truth" in terms of facts about these categories. For example, consider the statement, "Tammy is a vixen". What does it mean? The traditional answer is that there exists a category of creatures called "vixens" and the name "Tammy" refers to one of them. The proposition is true just in the case that "Tammy" really is in the category of "vixen". If she isn't a vixen, but identifies as one, the statement would be false according to standard logic. Logical consequence is therefore obtained purely by facts of truth and not by process of reasoning. Consequently, it can't tell the difference between, say, the equations 4=4 and 4=((2 x 52 ) -10)/100 simply because they are both true, but most of us would notice a difference. If our theory of logic can't handle this, what hope do we have to teach more refined, more subtle thinking to AI? What hope do we have of figuring our what is right and what is wrong in the age of post-truth? Language and meaning Our new logic better represents modern speech. The roots of it can be traced to the radical philosophy of the eccentric Austrian philosopher Ludwig Wittgenstein, who in his 1953 book, Philosophical Investigations, wrote the following: "For a large class of cases of the employment of the word 'meaning' -though not for all - this word can be explained in this way: the meaning of a word is its use in the language." This notion makes meaning more about context and function. In the 1990s, the US philosopher Robert Brandom refined "use" to mean "inferential behaviour", laying the groundwork for inferentialism. Suppose a friend, or a curious child, were to ask us what it means to say "Tammy is a vixen". How we would you answer them? Probably not by talking about categories of objects. We would more probably say it means, "Tammy is a female fox". More precisely, we would explain that from Tammy being vixen we may infer that she is female and that she is a fox. Conversely, if we knew both those facts about her, then we may indeed assert that she is a vixen. This is the inferentialist account of meaning; rather than assuming abstract categories of objects floating around the universe, we recognise that understanding is given by a rich web of relationship between elements of our language. Consider controversial topics today, such as those around gender. We bypass those metaphysical questions blocking constructive discourse, such as about whether the categories of "male" or "female" are real in some sense. Such questions don't make sense in the new logic because many people don't believe "female" is necessarily one category with one true meaning. As an inferentialist, given a proposition such as "Tammy is female", one would only ask what one may infer from the statement: one person might draw conclusions about Tammy's biological characteristics, another about her psychological makeup, while yet another might consider a completely different facet of her identity. Inferentialism made concrete So, inferentialism is an intriguing framework, but what does it mean to put it in practice? In a lecture in Stockholm in the 1980s, the German logician Peter Schroeder-Heister baptised a field, based on inferentialism, called "proof-theoretic semantics". In short, proof-theoretic semantics is inferentialism made concrete. This has seen substantial development in the last few years. While the results remain technical, they are revolutionising our understanding of logic and comprise a major advancement in our understanding of human and machine reasoning and discourse. Large language models (LLMs), for example, work by guessing the next word in a sentence. Their guesses are informed only by the usual patterns of speech and by a long training programme comprising trial and error with rewards. Consequently, they "hallucinate", meaning that they construct sentences that are formed by logical nonsense. By leveraging inferentialism, we may be able to give them some understanding of the words they are using. For example, an LLM may hallucinate the historical fact: "The Treaty of Versailles was signed in 1945 between Germany and France after the second world war" because it sounds reasonable. But armed with inferential understanding, it could realise that "Treaty of Versaille" was after the first world war and 1918, not the second world war and 1945. This could also come in handy when it comes to critical thinking and politics. By having a fit for purpose understanding of logical consequence, we may be able to automatically flag and catalogue nonsense arguments in newspapers and debates. For example, a politician may declare: "My opponent's plan is terrible because they have a history of making bad decisions." A system equipped with a proper understanding of logical consequence would be able to flag that while it may be true that the opponent has a history of poor decisions, no actually justification has been given for what is wrong with their current plan. By removing "true" and "false" from their pedestals we open the way for discernment in dialogue. It is based on these developments that we can claim that an argument - whether in the heated arena of political debate, during a spirited disagreement with friends, or within the world of scientific discourse - is logically valid.
Share
Copy Link
A global team of researchers has developed a new system of logic called 'inferentialism' that could enhance AI capabilities and improve critical thinking by focusing on the use and context of language rather than abstract categories.
Researchers from various disciplines have introduced a novel system of logic called 'inferentialism,' which could potentially revolutionize our understanding of language, reasoning, and artificial intelligence. This new approach addresses the limitations of traditional logic systems that have been in place for millennia 123.
For over two thousand years, the foundation of logic has been based on the idea that meaning is derived from what words refer to, assuming the existence of abstract categories. However, this approach has fundamental issues:
Inferentialism offers a new perspective on logic and meaning:
The new logic system has potential applications in various fields:
Peter Schroeder-Heister introduced "proof-theoretic semantics" as a concrete application of inferentialism:
The implementation of inferentialism could lead to significant improvements in AI and human reasoning:
As researchers continue to develop and refine this new system of logic, its potential applications in AI, education, and critical discourse could reshape how we approach complex problems and communicate in an increasingly nuanced world.
Elon Musk's AI company xAI has open-sourced the Grok 2.5 model on Hugging Face, making it available for developers to access and explore. Musk also announced plans to open-source Grok 3 in about six months, signaling a commitment to transparency and innovation in AI development.
7 Sources
Technology
1 day ago
7 Sources
Technology
1 day ago
Nvidia announces plans to implement silicon photonics and co-packaged optics for AI GPU communication by 2026, promising higher transfer rates and lower power consumption in next-gen AI data centers.
2 Sources
Technology
8 hrs ago
2 Sources
Technology
8 hrs ago
Netflix has released new guidelines for using generative AI in content production, outlining low-risk and high-risk scenarios and emphasizing responsible use while addressing industry concerns.
2 Sources
Technology
8 hrs ago
2 Sources
Technology
8 hrs ago
Scientists at KIST have developed a new device principle that utilizes "spin loss" as a power source for magnetic control, potentially revolutionizing the field of spintronics and paving the way for ultra-low-power AI chips.
2 Sources
Technology
8 hrs ago
2 Sources
Technology
8 hrs ago
Cloudflare introduces new features for its Cloudflare One zero-trust platform, aimed at helping organizations securely adopt, build, and deploy generative AI applications while maintaining security and privacy standards.
2 Sources
Technology
8 hrs ago
2 Sources
Technology
8 hrs ago