Browse course

Tutorial 1. Logic and AI

Exercise sheet

Definitions  

Note: A definition needs to fully explain the meaning of a concept. This is typically done by saying under which precise conditions the concept applies. That is, the answer should have the form like “A inference is … and consists of…” or “An AI system is symbolic just in case …”.

a.) Inference

Give a definition of inference. Include examples (where your definition applies) and non-examples (where it does not). Make sure to explain why your definition applies or not.

b.) Symbolic AI

Give a definition of symbolic AI. Include examples and non-examples with explanations.

Solution

These are examples of good answers.

a.) An inference is a process in which one or more premises are taken to support a conclusion, such that the conclusion follows from the premises according to a rule of reasoning. More precisely, an inference occurs just in case a transition from premises to conclusion is made under a rule that is truth-preserving (in deductive cases) or likelihood-enhancing (in inductive cases), and the premises are treated as grounds or reasons for endorsing the conclusion.

In deductive cases, an inference consists of a finite set of well-formed formulas (the premises) and a conclusion such that the conclusion is derivable from the formulas under a given system of inference rules. Then, the specific is derived from something general. For instance: all humans are mortal, Socrates is a human, so Socrates is mortal. Hence, the conclusion necessarily follows from the premises under standard first-order logic (deductive logic)

An inductive inference is one that is probabilistic: from the formulas (premises), one infers the conclusion, not because the conclusion is necessarily true, but because the premises makes the conclusion more plausible. To illustrate: the last hundred swans observed in this region were white, thus the next swan observed in this region will also be white. Consequently, the conclusion is not guaranteed but supported by a pattern in empirical data.

A non-example is the following: The moon shines at night, so Socrates is a philosopher. This is not an inference because there is no logical or probabilistic connection between the premise and the conclusion. The provides no reason to accept the conclusion.

b.) An AI system is symbolic just in case it represents knowledge using explicit, human-interpretable symbols (e.g., words, logical predicates, rules), and manipulates these symbols via formal rules of inference or transformation. The operation of such a system is grounded in symbolic representation and rule-based reasoning, typically following paradigms from logic, production systems, or semantic networks.

A symbolic AI system consists of three elements: a knowledge base (an explicit set of facts and rules, often written in a logical language), an inference engine (a mechanism that applies inference rules to derive new knowledge or make decisions) and an explicit representation of concepts and relationships.

An example of an expert system that is a symbolic system is MYCIN. This system used rules, e.g., “If the infection is bacterial and the organism is gram-positive, then prescribe penicillin. Besides, it used explicit symbolic facts, e.g., “Patient has streptococcus.” The system reasons via if-then rules to derive conclusions about diagnosis or treatment.

A deep neural network, such as ChatGPT, is not symbolic AI. It processes input as high-dimensional numerical vectors and performs operations like matrix multiplication, without explicit symbolic rules or interpretable representations. This is not-symbolic, because the symbols are not readable for humans, no logical inference is engine and no explicit rule-based reasoning is used.

Examples of indicators  

Give at least 5 additional examples of:

a.)

premise indicators,

b.)

inference indicators.

Solution

a.) Since, because, for, whereas, as, given that, assuming that, considering that, due to.

b.) It follows that, thus, hence, consequently, we know that, therefore, so, accordingly, then, as a result, this implies that.

Recognizing indicators  

Name the premise and inference indicators in the following paragraphs.

a.)

Because IA   knows that location of every book in the library, and the student asks for a specific title, it follows that IA  can guide the user directly to the book’s shelf.

b.)

If IA   detects that a student struggles with mathematics, then it recommends extra practice problems. Since the student’s last test scores were low, so IA   suggest additional exercises.

c.)

IA   thinks that he recognizes the face of a student entering the classroom. When a face is recognized by IA  , then it marks attendance automatically in the specific class. IA   remembers the face of the student, but the student is actually in another class. Consequently, IA   the attendance is not recorded.

e.)

If IA   detects pronunciation errors in a student’s speech, it will provide corrective feedback. Given that the student is practicing regularly, therefore IA   improves their pronunciation over time.

Solution

a.) Premise indicators: because, and. Conclusion indicator: it follows that.

b.) Premise indicators: if, then, since. Conclusion indicator: so.

c.) Premise indicators: when, then. Conclusion indicator: consequently.

d.) Premise indicators: if, given that. Conclusion indicator: therefore.

Inferences  

Define the premise(s) and conclusion of each of the following inferences.

a.)

The robot IA   rides towards entering of the supermarket. The supermarket’s automatic sliding doors open, when someone stands in front of the sensor. The sensor detects IA  . Therefore, the supermarket’s automatic doors open.

b.)

Since the robot IA   nests itself in the phone, is it able to take over the function of Face ID. If IA   recognizes the face, it will unlock the phone. A phone’s user, ∃n, picks up the phone and tries to unlock it by showing its face to the phone. However, IA   doesn’t know ∃n, so the phone stays unlocked.

c.)

If IA   notices that the milk is running low, it directly orders new milk. The ordered milk will arrive tomorrow, because ∃n delivers milk the day after the order is made by IA  .

d.)

IA   (in the personality of an autonomous driving car), at a red traffic light either brakes or chooses an alternative route. Since IA   isn’t braking, it must currently be still driving. Thus, it should have chosen another route.

e.)

If IA   notices that energy consumption is low at night, it turns off devices, and if IA   detects peak consumption, it shifts usage to cheaper hours. But energy consumption is now either low or there is peak consumption. So, IA   either turns off devices or shifts usage to cheaper hours.

f.)

If IA   detects plagiarism in an assignment, the teacher is alerted. The teacher was not alerted. Consequently, AI did not detect plagiarism.

g.)

IA   is a useful additional teacher to students, because IA   carefully analyzes the submitted homework to spot areas where the student struggles. Since IA   can identify these specific weaknesses, it knows exactly what advice to give to help the student improve. Moreover, thanks to the help of IA  , the teacher can focus on more structural or analytical explanations.

Solution

a.) P1: The robot IA   rides towards entering of the supermarket. P2: The supermarket’s automatic sliding doors open, when someone stands in front of the sensor.P3: The sensor detects IA  . C: the supermarket’s automatic doors open.

b.) P1: Since the robot IA   nests itself in the phone, is it able to take over the function of Face ID. P2: If IA   recognizes the face, it will unlock the phone. P3: A phone’s user, ∃n, picks up the phone and tries to unlock it by showing its face to the phone. P4: IA   doesn’t know ∃n C: The phone stays unlocked.

c.) P1: If IA   notices that the milk is running low, it directly orders new milk. P2: ∃n delivers milk the day after the order is made by IA  . C: The ordered milk will arrive tomorrow.

d.) P1: IA   (in the personality of an autonomous driving car), at a red traffic light either brakes or chooses an alternative route. P2: Since IA   isn’t braking, it must currently be still driving. C: It should have chosen another route.

e.) P1: If IA   notices that energy consumption is low at night, it turns off devices. P2: if IA   detects peak consumption, it shifts usage to cheaper hours. P3: Energy consumption is now either low or there is peak consumption. C: IA   either turns off devices or shifts usage to cheaper hours.

f.) P1: If IA   detects plagiarism in an assignment, the teacher is alerted. P2: The teacher was not alerted. C: IA   did not detect plagiarism.

g.) P1: IA   carefully analyzes the submitted homework to spot areas where the student struggles. P2: IA   can identify these specific weaknesses, it knows exactly what advice to give to help the student improve. P3: thanks to the help of IA  , the teacher can focus on more structural or analytical explanations. C: IA   is a useful additional teacher to students.

Symbolic or subsymbolic  

Below are technical descriptions of different AI systems taken from Wikipedia. In each case, classify the system as either symbolic or sub-symbolic system based on the description. Justify your answer by referencing the relevant aspects.

a.)

MYCIN was an early backward chaining expert system that used artificial intelligence to identify bacteria causing severe infections, such as bacteremia and meningitis, and to recommend antibiotics, with the dosage adjusted for patient’s body weight. MYCIN operated using a fairly simple inference engine and a knowledge base of ~600 rules by obtaining individual inferential facts identified by experts and encoding such facts as individual production rules. No other AI program at the time contained as much domain-specific knowledge clearly separated from its inference procedures as MYCIN. It would query the physician running the program via a long series of simple yes/no or textual questions. At the end, it provided a list of possible culprit bacteria ranked from high to low based on the probability of each diagnosis, its confidence in each diagnosis’ probability, the reasoning behind each diagnosis (that is, MYCIN would also list the questions and rules which led it to rank a diagnosis a particular way), and its recommended course of drug treatment.

b.)

Vampire is an automatic theorem prover for first-order classical logic developed in the Department of Computer Science at the University of Manchester. Vampire’s kernel implements the calculi of ordered binary resolution and superposition (for handling equality). The splitting rule and negative equality splitting can be simulated by the introduction of new predicate definitions and dynamic folding of such definitions. A DPLL-style algorithm splitting is also supported. A number of standard redundancy criteria and simplification techniques are used for pruning the search space: tautology deletion, subsumption resolution, rewriting by ordered unit equalities, basicness restrictions and irreducibility of substitution terms. The reduction ordering on terms is the standard Knuth–Bendix ordering.

c.)

A self-organizing map (SOM) or self-organizing feature map (SOFM) is an unsupervised machine learning technique used to produce a low-dimensional (typically two-dimensional) representation of a higher-dimensional data set while preserving the topological structure of the data. For example, a data set with p variables measured in n observations could be represented as clusters of observations with similar values for the variables. These clusters then could be visualized as a two-dimensional “map” such that observations in proximal clusters have more similar values than observations in distal clusters. This can make high-dimensional data easier to visualize and analyze.

d.)

ELIZA is an early natural language processing computer program developed from 1964 to 1967 at MIT by Joseph Weizenbaum. Created to explore communication between humans and machines, ELIZA simulated conversation by using a pattern matching and substitution methodology that gave users an illusion of understanding on the part of the program, but had no representation that could be considered really understanding what was being said by either party. ELIZA itself examined the text for keywords, applied values to said keywords, and transformed the input into an output; the script that ELIZA ran determined the keywords, set the values of keywords, and set the rules of transformation for the output.

e.)

In natural language processing, latent Dirichlet allocation (LDA) is a Bayesian network (and, therefore, a generative statistical model) for modeling automatically extracted topics in textual corpora. The LDA is an example of a Bayesian topic model. In this, observations (e.g., words) are collected into documents, and each word’s presence is attributable to one of the document’s topics. Each document will contain a small number of topics. Documents are represented as random mixtures over latent topics, where each topic is characterized by a distribution over all the words.

f.)

In machine learning, support vector machines (SVMs, also support vector networks) are supervised max-margin models with associated learning algorithms that analyze data for classification and regression analysis. SVMs can efficiently perform non-linear classification using the kernel trick, representing the data only through a set of pairwise similarity comparisons between the original data points using a kernel function, which transforms them into coordinates in a higher-dimensional feature space. Thus, SVMs use the kernel trick to implicitly map their inputs into high-dimensional feature spaces, where linear classification can be performed.

Solution

a.) MYCIN is a symbolic system, because its knoweldge is encoded explicitly as about 600 rules, each of which has a claer antecedent (if-condition) and consequent (then-condition) structure: “if [symptom/criteria], then [diagnostic or treatment fact]. During a consultation, MYCIN’s inference engine performs backward chaining over those rules: it starts from a hypothesis and works backward, asking the user yes/no or textual questions to verify which rules’ conditions hold. At the end of a session, MYCIN prints out exactly which questions it asked, which rules fired, how it computed confidence scores, and why it ranked bacteria in a certain order. Because everything is represented as discrete symbols, and because the inference mechanism manipulates those symbols according to logic production rules, MYCIN is one of the typical symbolic AI systems.

b.) Vampire is a symblic system, since it is an automatic theorem prover for first-order classical logic, and it operates entirely by manipulating symbolic clauses. Its core kernel implements ordered binary resolution and superposition, both of which are pure logical inference methods over symbolic expressions. When Vampire searches for a proof, it treats every proposition as a structure of function symbols, predicate symbols, and variables. It applies resolution rules to those literals, tries to unify terms symbolically, and prunes search branches using well-defined redundancy criteria (e.g., subsumption, tautology deletion, rewrite rules). All of these operations occur on discrete symbols rather than on numeric vectors or statistical parameters. Because Vampire’s entire “knowledge” is the set of logical axioms and conjectures expressed symbolically, and because its inference mechanism is purely rule-based symbol manipulation, Vampire is a symbolic system.

c.) SOFM is an example of a subsymbolic system, because the computations the system makes are governed by continuous weight vectors and iterative adjustment via a neighborhood function. There is no explicit, human-readable rule like “if [some symbolic condition] then [assign to cluster X]”. Instead, each neuron in the map has a weight vector, and during training the network iteratively adjusts those weights so that nearby neurons in the two-dimensional lattice come to represent similar input patterns. Since the system’s knowledge is embedded in continuous numeric parameters and the learning algorithm is based on iterative numeric adjustments rather than symbolic rules, SOFMs are subsymbolic.

d.) ELIZA contains no numeric learning or statistical modeling, but relies entirely on a set of pattern-matching rules, which makes it a symbolic system. Each rule consists of patterns (literal strings) and associated response templates. For example, if the user’s input matches the pattern “I am feeling X”, ELIZA might respond by transforming “I am feeling X” into “Why are you feeling X?” or “Do you often feel X?” All transformations are simple text substitutions based on pattern matching. Although the output often gives an illusion of understanding, under the hood of ELIZA never computes any numeric score or probabilistic weight, it just looks for matching keywords and applies the corresponding transformation rules. Since all its behavior comes from explicit scripts, and because there is no hidden continuous state, ELIZA is purely symbolic.

e.) LDA is an example of a subsymbolic system which is a generative probabilistic model in which topics are latent (hidden) variables, and documents are assumed to be mixtures over those topics. The core of LDA is a graphical (Bayesian) network in which each document has a probability distribution over K topics, and each topic is itself a probability distribution over the vocabulary. Although topics are sometimes describing in words, those are numeric probability vectors. There are no explicit it-then-rules. Instead, topics are soft distributions over words. Since LDA’s behavior is entirely driven by numeric latent variables, and it lacks any explicit discrete rule set, it is subsymbolic.

f.) SVMs is a subsymbolic system, because it relies on continuous parameters learned via optimalization, and because knowledge is hidden in support-vector coefficients rather than expressed as logical rules. SVMs are supervised learning models that learn a decision boundary (hyperplane) by solving a convex quadratic optimalization problem. Each training instance is mapped into a feature space, and the algorithm finds the maximum-margin hyperplane that separates classes. Because SVM is a maximum-margin numeric classifier, it is a typical example of a symbolic system.

Thinking, fast and slow  

Give at least 3 new examples of system 1 and system 2 thinking and explain why.

a.)

System 1

b.)

System 2

Solution

a.) System 1 thinking, which is fast, automatic, intuitive, unconscious, associative. For example:

  • To localize the source of a specific sound is a typical example of System 1 thinking, since the process happens unconscious and automatic. The brains register auditory stimuli and, within a fraction of a second, converts them into a rough spatial estimate (e.g., “it sounds like it’s behind me”). This happens without any conscious analysis of frequencies or time differences. Actually, it is largely a reflective, evolutionarily conditioned reaction that helps one detect danger or other relevant cues quickly.
  • The understanding of simple sentences in a native language is an example of System 1 thinking if one masters the language fluently. Then, one recognizes the structures of sentences, word order and semantics so quickly that one doesn’t need to consciously think about them. A sentence like “The book is on the table” is understood almost immediately, without any explicit grammatical analysis. However, if the sentences is grammatically complex, ambiguous or new, one would use System 2 thinking to break it down, even though the sentence is in someone’s native language.
  • Reaching out to grab something that falls, is a reflexive motor pattern. As soon as one sees that something is about to fall, visual and proprioceptive signals send information to your hand-eye coordination System, causing one to reach out. There is no time or space for conscious thought, this happens unconsciously in a split second.
  • Forming the mental image associated with the word “book” is an example of System 1 thinking, because, when one hears or reads the word “book”, the brain instantly and automatically triggers a visual and semantic schema which creates the image of a cover, pages, printed or written text and ensembles this to an image of a book. This requires almost no conscious or rational interpretation, since it comes from automated, associative memory links. On the other hand, one might object that forming a mental image could be a form of System 2 thinking as well. For instance, if one reflects on which specific book is meant and how that book looks like, then one needs a deliberate step-by-step mental construction.
  • An AI System for facial recognition on a phone recognizes faces automatically without “thinking” about what a face looks like. This is a System 1 kind of thinking for the user of the phone. The phone unlocks almost instantly when one’s face is detected, without one thinking consciously about what exactly constitutes a “face”. The algorithm processes facial features rapidly via trained neural networks: the associations between pixel patterns and facial characteristics occur largely unconsciously.
  • Algorithms that recommended content based on viewing history, without explanation why this is recommended, is an example of System 1 thinking. For the user, it feels like an immediate suggestion that catches one’s eye without needing to think about the why or how. The algorithm itself relies on generated correlations and patterns from datasets, which are performed quickly. So the recommendation pops up right away.
  • Voice assistant like Siri or Alexa use System 1 thinking: the voice assistant instantly interprets and responds to a simple command like “What time is it?” When one speaks a phrase to the voice assistant, almost immediately the answer appears. Siri or Alexa doesn’t think about it. The process of speech recognition, intent interpretation and response generation runs in an optimized and automated sequence. For the user, there is no slow, conscious analysis. Instead, it feels intuitive and instantaneous.
  • AI model that detects spam in emails uses System 1 thinking: based on pattern recognition and prior data, it automatically classifies a message as spam. As soon as an email arrives, it seamlessly passes through a pertained classification model and ends up in the spam folder or inbox without setting explicit rules. Thinking about whether something is spam, is not necessary, the system recognizes it unconsciously.
  • Simple chatbots, such as ChatGPT-3.5 that react fast and do not reason much, which make them an example of System 1 thinking. A user asks a simple question and almost immediately receives a (hopefully) coherent answer. The chatbot relies mostly on automatic, subsymbolic processes to generate text, without reasoning about the content. Besides, the user doesn’t see the internal logic or reasoning steps. However, the boundary is blurry: when the chatbot has to perform multilayered reasoning, it begins to use System 2 thinking.

b.) System 2 thinking, which is slow, deliberate, conscious, logical, calculating. For example:

  • Solving a complex arithmetic calculation requires focused attention, step-by-step reasoning and memory. These are aspects that indicate System 2 thinking, because one needs to deliberately apply arithmetic rules and keep track of intermediate steps. However, for some people who are really capable of mathematics, this complex calculation can be System 1 thinking as well. They can perform complex arithmetic rapidly and seemingly effortlessly, often without needing to go through conscious, step-by-step logic each time. Indeed, they may use mental shortcuts or visual strategies that bypass traditional calculation steps or their working and long-term memory for numbers and patterns are highly trained.
  • Finding the solution for a multi-criteria optimalization problem is typically System 2 thinking. These problems involve evaluating trade-offs between competing variables, often using algorithms or logical reasoning. Optimalization requires listing out each criterion, assigning relative weights and then comparing all alternatives systematically. This approach is slow, deliberate and explicitly logical.
  • Trying to identify the source of an unfamiliar sound or analyze a musical chord structure are examples of System 2 thinking, because it involves methodical processes without memorizing. Breaking down a chord means consciously naming each pitch, recognizing interval relationships, and mapping it back onto music theory. Both tasks demand focused attention, conscious comparison to mental templates and stepwise inference. On the other hand, an experiences mechanic or audiophile often instantly knows that a ratcheting click at a particular frequency means a loose piston pin. Likewise, a trained musician or ear-trainer can identify a complex chord by ear almost reflexively. Years of exposure have bundled the steps into an automatic pattern match, now functioning like System 1.
  • Count the number of A’s in a certain text requires full concentration. One must consciously move letter by letter, keep an accurate count, perhaps use pen-and-paper, and maintain vigilance to avoid losing track.
  • Determine the price/quality ratio of two phones is an example of System 2 thinking, because it demands conscious, logical and calculating thinking. If one determines the ratio, one first getters the important information and relevant specification that determine the quality of the phone. Then one should normalize the specification, to make sure all variables are converted to a common scale. One makes use of its mental competences and step-by-step logic, while computing the price/quality ratio. The last step is the interpretation of those ratios which requires conscious consideration. This all is System 2 thinking, one actively reflects on what is or is not relevant to the context.
  • To determine the validity of a complex logical reasoning, one should consciously map premises, consider contrapositives, check for fallacies, possibly draw a truth table, and ensure that each inference follows logically. This conscious, logical proof is typically System 2 thinking.
  • AI model that can play chess, like AlphaZero , that analyzes all possible moves and learns every strategy, can be considered as an example of System 2 thinking. When a human grandmaster faces a novel position, he might engage in complex “tree search”, calculating forcing variation, evaluating material imbalances, and computing candidate lines many moves deep. This exhaustive mental process matches the description of System 2, as it is a slower, effortful and rule-based analysis. AI models like AlphaZero use a neural network that doesn’t literally “reason” move-by-move as humans do, but it uses a learned value network to evaluate positions quickly. The algorithm searches subtrees.
  • ChatGPT-4 is able to use reasoning for functional programming, which van be considered as System 2 thinking. If one prompts GPT-4 to write a Python function that finds all prime factors of a large integer, it stimulates internally, step-by-step, the algorithm: trial division, recursion, memorization of known primes. It assembles correct code tokens based on logical patterns. This chain-of-thought that GPT-4 deploys, especially when it is asked for explanation, resembles System 2: deliberate, token-by-token reasoning about how to structure, map, filter and recursion.
  • AI models like WolframAlpha are used to solve complex mathematical problems and analyze data, which is an example of System 2 thinking. WolframAlpha frequently shows a stepwise solution. Each intermediate, applying integration by parts, simplifying each term, substituting limits, is to System 2 reasoning. It explicitly follows mathematical rules and applies formulae.
Research  

Note: These questions require you to do your own research, using reliable academic sources. You need to reference your sources!

Decidability

Are all logical systems undecidable?

Symbolic AI

Give 2 examples of existing expert systems, one that “failed” and one that’s still in use today. Explain why the first one failed.

Solution

These are examples of possible answers.

a.) No, not all logical systems are undecidable. A logical system is said ro be decidable if there exists an effective algorithm that, for any arbitrary formula within the system, determines in finite time whether that formula is derivable (i.e., a theorem) in the system. However, some logical systems are decidable, especially those with carefully limited expressive power. For instance, propositional logic is decidable. It has a finite number of variables and truth assignments, and its satisfiability and validity can be determined algorithmically via truth tables. Moreover, monadic first-order logic, which is first-order logic that is restricted to unary predicates and without function symbols, is also decidable.

In contrast, “Full First-Order Logic” is undecidable, as shown by Church and Turing (1935 & 1936). There is no general algorithm to decide the validity of arbitrary first-order formulas.

Undecidability is a feature of certain (especially expressive) logical systems, but not a universal property. Decidability depends on the expressive power and syntactic structure of the system in question.

Undecidability is a feature of certain (especially expressive) logical systems, but not a universal property. Decidability depends on the expressive power and syntactic structure of the system in question.

b.) An example of an expert system that failed is MYCIN, which was focused on diagnosis and treatment of bacterial infections. This system used rule-based reasoning with around 450 if-then rules. Moreover, it used certainty factors to handle uncertainty. Despite outperforming some human doctors in limited tests, MYCIN was never used in clinical settings. Indeed, medical professionals hesitated to rely on a black-box system in high-stakes situations without knowing how it arrived at its conclusions.

An example of an expert system that is still used today is Cyc, which aims to codify a vast base of common-sense knowledge using formal logic, expressed in the Cyc language—a richly expressive predicate logic tailored to support ontological reasoning. The system’s knowledge base includes millions of rules and assertions about everyday concepts, causality, actions, and more. Cyc’s core strength lies in its ability to represent nuanced common-sense knowledge that purely statistical systems often struggle with (e.g., that people usually wear clothes, or that physical objects don’t pass through each other). Moreover, Cyc has been used in hybrid architectures, where statistical models are combined with logical reasoning to interpret results, correct errors, or validate inferences. Lastly, it has been applied in government, defense, and industrial settings. Therefore, this system is still useful today.

Discussion  

Note: The following questions require argumentative writing, i.e. you’re asked to take an informed view on the question, clearly state your position, and the reasons for it.

a.) Logic and system 2

Is all logical thinking system 2 thinking?

b.) Minds and machines

Does Gödel’s theorem show that we cannot build a logic-based AGI?

Solution

a.) Logical thinking is in most cases system 2 thinking but can be system 1 thinking as well. System 2 thinking refers to slow, deliberate, effortful, and conscious reasoning. However, system 1 is fast, automatic, intuitive, and largely unconscious.

System 1 Experts often exhibit system 1 logical reasoning. For instance, a trained mathematician might automatically detect a contradiction in a proof sketch without conscious deliberation. This is similar to chunking in cognitive science: trained subroutines become fast and intuitive.

System 2 Solving a difficult mathematical proof, debugging a computer program, or building a clear logical argument usually depends on System 2 thinking. These tasks take effort because they require one to hold several ideas in mind at once (working memory), think about one’s own thinking (meta-representation), and resist jumping to easy or intuitive answers that might be wrong (inhibitory control). In short, when a task is unfamiliar, complex, or goes against instincts, then slow, careful, and deliberate reasoning is needed to get it right. Therefore, most logical thinking is system 2 thinking.

b.) According to Gödel’s First incompleteness Theorem, any consistent formal system F that is capable of expressing elementary arithmetic is incomplete. Indeed, there exist true statements G in the language of F that cannot be proven within F.

The theorem tells us that no single formal system can be both complete and consistent for arithmetic. This does not imply that a machine cannot recognize or output Gödel sentences. It only constrains what provable conclusions the machine can produce within a fixed formalism. In fact, any Turing machine simulating a particular formal system is subject to the same limitation, but a Turing machine can switch systems, iterate meta-reasoning, or simulate alternative logics.

A logic-based AGI can be equipped with meta-logical capabilities, engage in proof search, probabilistic reasoning or non-monotonic logic, and recognize the limitations of a system and extend it. Thus, Gödel’s theorem does not block the possibility of adaptive, logic-using AGI. Gödel’s theorem places fundamental constraints on formal systems but does not rule out a logic-based AGI. It rules out completeness, not intelligence.