Theory of Computer Science C8. Type-1 and Type-0 Languages: Closure - - PowerPoint PPT Presentation

theory of computer science
SMART_READER_LITE
LIVE PREVIEW

Theory of Computer Science C8. Type-1 and Type-0 Languages: Closure - - PowerPoint PPT Presentation

Theory of Computer Science C8. Type-1 and Type-0 Languages: Closure & Decidability Gabriele R oger University of Basel April 15, 2020 Turing Machines vs. Grammars Closure and Decidability Summary Overview Languages & Grammars


slide-1
SLIDE 1

Theory of Computer Science

  • C8. Type-1 and Type-0 Languages: Closure & Decidability

Gabriele R¨

  • ger

University of Basel

April 15, 2020

slide-2
SLIDE 2

Turing Machines vs. Grammars Closure and Decidability Summary

Overview

Automata & Formal Languages Languages & Grammars Regular Languages Context-free Languages Context-sensitive & Type-0 Languages Turing machines Closure properties & decidability

slide-3
SLIDE 3

Turing Machines vs. Grammars Closure and Decidability Summary

Turing Machines vs. Grammars

slide-4
SLIDE 4

Turing Machines vs. Grammars Closure and Decidability Summary

Reminder: Turing Machines – Conceptually

. . . b a c a c a c a . . . infinite tape read-write head

slide-5
SLIDE 5

Turing Machines vs. Grammars Closure and Decidability Summary

Reminder: Nondeterministic Turing Machine

Definition (Nondeterministic Turing Machine) A nondeterministic Turing machine (NTM) is given by a 7-tuple M = Q, Σ, Γ, δ, q0, , E with: Q finite non-empty set of states Σ = ∅ finite input alphabet Γ ⊃ Σ finite tape alphabet δ : (Q \ E) × Γ → P(Q × Γ × {L, R, N}) transition function q0 ∈ Q start state ∈ Γ \ Σ blank symbol E ⊆ Q end states

slide-6
SLIDE 6

Turing Machines vs. Grammars Closure and Decidability Summary

One Automata Model for Two Grammar Types?

Don’t we need different automata models for context-sensitive and type-0 languages?

Picture courtesy of stockimages / FreeDigitalPhotos.net

slide-7
SLIDE 7

Turing Machines vs. Grammars Closure and Decidability Summary

Linear Bounded Automata: Idea

Linear bounded automata are NTMs that may only use the part of the tape occupied by the input word.

  • ne way of formalizing this: NTMs where blank symbol

may never be replaced by a different symbol

slide-8
SLIDE 8

Turing Machines vs. Grammars Closure and Decidability Summary

Linear Bounded Turing Machines: Definition

Definition (Linear Bounded Automata) An NTM M = Q, Σ, Γ, δ, q0, , E is called a linear bounded automaton (LBA) if for all q ∈ Q \ E and all transition rules q′, c, y ∈ δ(q, ) we have c = .

German: linear beschr¨ ankte Turingmaschine

slide-9
SLIDE 9

Turing Machines vs. Grammars Closure and Decidability Summary

LBAs Accept Type-1 Languages

Theorem The languages that can be accepted by linear bounded automata are exactly the context-sensitive (type-1) languages. Without proof. proof sketch for grammar ⇒ NTM direction: computation of the NTM follows the production of the word in the grammar in opposite order accept when only start symbol (and blanks) are left on the tape because language is context-sensitive, we never need additional space on the tape (empty word needs special treatment)

slide-10
SLIDE 10

Turing Machines vs. Grammars Closure and Decidability Summary

LBAs Accept Type-1 Languages

Theorem The languages that can be accepted by linear bounded automata are exactly the context-sensitive (type-1) languages. Without proof. proof sketch for grammar ⇒ NTM direction: computation of the NTM follows the production of the word in the grammar in opposite order accept when only start symbol (and blanks) are left on the tape because language is context-sensitive, we never need additional space on the tape (empty word needs special treatment)

slide-11
SLIDE 11

Turing Machines vs. Grammars Closure and Decidability Summary

NTMs Accept Type-0 Languages

Theorem The languages that can be accepted by nondeterministic Turing machines are exactly the type-0 languages. Without proof. proof sketch for grammar ⇒ NTM direction: analogous to previous proof for grammar rules w1 → w2 with |w1| > |w2|, we must “insert” symbols into the existing tape content; this is a bit tedious, but not very difficult

slide-12
SLIDE 12

Turing Machines vs. Grammars Closure and Decidability Summary

NTMs Accept Type-0 Languages

Theorem The languages that can be accepted by nondeterministic Turing machines are exactly the type-0 languages. Without proof. proof sketch for grammar ⇒ NTM direction: analogous to previous proof for grammar rules w1 → w2 with |w1| > |w2|, we must “insert” symbols into the existing tape content; this is a bit tedious, but not very difficult

slide-13
SLIDE 13

Turing Machines vs. Grammars Closure and Decidability Summary

Deterministic Turing Machines

Definition (Deterministic Turing Machine) A deterministic Turing machine (DTM) is a Turing machine M = Q, Σ, Γ, δ, q0, , E with δ : (Q \ E) × Γ → Q × Γ × {L, R, N}.

German: deterministische Turingmaschine

slide-14
SLIDE 14

Turing Machines vs. Grammars Closure and Decidability Summary

Deterministic Turing Machines vs. Type-0 Languages

Theorem For every type-0 language L there is a deterministic Turing machine M with L(M) = L. Without proof. proof sketch: Let M′ be an NTM with L(M′) = L. It is possible to construct a DTM that systematically searches for an accepting configuration in the computation tree of M′. Note: It is an open problem whether an analogous theorem Note: holds for type-1 languages and deterministic LBAs.

slide-15
SLIDE 15

Turing Machines vs. Grammars Closure and Decidability Summary

Deterministic Turing Machines vs. Type-0 Languages

Theorem For every type-0 language L there is a deterministic Turing machine M with L(M) = L. Without proof. proof sketch: Let M′ be an NTM with L(M′) = L. It is possible to construct a DTM that systematically searches for an accepting configuration in the computation tree of M′. Note: It is an open problem whether an analogous theorem Note: holds for type-1 languages and deterministic LBAs.

slide-16
SLIDE 16

Turing Machines vs. Grammars Closure and Decidability Summary

Questions Questions?

slide-17
SLIDE 17

Turing Machines vs. Grammars Closure and Decidability Summary

Closure Properties and Decidability

slide-18
SLIDE 18

Turing Machines vs. Grammars Closure and Decidability Summary

Overview

Automata & Formal Languages Languages & Grammars Regular Languages Context-free Languages Context-sensitive & Type-0 Languages Turing machines Closure properties & decidability

slide-19
SLIDE 19

Turing Machines vs. Grammars Closure and Decidability Summary

Closure Properties

Intersection Union Complement Concatenation Star Type 3 Yes Yes Yes Yes Yes Type 2 No Yes No Yes Yes Type 1 Yes(2) Yes(1) Yes(2) Yes(1) Yes(1) Type 0 Yes(2) Yes(1) No(3) Yes(1) Yes(1) Proofs? (1) proof via grammars, similar to context-free cases (2) without proof (3) proof in later chapters (part D)

slide-20
SLIDE 20

Turing Machines vs. Grammars Closure and Decidability Summary

Decidability

Word problem Emptiness problem Equivalence problem Intersection problem Type 3 Yes Yes Yes Yes Type 2 Yes Yes No No Type 1 Yes(1) No(3) No(2) No(2) Type 0 No(4) No(4) No(4) No(4) Proofs? (1) same argument we used for context-free languages (2) because already undecidable for context-free languages (3) without proof (4) proofs in later chapters (part D)

slide-21
SLIDE 21

Turing Machines vs. Grammars Closure and Decidability Summary

Questions Questions?

slide-22
SLIDE 22

Turing Machines vs. Grammars Closure and Decidability Summary

Summary

slide-23
SLIDE 23

Turing Machines vs. Grammars Closure and Decidability Summary

Summary

Turing machines accept exactly the type-0 languages. This is also true for deterministic Turing machines. Linear bounded automata accept exactly the context-sensitive languages. The context-sensitive and type-0 languages are closed under almost all usual operations.

exception: type-0 not closed under complement

For context-sensitive and type-0 languages almost no problem is decidable.

exception: word problem for context-sensitive lang. decidable

slide-24
SLIDE 24

Turing Machines vs. Grammars Closure and Decidability Summary

What’s Next?

contents of this course:

  • A. background

⊲ mathematical foundations and proof techniques

  • B. logic

⊲ How can knowledge be represented? ⊲ How can reasoning be automated?

  • C. automata theory and formal languages

⊲ What is a computation?

  • D. Turing computability

⊲ What can be computed at all?

  • E. complexity theory

⊲ What can be computed efficiently?

  • F. more computability theory

⊲ Other models of computability

slide-25
SLIDE 25

Turing Machines vs. Grammars Closure and Decidability Summary

What’s Next?

contents of this course:

  • A. background

⊲ mathematical foundations and proof techniques

  • B. logic

⊲ How can knowledge be represented? ⊲ How can reasoning be automated?

  • C. automata theory and formal languages

⊲ What is a computation?

  • D. Turing computability

⊲ What can be computed at all?

  • E. complexity theory

⊲ What can be computed efficiently?

  • F. more computability theory

⊲ Other models of computability