COLLEGE NAME, BHOPAL DEPARTMENT OF COMPUTER SCIENCE & ENGINEERING
COURSE FILE Program
: B.E.
Semester
: VII
Course Code
:
Subject Name
: Automata and Compiler Design
Prepared By:
Approved By
Index S.No.
Contents
1.
Scheme
2.
Syllabus
3.
Time Table
4.
Lecture Plan
5.
List of Books
6.
Mid Semester Exam Question Papers
7.
RGPV Question Paper
8.
Tutorial Questions
9.
Assignment Questions
10. Hand-Written Notes 11. Transparencies/PPT Slides 12. Mid Semester Exam Result 13. Attendance Sheet
Page No.
PROGRAMME: B.E. Information Technology VII Semester IT 704 Elective –I (IT- 713- Automata and Compiler Design) Unit I: Introduction: Alphabets, Strings and Languages; Automata and Grammars, Deterministic finite Automata (DFA)-Formal Definition, Simplified notation: State transition graph, Transition table, Language of DFA, Nondeterministic finite Automata (NFA), Equivalence of NFA and DFA, Minimization of Finite Automata, Regular Expressions, Arden’s theorem. Unit II: Compiler Structure: Compilers and Translators, Various Phases of Compiler, Pass Structure of Compiler, Bootstrapping of Compiler. Lexical Analysis: The role of Lexical Analyzer, A simple approach to the design of Lexical Analyzer, Implementation of Lexical Analyzer. The Syntactic Specification of Programming Languages: CFG, Derivation and Parse tree, Ambiguity, Capabilities of CFG. Basic Parsing Techniques: Top-Down parsers with backtracking, Recursive Descent Parsers, Predictive Parsers, Unit III: Bottom–up Parsers, Shift-Reduce Parsing, Operator Precedence Parsers, LR parsers (SLR, Canonical LR, LALR) Syntax Analyzer Generator: YACC, Intermediate Code Generation: Different Intermediate forms: three address code, Quadruples & Triples. Syntax Directed translation mechanism and attributed definition. Translation of Declaration, Assignment, Control flow, Boolean expression, Array References in arithmetic expressions, procedure calls, case statements, postfix translation. Unit IV: Run Time Memory Management: Static and Dynamic storage allocation, stack based memory allocation schemes, Symbol Table management Error Detection and Recovery: Lexical phase errors, Syntactic phase errors, Semantic errors. Unit V: Code Optimization and Code Generation: Local optimization, Loop optimization, Peephole optimization, Basic blocks and flow graphs, DAG, Data flow analyzer, Machine Model, Order of evaluation, Register allocation and code selection References: 1.Louden, “Compiler construction”, Cengage learning . 2. Alfred V Aho, Jeffrey D. Ullman, “Principles of Compiler Design”, Narosa. 3. A.V. Aho, R. Sethi and J.D Ullman, “Compiler: principle, Techniques and Tools”, AW. 4. Michal Sipser, “Theory of Computation”, Cengage learning. 5. H.C. Holub, “Compiler Design in C”, Prentice Hall Inc. 6. Hopcroft, Ullman, “Introduction to Automata Theory, Languages and Computation”, Pearson Education.
Time Table
Department Name of Teacher Subject
Information Technology
Compiler Design
Session :
2014
Sem
VII
Sub. Code
IT-713
(B) TIME SCHEDULE : Total expected periods:___, Extra periods (if required)_____
Lecture Plan Day
Mon
Tue
Wed
Thu
Fri
Sat
No. of Periods
01
01
01
01
01
01
Lecture.
Topics to be covered
Max. availab le Planned Date of Completion
Remarks
UNIT-I 1
Alphabets, Strings and Languages
2
Automata and Grammars
3
Deterministic finite Automata (DFA)
4
State transition graph, Transition table
5
Nondeterministic finite Automata (NFA
6
Equivalence of NFA and DFA,
R1:1, R2:1 R1:3, R2:1 R1:10 , R2:5 R1:84,8 8,R2:74 R1:105, R2:103 R2:20
7
Minimization of Finite Automata
R1:725 ,R2:24
8
Regular Expressions, Arden’s theorem
NOTES Unit-II
9
Compilers and Translators
R2:126
10
Various Phases of Compiler
R2-147
11
Pass Structure of Compiler, Bootstrapping of Compiler
12
The role of Lexical Analyzer
13
A simple approach to the design of Lexical Analyzer,
14
CFG, Derivation and Parse tree,
15
Ambiguity, Capabilities of CFG. Basic Parsing
16
Top-Down parsers with backtracking
17
Recursive Descent Parsers
R1:215, 193, R2:158 R1:215, 193, R2:158 R1:215, 193, R2:158 R1:52,2 07, R2:146 R1:215, 193, R2:158 R1:227, 233,R3: 148 R1:257
18
Predictive Parsers
R1:299, R5:154 Unit-III
24
Bottom–up Parsers
25
Shift-Reduce Parsing
R1:360, R4:8389 R1:-350
26
Operator Precedence Parsers
R1-366
27
LR parsers
NOTES
28
Syntax Analyzer Generator
29
YACC, Intermediate Code Generation
30
three address code
31
Quadruples & Triples.
32
. Translation of Declaration, Assignment
R1:408, R5:252 R1:413, R5:278 R2:328, R2:336 R1:478, R2:254 R1:500, R2:271
33
Control flow, Boolean expression
NOTES
34
Array References in arithmetic expressions
35
procedure calls
R1:526, R2:521 R1:28
36
case statements
R1:526,
37
postfix translation
NOTES Unit –IV
39
Run Time Memory Management
R2-408
40
Static and Dynamic storage allocation
41 42
stack based memory allocation schemes Symbol Table management Error Detection and Recovery
R1:608,R2 :410 R1:608,R2 :410 R2-408
43
Lexical phase errors
R3:661
44
Syntactic phase errors, Semantic errors
R2-408 Unit –V
45
Code Optimization and Code Generation
R2-408
46
Local optimization
47
Loop optimization
48 49
Peephole optimization Basic blocks and flow graphs
R1:608,R2 :410 R1:608,R2 :410 R2-408
50
DAG
R2-408
51
Data flow analyzer
R1:608,R2 :410
52
Machine Model
R1:608,R2 :410
53
Order of evaluation
R2-408
54
Register allocation and code selection
R3:661
R3:661
REFERENCE BOOKS R1: Compilers Principles, Techniques, and Tools by Aho,Ullman & Sethi, Pearson Education R2: Principles of Compiler Design by Aho & Ullman, Narosa Publishing House R3:Compiler Construction 2/e: Dhamdhere R4: Compiler Design by Santanu Chattopadhyay, PHI R5: Compilers Construction & Design by Rajni Jindal, Umesh Publications R6: Compiler Design by O. G. Kakde, Laxmi Publications (p) LTD
Websites: 1. www.cs.uccs.edu/~abudjen/classsnotes.doc 2. www.os.iitb.ac.in/~sri/notes/lexical.pdf 3. www.iitb.ac.in/~sri/notes/compiler/regex.pdf
List of Books 1. A. V. Aho, R. Sethi, and J. D. Ullman. Compilers: Principles, Techniques and Tools , Pearson Education 2 Raghavan, Compiler Design, TMH Pub. 3. Louden. Compiler Construction: Principles and Practice, Cengage Learning 4. A. C. Holub. Compiler Design in C , Prentice-Hall Inc., 1993. 5. Mak, writing compiler & Interpreters, Willey Pub.
COLLEGE NAME MID SEMESTER I BRANCH- IT SEMESTER: VII SUBJECT: COMPILER DESIGN SUBJECT CODE: IT-713 Max. Marks: 40 Note: All questions are compulsory carry equal marks.
Time: 2 Hours
1. (a) What is Ad hoc network and how it differs from other network? (b) Explain various challenges of mobile Ad hoc network. 2. What is Source and Receiver oriented MAC Routing Protocol? 3. Write the difference betweena. Cellular and Mobile Ad hoc Network b. Proactive and Reactive Routing protocol 4. Explain various design issues of Ad hoc Network in brief. 5. Explain the Working of DSDV Routing protocol
COLLEGE NAME MID SEMESTER II BRANCH- IT SEMESTER: VII SUBJECT: COMPILER DESIGN SUBJECT CODE: IT-713
Max. Marks: 40 Note: Attempt any four questions.
Time: 2 Hours
1. (a) Explain dynamic source routing (DSR) protocols with its advantages and disadvantages. (b) Evaluate the route discovery (RD) time parameter in the communication performance of ad-hoc network.
5
2. (a) Describe the power management at various layers. (b) What are smart batteries? What are its characteristics ?
5 5
3. (a) Explain the ATM cell header at UNI and at NNI.
5
(b) What are the advantages and disadvantages of packet switching over circuit switching? 4. (a) Draw and discuss the ATM protocol architecture model. (b) With neat sketch, explain architecture of 802.11 LAN. Also, explain its MAC logic. 5.
Write short note on (any two) i) ZRP protocol ii) EED performance in ad-hoc network iii) X.25 iv) AAL
5
5 5 5 10
COLLEGE NAME, Bhopal Department of Information Technology Assignment-1
Subject: Compiler Design Subject Code: IT-713 Unit-1 1. 2. 3. 4. 5. 6.
Give the reasons for the separation of scanner and parser in separate phases of compiler. Describe the role of lexical analyzer in recognizing tokens. Explain the concept of input buffering. Explain how tokens are recognized. What is simple approach to design of lexical analyzer for an identifier. What’s LEX? Describe auxiliary definitions and translation rules for LEX with suitable example. 7. What are the tasks performed by the compiler in lexical and syntax analysis phases. Explain with help of examples. 8. Explain role of symbol table in various phases of compiler
COLLEGE NAME, Bhopal Department of Information Technology Assignment-2 Subject: Compiler Design Subject Code: IT-713 Unit-2 1. 2. 3. 4. 5.
Write in brief about the error recovery procedure in LL and LR parsing. Write short note on automatic parser generator. Describe the function of LALR parser generator YACC. What is meant by syntax directed translation? Explain. Give the parse tree and translation for expression 23* 5+4 according to the syntax directed translation scheme. 6. Differentiate between synthesized translation and inherited translations. 7. Let the synthesized attribute “val”, give the integer value associated with non terminals in following grammarL→ E E →E +T | T T→ T*F | F F→(E)| digit 8. Write a brief note on syntax tree. 9. For the following grammar find FIRST and FOLLOW sets for each non terminalS→aAB |bA|€ A→aAb|€ B→bB|€ Where € is Null string. 10. What is Shift-Reduce and Reduce-Reduce conflict? How these can be resolved? With examples explain in which condition S-R and R-R conflict can occur in SLR, Canonical LR and LALR parsers. (Make use of LR(0), LR(1) items.
COLLEGE NAME, Bhopal Department of Information Technology Assignment-3 Subject: Compiler Design Subject Code: IT-713 Unit-3 1. What do you mean by heap allocation? Explain the following terms related to heap allocation(i) Fragmentation (ii) Free list (iii) Reference counts 2. Explain the difference between static, stack and heap allocation. 3. What is the difference between dynamic and static storage management? 4. What are different parameter passing mechanisms? 5. Explain with a suitable example, mechanisms, used by the compiler to handle procedure parameters. 6. Write short note on symbol table organization. 7. Explain various symbol table management techniques. 8. Explain various data structures used for implementing the symbol table and compare them. 9. What is hashing? What are different types of hashing techniques available?
COLLEGE NAME, Bhopal Department of Information Technology Assignment-4 Subject: Compiler Design Subject Code: IT-713 Unit-4 1. 2. 3. 4. 5. 6. 7.
Define Leaders. Explain DAG construction. What are applications of DAGs? Write advantages of DAG. Write short note on application of DAG in code generation. Discuss the various methods of translating Boolean expression. Construct DAG of basic block after converting code in 3-address representationBegin Prod:=0; i :=1; do begin prod:= prod+a[i]*b[i]; i:=i+1; end while i<=20 end
8. Translate the following expression into quadruples, triples and indirect triples.
COLLEGE NAME, Bhopal Department of Information Technology Assignment-5 Subject: Compiler Design Subject Code: IT-713 Unit-5 1. What is global data flow analysis? What is its use in code optimization? 2. Describe global data flow analysis. 3. Write the criteria for code improving transformations. Explain the principal sources of optimization. 4. Define dominators and write short note on loops in flow graph.
COLLEGE NAME, Bhopal Department of Information Technology Tutorial-1 Subject: Compiler Design Subject Code: IT-713 Topic to be covered: Unit-1 9. Explain why one should study about the compilers. 10. Explain working of compilers drawing its block diagram. 11. What are the features of good compiler? 12. Compare and contrast the features of single pass compiler with multi pass compiler. 13. Write short note on bootstrapping. 14. Compiler uses two passes scheme of compilation, performing analysis, in first pass and synthesis in second pass. Explain what use is made of the symbol table in two passes of compiler. 15. Draw and explain various phases of compiler 16. Draw the block diagram of lexical and syntax analyzer and state clearly input , output and task performed by them.
COLLEGE NAME, Bhopal Department of Information Technology Tutorial-2 Subject: Compiler Design Subject Code: IT-713 Unit-2 1. Write short note on grammar. What do you mean by ambiguity of grammar? 2. What is top down parsing? What are the difficulties encountered in this and how they overcome? 3. Write short note on recursive decent parser. 4. What is predictive parser? How a parser is controlled by a program? 5. Differentiate top down and bottom up parsing. Give examples of method of each type. 6. What is operator precedence grammar? 7. What Describe operator precedence parsing algorithm. 8. What are the limitations of operator precedence parsing? 9. What do you understand by LR(k) grammar? 10. Write the Algo/ Procedure for construction of canonical LR parsing table.
COLLEGE NAME, Bhopal Department of Information Technology Tutorial-3 Subject: Compiler Design Subject Code: IT-713 Unit-3 1. Give a brief description of type checking. 2. Give the difference between implicit type conversion and explicit type conversion with help of an example. 3. Differentiate between static and dynamic binding. 4. Explain the importance of run time storage management in compiler. 5. What do you mean by activation record? Why this record is maintained by the compiler? 6. Explain various fields of activation record. 7. Describe storage allocation strategies. 8. What do you mean by static allocation? What are its drawback?
COLLEGE NAME, Bhopal Department of Information Technology Tutorial-4 Subject: Compiler Design Subject Code: IT-713 Unit-4 9. Write a translation scheme to generate intermediate code for assignment statement with array references. 10. Write syntax directed definition to translate ‘switch’ statement. With a suitable example, show translation of the source language ‘switch’ statement. 11. Write short note on back patching. 12. Write short note on code generation. 13. What are general issues in designing a code generator? 14. Explain code generation algorithm. 15. What is basic block? With suitable example discuss various transformations on the basic blocks. 16. What are the different types of intermediate codes? Explain in brief.
COLLEGE NAME, Bhopal Department of Information Technology Tutorial-5 Subject: Compiler Design Subject Code: IT-713 Unit-5 1. Explain why code optimization is called optional phase? 2. Explain the principal sources of optimization with suitable example. 3. Describe the necessary and sufficient condition performing constant propagation and dead code eliminations. 4. Explain the following optimizing transformations with suitable examples(i) Code movement (ii) Common sub-expression elimination 5. Define dominators.
Unit wise blow- Up UNIT – I A language translator is a program which translates programs from source language in an equivalent program in an object language. The source language is usually a high-level programming language and the object language is usually the machine language of an actual computer. A compiler is a translator whose source language is a high-level language and whose object language is close the machine language of an actual computer. The typical compiler consists of an analysis phase and a synthesis phase. In contrast with compilers an interpreter is a program which simulates the execution of programs written in a source language. Interpreters may be used either at the source program level or an interpreter may be used it interpret an object code for an idealized machine. The typical compiler consists of several phases each of which passes its output the next phase (i)
(ii) (iii)
(iv) (v) (vi)
The lexical phase (scanner) groups characters in lexical units or kens. The input the lexical phase is a character stream. The output is a stream of kens. Regular expressions are used define the kens recognized by a scanner (or lexical analyzer). The scanner is implemented as a finite state machine. The parser groups kens in syntactical units. The output of the parser is a parse tree representation of the program. Context-free grammars are used define the program structure recognized by a parser. The parser is implemented as a push-down automata. The contextual analysis phase analyzes the parse tree for context-sensitive information often called the static semantics. The output of the contextual analysis phase is an annotated parse tree. Attribute grammars are used describe the static semantics of a program. The optimizer applies semantics preserving transformation the annotated parse tree simplify the structure of the tree and facilitate the generation of more efficient code. The code general transforms the simplified annotated parse tree in object code using rules which denote the semantics of the source language. The peephole optimizer examines the object code, a few instructions at a time, and attempts do machine dependent code improvements.
Types of Compiler1. 2. 3. 4.
One-pass compiler Multi-pass Compiler Load & Go Compiler Optimized Compiler
A one-pass compiler reads the program only once, and translates it at the same time as it is reading. A multi-pass compiler reads the program several times, each time transforming it in a different form and usually in a different data structure. Cross Compiler-A cross compiler compiles a target language different from the language of the machine it runs on
Bootstrapping- Bootstrapping describe the techniques involved in writing a compiler or assembler) in the target programming language which it is intended compile.
LEX- LEX is a program general designed for lexical processing of character input streams. It accepts a high-level, problem oriented specification for character string matching, and produces a program in a general purpose language which recognizes regular expressions. The regular expressions are specified by the user in the source specifications given LEX. The LEX written code recognizes these expressions in an input stream and partitions the input stream in strings matching the expressions
UNIT II The parsing is a process of finding a parse tree for a string of kens. Equivalently, it is a process of determining whether a string of kens can be generated by a grammar. There are two types of Parsing
Top-down Parsing (start from start symbol and derive string) A Top-down parser builds a parse tree by starting at the root and working down wards the leaves. o Easy generate by hand. o Examples are- Recursive- descent, Predictive. Bottom-up Parsing (start from string and reduce start symbol) A bottom- up parser builds a parser tree by starting at the leaves and working up wards the root. o Not easy handle by hands, usually compiler- generating software generate bottom up parser o But handles larger class of grammar o Example is LR parser. Operator precedence parsing- Bottom-up parsers for a large class of context-free grammars can be easily developed using operator grammars. Operator grammars have the property that no production right side is empty or has two adjacent non terminals. This property enables the implementation of efficient operator-precedence parsers SLR parser - Simple LR parser or SLR parser is an LR parser or which the parsing tables are generated as for an LR(0) parser except that it only performs a reduction with a grammar rule A → w if the next symbol on the input stream is in the follow set of A. Such a parser can prevent certain shift-reduce and reduce-reduce conflicts. That occur in LR (0) parsers and it can therefore deal with more grammars. However, it still cannot parse all grammars that can be parsed by an LR(1) parser. A grammar that can be parsed by an SLR parser is called a SLR grammar. Yacc-YACC stands for “Yet Another Compiler-Compiler”. this is because this kind of analysis of text files is normally associated with writing compilers. Yacc provides a general ol for imposing structure on the input a computer program. The Yacc user prepares a specification of the input process; this includes rules describing the input structure, code be invoked when these rules are recognized, and a low-level routine do the basic input. Yacc then generates a function control the input process
UNIT III A syntax- directed translation is used define the translation of a sequence of kens some other value, based on a CFG for the input. A syntax- directed translation is defined by associating a translation rule with each grammar rule. A translation rule defines the translation of the left- hand -side conterminal as a function of eight-hand- side no terminals' translations, and the values of the right-hand-side terminals. compute the translation of a string, build the parse tree, and use the translation rules compute the translation of each conterminal in the tree, bottom-up; the translation of the string is the translation of the root conterminal. There is no restriction on the type of converted actions that manipulate the parser's semantic stack. Each action must pop all right-hand-side non terminals' translations from the semantic stack, then compute and push the left-hand-side non terminal's translation. Next, the actions are incorporated (as action numbers) in the grammar rules. Finally, the grammar is converted LL (1) form (treating the action numbers just like terminal or conterminal symbols).
Intermediate Code- The semantic phase of a compiler first translates parse trees in an intermediate representation (IR), which is independent of the underlying computer architecture, and then generates machine code from the IRs. This makes the task of retargeting the compiler another computer architecture easier handle 1. High-Level Intermediate Languages
Abstract syntax tree. Postfix notation is a linearized representation of an abstract syntax tree. EXAMPLE- a = b*c + d is a b c * d + assign
Dependence graphs. Medium-Level Intermediate Language
Reflect a range of features in a language-independent way. Represent source variables, temporaries, and registers. Appropriate for most of the optimizations done in compilers such as common-sub expression elimination, code motion, and algebraic simplification.
Three address code (x = y op z), o assignment statements x := y op z o assignment statements x := op y o copy x := y o unconditional jump go L o conditional jump if re loop go L
o param x and call p, n for procedure calls and return y -- where y representing a return value is optional o indexed statements of the form x:= y[i], x[i] := y o address and pointer assignments of the form x := &y and x := *y
Low-Level Intermediate Languages one--one correspondence target machine instructions, and are usually architecture dependent. Usually deviate only when there are alternatives for the most effective code generate for them. E.g., multiply instruction or more complex addressing modes.
UNIT IV Activation records Created every time a procedure is called Must be accessible both the caller and the callee Allocates space for o Parameters o Local variables o Return address o Other links and pointers provide access non-local data Other issues o Initializing local variables o Stack vs. heap allocated o Optimizing activation records by coalescing Symbol Table keeps track of scope and other attributes of named program entities key operations o void insert(symbol s); o symbol lookup(string name); o void enter_scope(void); o void exit_scope(void); implementations o list o hash table o stack of tables for some languages, the symbol table must handle overloading o each identifier contains a list of symbols o when entering new scope, chain symbols with same name in previous scope
UNIT V Directed Acyclic Graph (DAG)– Directed acyclic graph DAG• Contraction of AST that avoids duplication- identical sub trees are reused. • exposes redundancies- changes (assignments, calls) ? • smaller memory footprint o Example- a x (a-b) + c x (a-b)
Code optimization together with code generation form the back end of the compiler. In compilers with a very extensive optimization, the optimization phase is distinguished as a middle end. The goal of the compiler's optimizer is transform the IR program created by the front end in an IR program that computes the same results in a better way. Here "better" can take on many meanings. It usually implies faster code, but it might imply more compact code, less power when it runs or costs less run under some model. Ideally, compilers should produce target code that is as good as can be written by hand. The reality is that this goal is achieved only in limited cases, and with difficulty. However, the code produced by straightforward compiling algorithms can often be made run faster or take less space, or both. This improvement is achieved by program transformations that are traditionally called optimizations, although the term "optimization" is a misnomer because is rarely a guarantee that the resulting code is the best possible. Most of the compilers involve some optimization. Compilers that apply code-improving transformations are called optimizing compilers. Code generation is the final phase in compilation. It takes as input an intermediate representation of the source program and produces as output an equivalent target program. The code generation techniques can be used whether or not an optimization phase occurs before code generation. The requirements traditionally imposed on a code generar are severe. The output code must be correct and of high quality, meaning that it should make effective use
of the resources of the target machine. Moreover, the code generar itself should run efficiently. Mathematically, the problem of generating optimal target code is (TM) undecidable! In practice, we must be content with heuristic techniques that generate good, but not necessarily optimal, code. The choice of heuristics is important, because a carefully designed code generation algorithm can easily produce code that is several times faster that produced with ad hoc code generation techniques. As code generation begins, the program exists in IR form. The code generar must convert the IR program (perhaps, already optimized) in code that can run on the target machine. The code generation is performed typically as a sequence- instruction selection, instruction scheduling and register allocationo Instruction selection - selecting a sequence of target-machine operations that implement the IR operations. o Instruction scheduling - choosing an order in which the operations should execute. o Register allocation - deciding which values should reside in registers at each point in the program Most compilers handle each of these three processes separately. The term code generation is often used refer instruction selection only. When the level of abstraction of the IR and the target machine differ significantly, or the underlying computation models differ, instruction selection can play a critical role in bridging the gap. The extent which instruction selection can map the computation in the IR program efficiently the target machine will often determine the efficiency of the generated code. For example, consider three scenarios for generating code from an IRo A simple, scalar RISC machine - the mapping from IR assembly is straightforward. The code generar might consider only one or two assemblylanguage sequences for each IR operation. o A CISC processor - make effective use of a CISC's instruction set, the compiler may need aggregate several IR operations in a single targetmachine operation. o A stack machine - the code generar must translate from the register--register computational style of IR a stack based style with its implicit names and, in some cases, destructive operations. As the gap in abstraction between the IR and the target ISA grows, so does the need for ols help build code generars. While instruction selection can play an important role in determining code quality, the compiler writer must keep in mind the enormous size of the search space that the instruction select might explore. As we shall see, even moderately sized instruction sets can produce search spaces that contain hundreds of millions of states. Clearly, the compiler cannot afford explore such spaces in either a careless or an exhaustive way.
Previous Univ. Exam Paper
CS – 701 B.E. (Seventh Semester ) EXAMINATION , June, 2009 (Common for CS & IT Engg.) COMPILER DESIGN Time - Three Hours Maximum Marks - 100 Minimum Pass Marks - 35 Note - Attempt all questions. All question carry equal marks. 1. (a) Consider the following LEX program Auxiliary Definitions (none) Translation Rules Token Pattern Key word Identifier Constant Operator
Action {Return KW} (Return id} {Return const.} {Return OP}
Implement the LEX program as DFA. (b) Describe the role of a lexical analyzer and also explain the concept of Input Buffering. OR 2. (a) Consider a finite – state automata in which S = {q0, q1, q2, q3}, Σ = {0, 1} S0 = q0 F = {q0} And the transition function δ is given in the following table State q0 q1 q2 q3
0 q2 q3 q0 q1
1 q1 q0 q3 q2
a) Give the entire sequence of states for the input string 101101. b) Find out the string among the following string which are accepted by the given finite- state automation c) 101101 d) 11111 e) 000000 (b) Explain why should one study about compilers. 3. (a) What do you mean by ambiguous and unambiguous grammar? Explain with example. (b) Construct the parsing table for the following grammar. S aXYb X c/ € Y d/€ OR 4. (a) Consider the grammar S ACB/ C b B/ B a A da /BC B g / € C h / € Calculate FIRST and Follow. (b) Show that the following grammar is U (1), by constructing its parse tableS a AB / b A / € A a A b / € B b B / € 5. (a) Translate the following statement to quadruple If a > b then x = a + b else x = a - b (b) Translate the following expression to quadruple, triple- (x + y) * (z + c) – (x + y + z) OR 6. Write short notes on the following -
7. S- attributed definitions 8. L- Attributed definition 9. Dependency graph 10. (a) Explain the role of symbol table in various phases of compiler. (b) What do you mean by heap allocation? Explain the following terms related to heap allocation11. Fragmentation 12. Free list 13. Bit map 14. Reference counts OR 15. (a) Explain various storage allocation strategies. Which storage allocation model is to be used if a language permits recursion ? (b) Describe parameter passing mechanisms for a procedure call. 16. (a) Describe the necessary and sufficient conditions for performing constant propagation and dead code eliminations. (b) Write and explain data flow equations. OR 17. Explain the following 18. Loop invariant conditionals 19. Conditional depending on loop index 20. Independent loop conditionals 21. Reduction loops ***********************************************************************
Model Paper CS-701 COMPILER DESIGN Time - 3 Hrs Max. Marks-100 Minimum Pass Marks - 35 Note - Solve any five questions. All questions carry equal marks. 1.(a) What is the basic task of scanning. What are the difficulties faced in delimeter oriented scanning. How can this be removed. (b) What is the structure of a compiler. Discuss phases of a compiler briefly. 2. (a) What are errors can be encountered by virtually all the phases of a compiler. (b) Explain the concept of transition diagram and its use in building a lexical analyzer. 3 (a) Why are multiple passes required in a compiler. Describe strategies for reducing the number of passes. (b) What are data structures used for symbol tables management. Explain each of them in brief. 4(a) Comment on the need for indirection in symbol table. (b) What is p-down parsing. What are the difficulties encountered in it and how are they overcome. 5(a) Write the quadruples, triples and indirection triples for the expression A B + C*D | E (b) Why is the dynamic srage allocation strategy particularly amendable the run-time usage requirements of block structured languages. Write the procedure for run-time address calculation in our dynamic srage allocation mode. 6. (a) Describe Global Data Flow analysis. How is it useful in code optimization. (b) Define the ambiguity of grammer. Suggest some possible solutions remove the ambiguity of a grammer. 7. Given a grammer with the following productions. SE ET|E–T TF|T*F F i | (E)
Generate the sets of LR (1) items Give a trtace of the Parse of the following input string (i – i) * i.
8. What is operar precedence grammer? For given the grammer: EE+T|T TT*F|F F (E) | id Obtain the following
Operator precedence matrix The parse for input string id + id