Test: Introduction, Process, Threads & CPU Scheduling- 2

15 Questions MCQ Test GATE Computer Science Engineering(CSE) 2022 Mock Test Series | Test: Introduction, Process, Threads & CPU Scheduling- 2

This mock test of Test: Introduction, Process, Threads & CPU Scheduling- 2 for Computer Science Engineering (CSE) helps you for every Computer Science Engineering (CSE) entrance exam. This contains 15 Multiple Choice Questions for Computer Science Engineering (CSE) Test: Introduction, Process, Threads & CPU Scheduling- 2 (mcq) to study with solutions a complete question bank. The solved questions answers in this Test: Introduction, Process, Threads & CPU Scheduling- 2 quiz give you a good mix of easy questions and tough questions. Computer Science Engineering (CSE) students definitely take this Test: Introduction, Process, Threads & CPU Scheduling- 2 exercise for a better result in the exam. You can find other Test: Introduction, Process, Threads & CPU Scheduling- 2 extra questions, long questions & short questions for Computer Science Engineering (CSE) on EduRev as well by searching above.

The only state transition that is initiated by the user process itself is


The only state transition that is initiated by user process itself is block.


Pre-emptive scheduling, is the strategy of temporarily suspending a running process


In pre-emptive scheduling suspanding temporarily a running process is done before the CPU time slice expires.


In Round Robin CPU scheduling, as the time quantum is increased, the average turn around time


The Round Robin CPU scheduling technique will become insignificance if the time slice is very small or very large. When it is large it acts as FIFO which does not have a fixed determination over the turn around time. If processes with small burst arrived earlier turn around time will be less else it will be more.


Suppose that a process spends a fraction p of its time in I/O wait state. With n processes in memory at once, the probability that all n processes are waiting for I/O is 


Considering that a process spends a fraction p of its time in I/O wait state. As n processes are there is memory at once, then Probability that all n processes are waiting for I/O = p x p x p x ... p (n times) = pn
Note: Moreover, the CPU utilization is given by the formula CPU utilization = 1 - pn.


In a multi-user operating system, 20 requests are made to use a particular resource per hour, on an average. The probability that no requests are made in 45 minutes is


The arrival pattern is a Poission distribution.


In which of the following scheduling policies doe context switching never take place?
1. Round-robin
2. Shortest job first(non pre-emptive)
3. Pre-emptive
4. First-cum-first-served


Context switching takes place when a process is preempted (forcefully) and another process goes into the running state. In FIFO and SJF (non- preemptive) techniques, the processes finishes their execution then only their context is switched to other processes.


Which of the following is the most suitable scheduling scheme in a real-time operating system?


In preemptive scheduling the process with higher priority is executed first and then next higher process with least priority is executed after the higher priority processes are executed, if occured simultaneously. This is the ideal technique to be used in real time O.S. in which critical or higher priority processes are to be entertained first.


Which of the following scheduling algorithms gives minimum average waiting time?


SJF will gives minimum average waiting time. SJF ensures that all smaller processes will finish first. So, less processes will be waiting since more smaller processes will end up in less number of time. Use processes waiting means less waiting time overall which will decrease the average waiting time.


A process executes the following code: 

The total number of child processes created is


A fork () call creates the child processes for the following loop creates 2n - 1 child processes, excluding the parent process. And including parent it is 2n.


While designing a kernel, an operating system designer must decide whether to support kernel- level or user-level threading. Which of the following statements is/are true?
1. Kernel-level threading may be preferable to user-level threading because storing information about user-level threads in the process control block would create a security risk.
2. User-level threading may be preferable to kernel-level threading because in user-level threading, if one thread blocks on I/O, the process can continue.


Kernel level threading may be preferrable to user level threading because storing information about user level thread in a PCB would create a security risk i.e. with each access to the non-critical services. We are going to the domain where both critical and non-critical services are residing. Any harm in this domain may creates problems to the critical services. Hence Kernel level threading is preferable.


Which of the following should be allowed only in Kernel mode?
1. Changing mapping from virtual to physical address.
2. Mask and unmask interrupts
3. Disabling all interrups 
4. Reading status of processor
5. Reading time of day


Only critical services must resides in the kernel. All services mentioned except reading status of processors and reading time of the day are critical.


Consider the following statements with respect to user-level threads and Kernel-supported threads
(i) Context switching is faster with Kernel- supported threads.
(ii) For user-level threads, a system call can block the entire process.
(iii) Kernel-supported threads can be scheduled independently.
(iv) User-level threads are transparent to the Kernel.
Which of the above statements are true?


Kernel level threads can be scheduled independently. For user level threads a system call can block the entire process and are not transparent to Kernel.


Consider the following statements about user level threads and kernel level threads. Which one of the following statements is FALSE?

  • User level threads are not recognised by operating system, hence do not need any hardware support. Since, user level threads are not recognised by operating system, no separate PCB (Process Control Block) information is to stored. Hence, context switching is faster.  
  • Since, Kernel level threads are like separate processes. So, blocking one thread will not block other threads.

Which of the following algorithms favour CPU bound processes?
1. Round-robin
2. Rirst-come-first-served
3. Multilevel feedback queues


Only FCFS and non-preempting algorithms favour CPU bound processes.


Consider a system contains two types of processes CPU bound processes and I/O bound processes, what will be the sufficient condition when ready queue becomes empty.


It all the processes are I/O bound, they will go or reside into BLOCKED state and hence READY QUEUE becomes empty.

Similar Content

Related tests