Skip to main content

Message Delivery Protocol

 

Message Delivery Protocol

·         A protocol is a set of rules and conventions shared by communicating entities .

·         Message passing can be blocking or non blocking .

·         In blocking protocol, a sender process is blocked till the message selivered to the . in this case , sender process has guarantee that the message sent by it ie delivered before it continues its execution .

A blocking protocol normally includes:

1.       Blocking send : the sender process is blocked until the message is received by the receiving process or by the mailbox.

2.       Blocking receive : the receiver nlocks until a message is available thus when both sender as well as receiver is blocked ,it is known as rebdevezevous . this combination allows for tight sysnchronization between process.

·         In non blockeing protocol, asender continues the execution after performing a send operation irrespective of whether the message is delivered or not

·         A non blocking protocol has the advantage that a sender is free to continue its execution immediately after sending a message

·         However , sender has no means of knoiwing when(and wheather )its message  is delivered to the receiver.

     

·         A non blocking send : the sending process sends the message and resumes operation.

·         Non blocking receive: the receiver retrieves either a valid message or a null.

 

Buffering

Wheather the communication is direct or indirect , message exchanged by communicating processes reside in a temporary queue.

There are  three different  ways of implementing such queue.

 

Zero capacity

·         Zero capacity queue cannot keep and message waiting in it , thus it has maximum length 0

·         For this , sending process must be blocked until the receiving process receives the message .

·         System with zero capacity queue is known as message system with no buffering.

 

Bounded capacity

·         Bounded capacity queue has finite length n. thus it can have n message.

waiting in it .

·         If the queue is not full , new message can be placed in the queue and the sending process is not blocked , it not blocked , it can continue execution without waiting.

·         If the link or queue is full , then the sender must be blocked until space is available in the queue.

·         This system is also known as automatic buffering .

 Unbounded capacity

·         Unbounded capacity queue has infinite length . thus any number of messages can wait in it .

In such a system , sending process is never blocked , can continue execution without waiting

Shared memory model

·         Normally , process use memory areas within the scope of virtual memory space . also memory management systems ensure that every process has a  well-defined and separate data and separate data and code area.

·         In shared memory model parts of private address space of two or more rae processes are mapped to same physical memory.

·         Such a memory areas are called shared memory regions.


Comments

Popular posts from this blog

Thread

Thread What is Thread? A thread is a flow of execution through the process code, with its own program counter that keeps track of which instruction to execute next, system registers which hold its current working variables, and a stack which contains the execution history. A thread shares with its peer threads few information like code segment, data segment and open files. When one thread alters a code segment memory item, all other threads see that. A thread is also called a  lightweight process . Threads provide a way to improve application performance through parallelism. Threads represent a software approach to improving performance of operating system by reducing the overhead thread is equivalent to a classical process. Each thread belongs to exactly one process and no thread can exist outside a process. Each thread represents a separate flow of control. Threads have been successfully used in implementing network servers and web server. They also provide a suitable foundation ...

Multi Level Queue Scheduling (MLQ)

  Multi Level Queue Scheduling (MLQ) ·          Multilevel queue scheduling classifies the processes according to their types for example, a multilevel queue scheduling algorithm makes a common. ·          In this scheduling ready queue is divided into various queue that are called sub queues. A subqueue is a distinct operational queue ·          The process are permanently assigned to subqueues, generally based on some property of the process such as memory size,priority or process type ·          Each subqueue has its process sucheduling algorithm. For example interactive process at the foreground may use round robin scheduling while batch jobs at the background may use the FCFS method ·          For example, consider a system with four different queues 1.   ...

Distributed system

  Distributed system ·           A distributed system is a collection of processor located in geographical dispersal physical location. ·           In this system, the workhold is distributed or divided between two or more computers that are linked together by communication network . that is the different processors communicate using communication links, such as telephone lines and buses3 ·           The various processors do not share memory, clock ,or peripheral devices instead each processors has its local memory. ·           The purpose of distributed system is to provide an efficient and convenient environment for sharing of resources. ·           Some example of distributed operating system are amoeba, angle, chorus, much, alpha kernel. ·   ...

Layered Operating System

  Layered Operating System Layered Structure is a type of system structure in which the different services of the  operating system  are split into various layers, where each layer has a specific well-defined task to perform . It was created to improve the pre-existing structures like the Monolithic structure ( UNIX ) and the Simple structure ( MS-DOS ). Example –  The Windows NT operating system uses this layered approach as a part of it . Design Analysis : The whole Operating System is separated into several layers ( from 0 to n ) as the diagram shows. Each of the layers must have its own specific function to perform. There are some rules in the implementation of the layers as follows. The outermost layer must be the User Interface layer. The innermost layer must be the Hardware layer. A particular layer can access all the layers present below it but it cannot access the layers present above it. That is layer n-1 can access all the layers from n-2 to 0 but it canno...

Classification of Operating System

  Classification of operating systems The operating systems may be classified into different types depending upon the nature of interaction between the user and his/her program. The various types of operating system are : 1.       single user operating system    2.        Multi user operating system   3.         Batch processing operating system 4.        Multi programming operating system   5.       Multi tasking operating system   6.         Multiprocessing operating system 7.         Time sharing operating system 8.       Real time system      Distributed system Multi threading operating system       Single user operating system ·          ...