Академический Документы
Профессиональный Документы
Культура Документы
Submitted by
B. KOWSALYA (B5C15065)
A. VINITHA (B5C15097)
Assistant professor
Assistant professor,
Paravai, Madurai-625402
CERTIFICATE
This is to certify that the report entitle “ON THE JOB TRAINING REPORT
ON AAVIN MILK COMPANY” submitted by B. KOWSALYA (B5C15065) &
A. VINITHA (B5C15097)is a bonafide record of their original work carried out
under my supervision. This work has not been submitted elsewhere for any other
degree or diploma.
A. VINITHA(B5C15097)
IIIB.COM (CA)
DECLARATION
We here declare that the report entitled “ON THE JOB TRAINING REPORT
ON AAVIN MILK COMPANY AT MADURAI” is a record of an original work
done by us under the guidance of Ms. M. MANJULA DEVI M.COM.,M.Phil.,
Assistant professor Department of Computer Application and this project work is
submitted in the partial fulfillment of the requirement for the award of the Degree
of Bachelor of Commerce with Computer Application. The results embodied in
this report have not been submitted to any university or institution for award of any
degree or diploma.
Signature of thecandidates
It was God who saw us this project report. The Path was zigzagged and
rough but the almighty god did not let me waver. So we thank Almighty for the
blessings showered on us.
We would like to express our special appreciation and thanks to our advisor
Ms. M. MANJULA DEVI M.Com., M.Phil.,for being a tremendous mentor for
us.
First Generation
1
Second Generation
2
The period of second generation: 1959-1965. Transistor based.
Third Generation
3
The period of third generation: 1965-1971. Integrated Circuit based.
Fourth Generation
4
The period of fourth generation: 1971-1980. VLSI microprocessor based.
Fifth Generation
5
The period of fifth generation: 1980-onwards. ULSI microprocessor
based.
S.No. Type Specifications
PC (Personal Computer)
A PC can be defined as a small, relatively inexpensive computer designed for an individual
user. PCs are based on the microprocessor technology that enables manufacturers to put an
entire CPU on one chip. Businesses use personal computers for word processing, accounting,
desktop publishing, and for running spreadsheet and database management applications. At
home, the most popular use for personal computers is playing games and surfing the Internet.
Although personal computers are designed as single-user systems, these systems are normally
linked together to form a network. In terms of power, now-a-days high-end models of the
Macintosh and PC offer the same computing power and graphics capability as low-end
workstations by Sun Microsystems, Hewlett-Packard, and Dell.
Workstation
Workstations generally come with a large, high-resolution graphics screen, large amount of
RAM, inbuilt network support, and a graphical user interface. Most workstations also have
mass storage device such as a disk drive, but a special type of workstation, called diskless
workstation, comes without a disk drive.
Common operating systems for workstations are UNIX and Windows NT. Like PC,
workstations are also single-user computers like PC but are typically linked together to form a
local-area network, although they can also be used as stand-alone systems.
Minicomputer
It is a midsize multi-processing system capable of supporting up to 250 users simultaneously.
Mainframe
Mainframe is very large in size and is an expensive computer capable of supporting hundreds or
even thousands of users simultaneously. Mainframe executes many programs concurrently and
supports many simultaneous execution of programs.
Supercomputer
Supercomputers are one of the fastest computers currently available. Supercomputers are very
expensive and are employed for specialized applications that require immense amount of
mathematical calculations (number crunching).
For example, weather forecasting, scientific simulations, (animated) graphics, fluid dynamic
calculations, nuclear energy research, electronic design, and analysis of geological data (e.g. in
petrochemical prospecting).
Previous Page
Print
Next Page
Advertisements
Functionalities of a Computer
If we look at it in a very broad sense, any digital computer carries out the
following five functions −
High Speed
Computer is a very fast device.
Accuracy
In addition to being very fast, computers are very accurate.
Computers perform all jobs with 100% accuracy provided that the input
is correct.
Storage Capability
Memory is a very important characteristic of computers.
It can store any type of data such as images, videos, text, audio, etc.
Diligence
Unlike human beings, a computer is free from monotony, tiredness, and
lack of concentration.
It can work continuously without any error and boredom.
It can perform repeated tasks with the same speed and accuracy.
Versatility
A computer is a very versatile machine.
This machine can be used to solve the problems related to various fields.
Reliability
A computer is a reliable machine.
Automation
Computer is an automatic machine.
Disadvantages of Computers
Following are certain disadvantages of computers.
No I.Q.
A computer is a machine that has no intelligence to perform any task.
Dependency
It functions as per the user’s instruction, thus it is fully dependent on
humans.
Environment
The operating environment of the computer should be dust free and
suitable.
No Feeling
Computers have no feelings or emotions.
44.
Also, the considerations are the architecture, languages, modes of operation, etc.
The function performed by the computer and the speed of their operations have
been changing since the old days to the most modern computer. Based on the
instructions were made to order for the specific task for which the computer was
In the first generation computer, vacuum tubes as CPU, magnetic drum for data
storage , and machines languages were used for giving instruction.The computer
(0s and 1s). Afterward, assembly languages were developed and used in first
generation computer.
EDSAC.
2.) Second Generation Computer (1957-1963)
binary codes.
miniature size, smaller power consumption, and heat production rate. The
Transistors were an improvement over the vacuum tube, but they still generated
a great heat, which damaged the computer's sensitive parts. The quartz
computer became ever smaller as more components were squeezed on the chip.
IC was first designed and fabricated by Jack S Kilby at Texas Instrument and by
lithographic process.
Magnetic disks began to replace magnetic tape for auxiliary and video display
terminals were introduced for the output of data. Keyboards were used for the
input of data. A new operating system was introduced for automatic processing
and multi-programming.
These computers were highly reliable, relatively expensive and faster. High-
series.
The first microprocessor called Intel 4004 was developed by American Intel
Corporation 1971.
This computer has faster generation language and application software for
microcomputers became popular and allowed home and business users to adapt
their computers for word processing, spreadsheet manipulating, file handing and
graphics.
In this generation, the concept of computer networks and CD-ROMs came into
existence.
notebook etc.
The aim is to bring machines with genuine IQ, the ability to reason logically and
with real knowledge of the word. Thus, this computer will be totally different,
totally novel and totally new than last four generations of computer.
Fifth generation computer was based on Artificial Intelligence (AI) and that is
still developing process, but not yet a reality i.e this computer is incomplete.
These computers will be able to converse with people and will be able to mimic
Technology to be used: These machines will incorporate Bio-chip and VVLSI (Very Very
Large Scale Integration) or Ultra-Large Scale Integration (ULSI)
The computer will have Artificial Intelligence (AI).
Natural language will be used to develop programs.
The computer will have parallel processing in full fledge.
The operating speed will be in terms of LIPS(Logical Inference per Second)
These aim is to solve highly complex problems, which require great intelligence and
expertise when solved by people.
Quantum computation, molecular and nanotechnology will be used completely
The first generation computers were developed during 1943-1958. It used vacuum
tubes as the active electronic components and was therefore very large. However
some of the features are as follows-
a) They were extremely large and occupied a very large space.
b) They used vacuum tubes as memory device.
c) They were very expensive and consumed a lot of electrical power.
d) The operating speed was measured in milliseconds.
e) These computers had low level of accuracy and reliability.
f) Storage capacity was too small only 1 to 4Kb.
g) They used machine level programming language.
The examples are- UNIVAC, ENIAC, EDSAC, EDVAC, and UNIVAC.
The second generation computers were developed during 1959-1965. The invention
of the transistor by three scientists of Bell Telephone Laboratories in 1947 greatly
changed the development of computers. However some of the features are as
follows-
a) These computers used transistor.
b) They were smaller, faster and cheaper than first generation of computer.
c) They consumed less electrical power than first generation.
d) The operating speed was measured in microseconds.
e) They were more reliable and accurate than the first generation computers.
f) They could understand high level language such as COBOL.
g) Magnetic tapes were used as secondary storage media.
The examples are – IBM 1620, IBM 1401, and CDC 3600.
The third generation computers were developed during 1966-1973. The
development of Integrated Circuit (IC) signaled the beginning of the third
generation computers. However some of the features are as follows-
a) These computers used integrated circuits.
b) They were small, efficient and reliable.
c) Operating systems were developed.
d) Monitors and keyboards were introduced for input and output of data.
e) Magnetic disks were used for secondary storage.
f) The operating speed was measured in nano seconds.
g) They could understand large number of high level languages.
The examples are – IBM 360, ICL -1900, and IBM 370 etc.
The fourth generation computers were developed during 1974-1990. This
generation of computer is presently in use. The development of microprocessor
signaled the beginning of the fourth generation of computers. However some of
the features are as follows-
a) These computers use LSI and VLSI technologies.
b) Its sizes were reduced to desktop and laptop computer.
c) These computers are highly reliable and accurate.
d) They have a large memory and high functional speed.
e) The operating speed is measured in beyond picoseconds and MIPS (Million of
instruction per second)
f) Magnetic disk is the common source of external storage.
g) Multiprocessing and multiprogramming OS (operating system) are used.
h) 4GL are also used.
he examples are – IBM PC, Apple/Macintosh, 8086, 80286 SX, 80386 SX etc.
he computers having artificial intelligence (AI) and high
processing capacity undergoing on the development stage from 1990 are said to be
fifth generation computers. However some of the features are as follows-
a) The speed will be extremely high in fifth generation computer.
b) These computers will be using Ultra Large Scale Integration (ULSI) technology.
) The goal of fifth generation computers is to develop machines that will be able to
think and take decisions.
Advantages
Vacuum tubes were the only electronic component
available during those days.
Vacuum tube technology made possible to make
electronic digital computers.
These computers could calculate data in millisecond.
Disadvantages
The computers were very large in size.
They consumed a large amount of energy.
tubes.
They were not very reliable.
Non-portable.
storage.
Used punch cards for input.
Advantages
Smaller in size as compared to the first generation
computers.
The 2nd generation Computers were more reliable
computers.
Better speed and could calculate data in
microseconds
Used faster peripherals like tape drives, magnetic
language.
Accuracy improved.
Disadvantages
Cooling system was required
Constant maintenance was required
Commercial production was difficult
Only used for specific purposes
Costly and not versatile
Puch cards were used for input.
Third Generation Computers (1964-1975)
Image Courtesy: www.computerhistory.org
The Third generation computers used the integrated circuits (IC). Jack Kilby
developed the concept of integrated circuit in 1958. It was an important
invention in the computer field. The first IC was invented and used in 1961.
The size of an IC is about ¼ square inch. A single IC chip may contain
thousands of transistors. The computer became smaller in size, faster, more
reliable and less expensive. The examples of third generation computers
are IBM 370, IBM System/360,UNIVAC 1108 and UNIVAC AC 9000 etc.
Advantages
Smaller in size as compared to previous generations.
More reliable.
Used less energy
Produced less heat as compared to the previous two
generations of computers.
Better speed and could calculate data in
nanoseconds.
Used fan for heat discharge to prevent damage.
Maintenance cost was low because hardware failure is
reare.
Totally general purpose
Could be used for high-level languages.
Good storage
Versatile to an extent
Less expensive
Better accuracy
Commercial production increased.
Disadvantages
Air conditioning was required.
Highly sophisticated technology required for the
manufacturing of IC chips.
Fourth Generation Computers (1975-
Present)
Advantages
More powerful and reliable than previous generations.
Small in size
Fast processing power with less power consumption
Fan for heat discharging and thus to keep cold.
No air conditioning required.
Totally general purpose
Commercial production
Less need of repair.
Cheapest among all generations
type of computers
Disadvantages
The latest technology is required for manufacturing of
Microprocessors.
Fifth Generation Computers (Present &
Beyond)
Scientists are working hard on the 5th generation computers with quite a few
breakthroughs. It is based on the technique of Artificial Intelligence
(AI). Computers can understand spoken words & imitate human reasoning.
Can respond to its surroundings using different types of sensors. Scientists
are constantly working to increase the processing power of computers. They
are trying to create a computer with real IQ with the help of advanced
programming and technologies. IBM Watson computeris one example that
outsmarts Harvard University Students. The advancement in modern
technologies will revolutionize the computer in future.
Difference Between LAN, MAN and WAN
Comparison Chart
BASIS OF
LAN MAN WAN
COMPARISON
Network
Meaning A network that connects a It covers relatively It spans large locality and
Network
maintenance
Delay
BASIS OF
LAN MAN WAN
COMPARISON
LAN or Local Area Network links network devices in such a way that
personal computer and workstations can share data, tools and programs.
Data transmits at a very fast rate as the number of computers linked are
limited. LAN’s cover smaller geographical area and are privately owned.
One can use it for an office building, home, hospital, schools, etc. LAN is
easy to design and maintain.
A Communication medium used for LAN has twisted pair cables and
coaxial cables. It covers a short distance, and so the error and noise are
minimized.
MAN or Metropolitan area Network covers a larger area than that of a LAN
and smaller area as compared to WAN. It connects two or more
computers that are apart but resides in the same or different cities. It
covers a large geographical area and may serve as an ISP (Internet
Service Provider). It’s hard to design and maintain a Metropolitan Area
Network.
There are many advantages of LAN over MAN and WAN, such as LAN’s
provide excellent reliability, high data transmission rate, they can easily
be managed, and shares peripheral devices too. Local Area Network
cannot cover cities or towns and for that Metropolitan Area Network is
needed, which can connect city or a group of cities together. Further, for
connecting Country or a group of Countries one requires Wide Area
Network.
ninjacraze
more
Contact Author
CC BY 2.0 | Source
Just as humans use language to communicate, and different regions have different languages,
computers also have their own languages that are specific to them.
Different kinds of languages have been developed to perform different types of work on the
computer. Basically, languages can be divided into two categories according to how the
computer understands them.
Low-Level Languages
Low-level computer languages are either machine codes or are very close them. A computer
cannot understand instructions given to it in high-level languages or in English. It can only
understand and execute instructions given in the form of machine language i.e. binary. There
are two types of low-level languages:
Machine Language: a language that is directly interpreted into the hardware
Assembly Language: a slightly more user-friendly language that directly corresponds to
machine language
Machine Language
Machine language is the lowest and most elementary level of programming language and was
the first type of programming language to be developed. Machine language is basically the only
language that a computer can understand and it is usually written in hex.
In fact, a manufacturer designs a computer to obey just one language, its machine code, which
is represented inside the computer by a string of binary digits (bits) 0 and 1. The symbol 0
stands for the absence of an electric pulse and the 1 stands for the presence of an electric
pulse. Since a computer is capable of recognizing electric signals, it understands machine
language.
Advantages Disadvantages
It requires no translator to
translate the code. It is All memory addresses have
directly understood by the to be remembered.
computer.
Assembly Language
Assembly language was developed to overcome some of the many inconveniences of machine
language. This is another low-level but very important language in which operation codes and
operands are given in the form of alphanumeric symbols instead of 0’s and l’s.
These alphanumeric symbols are known as mnemonic codes and can combine in a maximum
of five-letter combinations e.g. ADD for addition, SUB for subtraction, START, LABEL etc.
Because of this feature, assembly language is also known as ‘Symbolic Programming
Language.'
This language is also very difficult and needs a lot of practice to master it because there is only
a little English support in this language. Mostly assembly language is used to help in compiler
orientations. The instructions of the assembly language are converted to machine codes by a
language translator and then they are executed by the computer.
Advantages Disadvantages
Assembly language is
Like machine language, it is
easier to understand and
also machine
use as compared to
dependent/specific.
machine language.
Since it is machine
It is easy to locate and dependent, the programmer
correct errors. also needs to understand
the hardware.
It is easily modified.
High-Level Languages
High-level computer languages use formats that are similar to English. The purpose of
developing high-level languages was to enable people to write programs easily, in their own
native language environment (English).
High-level languages are basically symbolic languages that use English words and/or
mathematical symbols rather than mnemonic codes. Each instruction in the high-level language
is translated into many machine language instructions that the computer can understand.
Advantages Disadvantages
time
A program written in a
high-level language can be
translated into many
machine languages and
can run on any computer
for which there exists an
appropriate translator
The language is
independent of the
machine on which it is
used i.e. programs
developed in a high-level
language can be run on
any computer text
It is very difficult to guess or know the time required by any job to complete. Processors of the batch
systems knows how long the job would be when it is in queue
Multiple users can share the batch systems
The idle time batch system is very less
It is easy to manage large work repeatedly in batch systems
Disadvantages of Batch Operating System:
The computer operators should be well known with batch systems
Batch systems are hard to debug
It is sometime costly
The other jobs will have to wait for an unknown time if any job fails
Examples of Batch based Operating System: Payroll System, Bank Statements etc.
2. Time-Sharing Operating Systems –
Each task has given some time to execute, so that all the tasks work smoothly. Each user gets time of CPU as
they use single system. These systems are also known as Multitasking Systems. The task can be from single
user or from different users also. The time that each task gets to execute is called quantum. After this time
interval is over OS switches over to next task.
Advantages of RTOS:
Maximum Consumption: Maximum utilization of devices and system,thus more output from all the
resources
Task Shifting: Time assigned for shifting tasks in these systems are very less. For example in older
systems it takes about 10 micro seconds in shifting one task to another and in latest systems it takes 3
micro seconds.
Focus on Application: Focus on running applications and less importance to applications which are in
queue.
Real time operating system in embedded system: Since size of programs are small, RTOS can also
be used in embedded systems like in transport and others.
Error Free: These types of systems are error free.
Memory Allocation: Memory allocation is best managed in these type of systems.
Disadvantages of RTOS:
Limited Tasks: Very few task run at the same time and their concentration is very less on few
applications to avoid errors.
Use heavy system resources: Sometimes the system resources are not so good and they are expensive
as well.
Complex Algorithms: The algorithms are very complex and difficult for the designer to write on.
Device driver and interrupt signals: It needs specific device drivers and interrupt signals to
response earliest to interrupts.
Thread Priority: It is not good to set thread priority as these systems are very less pron to switching
tasks.
Examples of Real-Time Operating Systems are: Scientific experiments, medical imaging systems,
industrial control systems, weapon systems, robots, air traffic control systems, etc.
Answer: Following are the differences between multiprocessing and
multiprogramming.
4 Less time taken to process More Time taken to process the jobs.
the jobs.
WhatIs.com
Follow:
Advantages:
• Solve bigger problems
• Run more problems in shorter time
• May save money
• Allows for virtual testing
• Can be used for R&D
Disadvantages:
• Can be expensive
• Takes up a lot of space
• May only be good for specific applications
• Does not replace physical testing
• Requires trained staff
NETWORK
There are so many different types of computer networks in existence, it can be hard to understand the
differences between them, particularly the ones with very similar-sounding names. This lesson
explains the structures and functions of some of the most popular computer networks.
Types of Networks
There are several different types of computer networks. Computer networks can be characterized by
their size as well as their purpose.
The size of a network can be expressed by the geographic area they occupy and the number of
computers that are part of the network. Networks can cover anything from a handful of devices
within a single room to millions of devices spread across the entire globe.
Some of the different networks based on size are:
In terms of purpose, many networks can be considered general purpose, which means they are used
for everything from sending files to a printer to accessing the Internet. Some types of networks,
however, serve a very particular purpose. Some of the different networks based on their main
purpose are:
Send a document to the printer in the office upstairs while you are sitting on the couch with
your laptop.
Upload a photo from your cell phone to your desktop computer.
Watch movies from an online streaming service to your TV.
If this sounds familiar to you, you likely have a PAN in your house without having called it by its
name.
Lower in cost
Sharing of Resources
High Speed
Security
Disadvantages
Disadvantages
Disadvantages
Conclusion
There are many advantages of LAN over MAN and WAN, such as
LAN’s provide excellent reliability, high data transmission rate,
they can easily be managed, and shares peripheral devices too.
Local Area Network cannot cover cities or towns and for that
Metropolitan Area Network is needed, which can connect city or a
group of cities together. Further, for connecting Country or a
group of Countries one requires Wide Area Network.
Multiple jobs are executed by the CPU by switching between them, but the
switches occur so frequently. Thus, the user can receive an immediate
response. For example, in a transaction processing, the processor executes
each user program in a short burst or quantum of computation. That is,
if n users are present, then each user can get a time quantum. When the
user submits the command, the response time is in few seconds at most.
The operating system uses CPU scheduling and multiprogramming to
provide each user with a small portion of a time. Computer systems that
were designed primarily as batch systems have been modified to time-
sharing systems.
Problem of reliability.
Multitasking is the ability of an operating system to execute more than one task
simultaneously on a single processor machine. Though we say so but in reality
no two tasks on a single processor machine can be executed at the same time.
Actually CPU switches from one task to the next task so quickly that appears as
if all the tasks are executing at the same time. More than one
task/program/job/process can reside into the same CPU at one point of time.
Multitasking System
Multiprocessing
Multiprocessing is the ability of an operating system to execute more than one
process simultaneously on a multi processor machine. In this, a computer uses
more than one CPU at a time.
Multiprocessing System
Multithreading
Multithreading is the ability of an operating system to execute the different
parts of a program called threads at the same time. Threads are the light wait
processes which are independent part of a process or program. In
multithreading system, more than one threads are executed parallely on a
single CPU.
Definition:
This is any operating system which is capable of running more than one program at a time.
Most modern machines are designed to be able to handle various programs which can work
together simultaneously (apart from various specifically developed machines which are designed for
use in specific businesses.
As with most technology today, multi tasking
operating systems are constantly evolving, and will
continue to change over time.
Features:
The main advantage for the user is that one has the ability to have several applications open at the
same time.
For example, you could be editing a file on one application while another application is recalculating
spreadsheet, or even downloading and installing a new application.
The an application developer, the advantage of multitasking is the ability to create applications that
use more than one process and to create processes that use more than one thread of execution.
For example, a process can have a user interface thread that manages interactions with the user
(keyboard and mouse input), and worker threads that perform other tasks while the user interface
thread waits for user input. If you give the user interface thread a higher priority, the application will
be more responsive to the user, while the worker threads use the processor efficiently during the
times when there is no user input.
The main use of a multi tasking operating system is to have the ability to basically use many different
applications on the computer at the same time.
This is something that had previously not been able to be done. but due to technology advancing, it
is now easy to use different programs at one time without running down your system.
WHAT IS
HARDWARE AND
SOFTWARE?
Home » Learn Technology » What is hardware and Software?
A computer, also called PC, is a device that allows us to process information and calculate
results,of course that this classic definition is somewhat outdated, as it is currently very wide range
of tasks that can be performed with this device.
Currently with a computer, we can draw, write, play and communicate with others around the
world without matter the distance, and being faithful to the original definition, we can process
information using pure data through software programs.
Are hundreds of thousands the applications that we can install on a computer, each user can choose
the necessary software that best fits your needs and requirements for each task begins.
What is a computer?
Basically a computer consists of two parts: hardware and software.
The hardware is a generic term used to designate all physical component parts of a
computer,namely the cabinet, monitor, motherboard, RAM, processor, video card and others.
In the other hand, the software is the code in form of program responsible for providing the
necessary instructions to the computer to perform a certain action. This category includes the
operating system, games, device drivers, etc.
What is the hardware?
As well as the computer is divided into hardware and software, the hardware is also divided into
two categories, called basic hardware and complementary hardware.
While the basic hardware includes all the necessary components for the operation of the
computer, such as motherboard, monitor, keyboard and mouse, being in this case the
motherboard, RAM and CPU the most important components of the whole.
On the other hand we have the complementary hardware, hardware that is not needed for a
computer to function normally. Examples of such hardware are printers, scanners, flash drives and
others.
Another way to differentiate the type of computer hardware is based on the role within it. In
the following paragraphs we will know how this hardware are divided.
Input Devices: Devices that allow the entry of information to the computer.
Chipset: Chip integrated on the motherboard, the chipset allows traffic information between the
microprocessor and other components that make up the PC.
CPU: This device has the task of interpreting and executing instructions sent from the operating
system, and manages the requests made by users through the programs.
RAM: The RAM, acronym of Random Access Memory, is where the computer stores the
programs, data and results processed by the CPU, allowing quick retrieval by the system, providing
better performance than retrieving it directly from the hard disk.
Storage drives: This devices are intended for storing data both the user and the system. I.e. they are
the physical space where are stored the files of the operating system, programs and user
documents. These can be hard disks, CD, DVD or Blu Ray, flash drives or external removable
drives.
Output Devices: These are the devices that show the user the results of its operations. This group
includes monitors, printers and everything that serves the purpose of offering the user the
ability to see the result of their work.
Mixed devices: This group includes all those devices that are capable of receiving and delivering
information. Examples of these devices are network cards, modems and communication
ports,among others.
What is the software?
The software is a computer code written in a programming language and compiled in the form
of application that you use to perform a particular task. Also as in the case of hardware, the
software should be divided into categories based on the role within the system.
Software Concepts
The software basically is a sequence of instructions that are interpreted and / or implemented to
manage, redirect or modify a data / information or event.
Software is also a product which is developed by software engineering, and includes not only the
program for your computer, but also manuals and technical documentation. Currently, these
manuals generally are provided in PDF format.
Computer software applications comprises a sequence of instructions that is interpreted and executed
by a processor or a virtual machine. In a functional software, that sequence follows specific
standards that result in a certain behavior.
In our computer, all applications we use are software: Windows, Outlook, Firefox, Internet Explorer,
Google Chrome, and even the antivirus apps.
A software can be executed by any device capable of interpreting and executing instructions for
which it is created. When software is represented as instructions that can be executed directly by a
processor, we say that is written in machine language.
The execution of a software can also be performed through an interpreter, responsible program to
understand and execute each of its instructions. A special category of interpreters and remarkable
are the virtual machines such as the Java (JVM) virtual machine.
The best-known device that has a processor is the computer. Nevertheless, there are other
programmable machines such as mobile phones, industrial automation machines, etc.
Software Rating
System Software: This group includes the operating system, device drivers, system utilities and all
the tools necessary to control the specific characteristics of these computers.
Software Implementation: It is called application software to all programs used by the user to
perform a particular task. This group includes the office software, graphic design, accounting and
electronic, to cite just one small fraction of all categories of applications that we can find in the
market.
Comparison chart
Hardware versus Software comparison chart
Hardware Software
Definition Devices that are required to store and execute (or Collection of instructions that enables a user to
run) the software. interact with the computer. Software is a
program that enables a computer to perform a
specific task, as opposed to the physical
components of the system (hardware).
Types Input, storage, processing, control, and output System software, Programming software, and
devices. Application software.
Function Hardware serve as the delivery system for software To perform the specific task you need to
solutions. The hardware of a computer is complete. Software is generally not needed to
infrequently changed, in comparison with software for the hardware to perform its basic level tasks
and data, which are “soft” in the sense that they are such as turning on and reponding to input.
readily created, modified, or erased on the comput
Examples CD-ROM, monitor, printer, video card, scanners , QuickBooks, Adobe Acrobat, Google Chrome,
label makers, routers and modems. Microsoft Word, Microsoft Excel, Apple Maps
Inter Hardware starts functioning once software is To deliver its set of instructions, Software is
dependency loaded. installed on hardware.
Failure Hardware failure is random. Hardware does have Software failure is systematic. Software does
increasing failure at the last stage. not have an increasing failure rate.
Durability Hardware wears out over time. Software does not wear out over time.
However, bugs are discovered in software as
time passes.
a. Batch Processing
The time delay between the collection of data and getting the result after the
batch process.
In the batch processing master file is not always kept up to date.
Here, a one-time process can be very slow.
Batch vs Real Time Processing
b. Real-Time Processing
3. Conclusion
As a result, we have seen a comparison between Batch Processing vs Real-Time
processing in spark in detail. Hence, making a decision of selecting method depends
on the current business system. Basically, there are various conditions on which it
depends, whether to use one over the other. For example, type & volume of data
and time that the data needs to be processed. Thus, select the one that best suits
your business system. Hope we have answered all the
questions regarding Batch Processing vs Real Time Processing.
For reference.
Transaction processing system
From Wikipedia, the free encyclopedia
Transaction processing is a way of computing that divides work into individual, indivisible operations,
called transactions.[1] A transaction processing system (TPS) is a software system, or
software/hardware combination, that supports transaction processing.
Contents
1History
2List of transaction processing systems
3Processing types
o 3.1Batch processing
o 3.2Real-time processing
o 3.3Transaction processing
4Transaction processing system features
o 4.1Performance
o 4.2Continuous availability
o 4.3Data integrity
o 4.4Ease of use
o 4.5Modular growth
5Types of transaction processing
o 5.1Processing in a batch
o 5.2Processing in real-time
6Databases for transaction processing
7Backup procedures
o 7.1Recovery process
o 7.2Types of back-up procedures
7.2.1Grandfather-father-son
7.2.2Backup plus journal
8See also
9References
10Further reading
History[edit]
The first transaction processing system was SABRE, made by IBM for American Airlines, which
became operational in 1970. Designed to process up to 83,000 transactions a day, the system ran
on two IBM 7090 computers. SABRE was migrated to IBM System/360 computers in 1972, and
became an IBM product first as Airline control Program (ACP) and later as Transaction Processing
Facility (TPF). In addition to airlines TPF is used by large banks, credit card companies, and hotel
chains.
The Hewlett-Packard NonStop system (formerly Tandem NonStop) was a hardware and software
system designed for Online Transaction Processing (OLTP) introduced in 1976. The systems were
designed for transaction processing and provided an extreme level of availability and data integrity.
Processing types[edit]
Transaction processing is distinct from other computer processing models – batch processing, time-
sharing, and real-time processing.[9]
Batch processing[edit]
Main article: Batch processing
"Real time systems attempt to guarantee an appropriate response to a stimulus or request quickly
enough to affect the conditions that caused the stimulus."[9] Each transaction in real-time processing
is unique; it is not part of a group of transactions.
Transaction processing[edit]
Main article: Transaction processing
A Transaction Processing System (TPS) is a type of information system that collects, stores,
modifies and retrieves the data transactions of an enterprise. Transaction processing systems also
attempt to provide predictable response times to requests, although this is not as critical as for real-
time systems. Rather than allowing the user to run arbitrary programs as time-sharing, transaction
processing allows only predefined, structured transactions. Each transaction is usually short duration
and the processing activity for each transaction is programmed in advance.
A database is an organized collection of data. Databases offer fast retrieval times for non-structured
requests as in a typical transaction processing application.
Databases for transaction processing may be constructed using hierarchical, network, or relational
structures.
A hierarchical structure.
A network structure.
The following features are desirable in a database system used in transaction processing systems:
Good data placement: The database should be designed to access patterns of data from many
simultaneous users.
Short transactions: Short transactions enables quick processing. This avoids concurrency and
paces the systems.
Real-time backup: Backup should be scheduled between low times of activity to prevent lag of
the server.
High normalization: This lowers redundant information to increase the speed and improve
concurrency, this also improves backups.
Archiving of historical data: Uncommonly used data are moved into other databases or
backed up tables. This keeps tables small and also improves backup times.
Good hardware configuration: Hardware must be able to handle many users and provide
quick response times.
Backup procedures[edit]
Main article: Backup
Journal: A journal maintains an audit trail of transactions and database changes. Transaction
logs and Database change logs are used, a transaction log records all the essential data for
each transactions, including data values, time of transaction and terminal number. A database
change log contains before and after copies of records that have been modified by transactions.
Checkpoint: The purpose of checkpointing is to provide a snapshot of the data within the
database. A checkpoint, in general, is any identifier or other reference that identifies the state of
the database at a point in time. Modifications to database pages are performed in memory and
are not necessarily written to disk after every update. Therefore, periodically, the database
system must perform a checkpoint to write these updates which are held in-memory to the
storage disk. Writing these updates to storage disk creates a point in time in which the database
system can apply changes contained in a transaction log during recovery after an unexpected
shut down or crash of the database system. If a checkpoint is interrupted and a recovery is
required, then the database system must start recovery from a previous successful
checkpoint. Checkpointing can be either transaction-consistent or non-transaction-
consistent (called also fuzzy checkpointing). Transaction-consistent checkpointing produces a
persistent database image that is sufficient to recover the database to the state that was
externally perceived at the moment of starting the checkpointing. A non-transaction-consistent
checkpointing results in a persistent database image that is insufficient to perform a recovery of
the database state. To perform the database recovery, additional information is needed, typically
contained in transaction logs. Transaction consistent checkpointing refers to a consistent
database, which doesn't necessarily include all the latest committed transactions, but all
modifications made by transactions, that were committed at the time checkpoint creation was
started, are fully present. A non-consistent transaction refers to a checkpoint which is not
necessarily a consistent database, and can't be recovered to one without all log records
generated for open transactions included in the checkpoint. Depending on the type of database
management system implemented a checkpoint may incorporate indexes or storage pages (user
data), indexes and storage pages. If no indexes are incorporated into the checkpoint, indexes
must be created when the database is restored from the checkpoint image.
Recovery Manager: A recovery manager is a program which restores the database to a correct
condition which allows transaction processing to be restarted.
Depending on how the system failed, there can be two different recovery procedures used.
Generally, the procedures involves restoring data that has been collected from a backup device and
then running the transaction processing again. Two types of recovery are backward
recovery and forward recovery:
Backward recovery: used to undo unwanted changes to the database. It reverses the changes
made by transactions which have been aborted.
Forward recovery: it starts with a backup copy of the database. The transaction will then
reprocess according to the transaction journal that occurred between the time the backup was
made and the present time.
See also: Checkpoint restart
I believe some of the main advantages and disadvantages or TPSs are as follows:
Advantages
o Can handle several thousand operations at once
o Access to untapped markets. Allows people across the world to access and buy a business’s
goods and services
Disadvantages
o Complex system may not be large enough to utilize TPS
o Security and hardware problems
o Perhaps too much consolidation is required