HISTORY OF DATA PROCESSING
Download
Report
Transcript HISTORY OF DATA PROCESSING
Evolution of Computing
1.1
Quit
First
Computers
PCs
introduced!
1960s
1950
Very
expensive
computers
for large
companies.
Computer
professionals
ran the show.
2001
mid1970s
Powerful PCs
on every
desktop.
Explosion of
applications.
PREHISTORY OF DATA PROCESSING
1.2
Quit
Calculating
devices:
pebbles, cowrie shells, bamboo sticks
'chou'
calculation tables (Medieval Europe)
Abacus and its counterparts in China
and Japan
PREHISTORY OF DATA PROCESSING
1.3
Quit
Computing
machines were designed
based on the technology of gears:
Blaise Pascal, 17th century,
permanent algorithm
G.W. Leibniz, 17th cent., a choice of
built-in algorithms
Charles Babbage, 19th century:
programmable "Analytical Engine"
PREHISTORY OF DATA PROCESSING
1.4
Quit
Jacquard's
loom 1801: weaving
loom was controlled by paper cards
with holes in them.
PREHISTORY OF DATA PROCESSING
1.5
Quit
Hollerith:
punched card machines
for US census 1890,
electromechanical, start of
calculation machines and IBM
addition and counting
1928 subtraction
1931
multiplication
HISTORY OF DATA PROCESSING :1940s
1.6
Quit
ENIAC:
first modern computer, vacuum
tubes
hexadecimal and binary systems
cryptography during World War II
Von Neumann architecture:
both commands and numerical data were
expressed and processed in the same form
and in the same devices, step by step
General purpose computer
technologies:
electro-mechanical relays, punched tape
magnetic drum (max capacity 500 characters
or 50 numbers)
HISTORY OF DATA PROCESSING :1950s
1.7
Quit
magnetic tape storage
ferromagnetic cores in central memory
stand-alone computers
1956
first computer made of transistors
magnetic disks
HISTORY OF DATA PROCESSING :1960s
1.8
Quit
1960
SSI: small scale integrated
circuit (100 transistors)
1966 MSI: medium scale integrated
circuit (1,000 transistors)
1969 LSI: large scale integration (up
to 10,000 components)
VLSI and so on
Increase in performance
Quit
Speed
Accuracy
Consistency
Reliability
Communications
Memory
HISTORY OF DATA PROCESSING: software
1.10
Quit
1950s
Machine
and Assembler language
programming
1960s
Multitasking operating systems, timesharing, batch-processing
Mainframe computers: terminals, on-line
use
Statistics, calculations, accounting
Cobol, Fortran, Algol programming
languages
HISTORY OF DATA PROCESSING : 1970s
1.11
Quit
minicomputers,
distributed systems,
networks
Basic, APL, Pascal and other
specialized programming languages
databases
division of labor:
operators, systems analysts, data
analysts, data entry clerks,
programmers, output handlers
Supercomputers
Quit
1.12
Processor-bound
applications
100 times faster than typical server
Used for
Scientific simulations and forecasting
Medicine
Advanced graphics
Supercomputers
HISTORY OF DATA PROCESSING : 1980s
Quit
1.13
microcomputers
IBM PC 1981
mainframes
and minicomputers,
workstations
http://news.bbc.co.uk/2/hi/technology/4780963.stm
HISTORY OF DATA PROCESSING : 1980s
1.14
Quit
Office
automation: text processing and
spreadsheets
Graphical user interfaces
Robotics
C programming, client /server databases,
SQL and relational model
Graphics, CAD, CAE, color printing
Integration of use, end-users, from
programming to applications
Client/Server Computing
1.15
Quit
Server
PC to Supercomputer
Data storage and applications software
Client
PC, workstation or thin client
Requests processing or support from server
Applications
Server
Computers
Software
Front-end (client)
Back-end (server)
Enterprise-wide
Systems
Proxy server computer
Cloud computing
HISTORY OF DATA PROCESSING : 1990s
1.16
Quit
LANs
(local area networks), electronic
mail
WANs (wide area networks), global
integration of systems
high-power workstations, GUIs (Graphical
user interface) standard
object-oriented model, C++ and Java
integration of applications
multimedia
Information deluge
Quit
1.17
US National Security Agency intercepts and
stores 1.7 billion e-mails, messages and phone
calls a day
decoding the human genome involves analysing
3 billion base pairs, which took ten years the first
time it was done, in 2003, but can now be
achieved in one week
WalMart, a retail giant, handles more than 1m
customer transactions every hour, feeding
databases estimated at more than 2.5 petabytes
Facebook is home to 40 billion photos.
"It is not surprising that people feel overwhelmed. There is
an immense risk of cognitive overload"
Data volumes
Quit
1.18
4,6 billion mobile phone subscriptions worldwide (the
world has 6,8 billion people)
1 billion to 2 billion people use the internet.
By 2013 the amount of traffic flowing over the internet
annually will reach 667 exabytes, according to Cisco, a
maker of communications gear.
According to a 2008 study by International Data Corp
(IDC), a market research firm, around 1 200 exabytes of
digital data will be generated this year.
Google handles around half the world’s internet
searches, answering around 35 000 queries every
second; data mining:
Google translation
spread of flu epidemics.
Source: The Economist 2010
Units of data
Unit
Bit
b
Byte
B
Kilobyte KB
Megabyte MB
Size
1 or 0
8 bits
1000 or 210 bytes
1000 KB; 220 bytes
Gigabyte GB
Terabyte TB
230 bytes
240 bytes
Petabyte PB
250 bytes
Exabyte (EB) 260 bytes
Zettabyte (ZB) 270 bytes
Yottabyte (YB) 280 bytes
What it means
short for “binary digit”
the basic unit of computing
from “thousand” in Greek
from “large” in Greek
a typical pop song is about 4MB
from “giant” in Greek.
from “monster” in Greek.
all the catalogued books in
America’s Library of Congress total 15TB
Google processes around 1PB
every hour
the total amount of information in
existence this year is forecast to be
around 1.2ZB
FUTURE OF DATA PROCESSING ?
1.20
Quit
Embedded systems
Robotics: industry & entertainment
GPRS and mobile use
Telemedicine
Intelligent buildings
RFID (radio frequency identification chips)
Global sharing
GPS and navigation
Storage capacity
Music, entertainment
Life on net
Intuitive user interfaces, touch
control, voice