TiredService

Download Report

Transcript TiredService

Web Server Support
for Tired Services
Telecommunication Management Lab
M.G. Choi
Content
1. Introduction
2. Servers and QoS
3. Architecture for Server Based QoS
4. Prototype
5. Result
6. Summary
Introduction




The Network QoS has been focused for Performance
The benefits of Non-isochronous applications such as
web pages from QoS
Example ; North America’s EC traffic increase
The necessity of Server QoS
• The mechanisms and policies are need for establishing and
Supporting QoS. Network QoS is not sufficient to support
end to end QoS

The hypothesis that servers become an important
element in delivering QoS
Introduction (Cont’d)

The research Question to be answered
•
•
•
•
•


Impact of Internet Workloads on servers.
Impact of server latency on end to end latency.
Server mechanism to improve Quality of Service
The way to protect server from overload
The way for server to support tiered user service level
with unique performance attribute
The Importance of requirement on servers and
networks
Show the increasing role of the servers to provide
end to end and the potentiality of tired services
Server and QoS

Empirical Study
• Instrument and monitor one of large ISPs in North
America in 1997
• Quantifying the delay components for web, news and
mail server
• The Network Typology [figure 1]
• A mixture of active and passive measurement
technique
• The nntp server response time [figure 2]
• The coast to coast network response time [figure3]
Trends affecting the complex E-commerce
Applications

The trends increasing network performances
• The decreasing Network latency due to increasing
capacity of network backbone
• The guaranteed network latency by the ISP
• The caches becomes more pervasive

The trends increasing Server latency time
• The Flash
• The new application technologies[JAVA, SSL, DB, M/W]
• The Media with much richer, larger, more image [Audio
Voice, Video]
The Overload Server Causing
poor end to end QoS



The Measurement of busy Web Sites
The response rate grows linearly until the server
nears maximum capacity in terms of HTTP request
[figure 4]
The HTTP and a User transaction [figure 5]
Over-Provision Server






The evolution of web applications grows very steeply in
the client demand curve
Now, Internet Applications have unaccountable client
population
No reasonable amount of H/W can guarantee
predictable performance for flash crowds
The over-provisioning of servers can not provide tired
services or application
Network QoS can not solve scheduling or bottleneck
problems at Server and ignored by server FIFO
The Server QoS mechanism supports tired service, and
to provide overload protection
Architecture for Server Based QoS



An architecture Servers consisting of multiple node
web, application and database
The philosophy to create a low overhead, scalable
infrastructure that is transparent to application,
web servers
The two goals to support two key capabilities
• The architecture manages effectively peaks in client
HTTP request rates
• To support tiered service levels that enable preferential
treatment of users or services (to improve performance
of premium tires)



The architecture to be presented [Figure 6]
The request class is introduced for tired service
The architecture supports integration with network
QoS mechanisms and management systems
Architecture for Server Based QoS
Related Work and Prototype

Related Work
• The operating systems control mechanism to ensure
class-based performance in web servers
• The scheduling of web server worker processes with the
same priority
• The research of intelligent switch or router

Prototype
• Modifying the FIFO servicing model of Apache
Ver.1.2.4.
• The identical worker processes that listen on a UNIX
socket for HTTP connections and serve requests
• The connection manager, request classifier,
admission controller, request scheduler and
resource scheduler
Connection Manager



A new unique acceptor process that intercepts all
requests
Classifying the request and Placing the request on
the appropriate tier queue
The connection manager must run frequently
enough to keep request queues full
• Worker processes may execute requests from lower
tires
• Premium requests are prohibited from establishing
a TCP connection and thus drop
Request Classification



To identify and classify the incoming requests of
each class
The classification mechanism are user class based
or target-class based
The User Class Based (Source of Request)
• Client IP address
• HTTP cookie
• Browser plug-ins

The Target Based
• URL Request type or file name path
• Destination IP address
Admission Control
`
When the server is saturated, the Admission
Control Rule (Premium > Basic)
 The two admission control trigger parameter

• Total Requests Queued
• Number of Premium Requests Queued

Rejection is done by simply closing the
connection
Request and Resource Scheduling



To process request, selection of requests is based on
the scheduling policy
The scheduling policy may have many options for
processing
Several Potential Policies
•
•
•
•
•

Strict Priority
Weighted Priority
Shared Capacity
Fixed Capacity
Earliest Deadline First
Resource Scheduling provide more resources to
premium request and less resources to basic request
Apache Source Modification

The number of Apache code changes is minimal

http_main.c modification
• Start the connection manager process, setup queues
• Change the child Apache process to accept request from
connection manager not HTTP socket

Additional connection_mgr.c is linked
• The classification policy, enqueue mechanism, dequeue
policy, connection manager process code

Additional shared memory and semaphores
• The state of queues, each class queue length, number of
requests executing in class
• last class to have a request dequeued, total count of waiting
requests on classes
• Access to shared memory is synchronized through
Semaphore
Results


The comparison of response time, throughput, error
rate for premium and basic clients with priority
scheduling
The comparison of performance in premium and in
basic clients
• The premium rate is fixed
• The premium request rate identical to the basic client



The quality of service in premium clients is better than
in basic clients in above both case
The four Clients, The one Server, 100 based Network
The httperf application is used by four clients
Summary

Contribution
• To motivate the need for Serer QoS to support tired user
service level
• To protect servers from client demand overload
• To develop architecture of WebQoS
• To show the benefit of architecture through experiment

The unsolved problems
• The tighter integration of server, network QoS and the ability
to communicate QoS attribute across network
• More Flexible admission control mechanisms
• Lightweight signalling mechanisms for high priority traffic
• What benefits can be obtained by the end to end QoS
Critique

The Strong points
• To show the Web bottleneck is the server side not network
• To present the architecture performing the differentiated
Service and verify the availability of the architecture
• To combine other differentiated service approaches through
showing architecture

The weak points
• The architecture presented in this paper may be deeply
influenced by the status of the connection manager and may
be bottleneck
• The experiment that is similar to real environment does not
provide the effectiveness of the architecture presented
enough.