The 2to3 tool converted socketserver (which I manually fixed up with an import fallback) and the print(e) line. The xrange option was converted to range, but it seems better to use the '*' operator here for simplicity. Signed-off-by: Peter Wu <peter@lekensteyn.nl>
		
			
				
	
	
		
			442 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable File
		
	
	
	
	
			
		
		
	
	
			442 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			Python
		
	
	
		
			Executable File
		
	
	
	
	
#!/usr/bin/python
 | 
						|
 | 
						|
# Copyright 2012 Google Inc. All Rights Reserved.
 | 
						|
#
 | 
						|
# Licensed under the Apache License, Version 2.0 (the "License");
 | 
						|
# you may not use this file except in compliance with the License.
 | 
						|
# You may obtain a copy of the License at
 | 
						|
#
 | 
						|
#    http://www.apache.org/licenses/LICENSE-2.0
 | 
						|
#
 | 
						|
# Unless required by applicable law or agreed to in writing, software
 | 
						|
# distributed under the License is distributed on an "AS IS" BASIS,
 | 
						|
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
						|
# See the License for the specific language governing permissions and
 | 
						|
# limitations under the License.
 | 
						|
#
 | 
						|
# Modified by Linus Nielsen Feltzing for inclusion in the libcurl test
 | 
						|
# framework
 | 
						|
#
 | 
						|
try:
 | 
						|
    import socketserver
 | 
						|
except:
 | 
						|
    import SocketServer as socketserver
 | 
						|
import argparse
 | 
						|
import re
 | 
						|
import select
 | 
						|
import socket
 | 
						|
import time
 | 
						|
import pprint
 | 
						|
import os
 | 
						|
 | 
						|
INFO_MESSAGE = '''
 | 
						|
This is a test server to test the libcurl pipelining functionality.
 | 
						|
It is a modified version if Google's HTTP pipelining test server. More
 | 
						|
information can be found here:
 | 
						|
 | 
						|
http://dev.chromium.org/developers/design-documents/network-stack/http-pipelining
 | 
						|
 | 
						|
Source code can be found here:
 | 
						|
 | 
						|
http://code.google.com/p/http-pipelining-test/
 | 
						|
'''
 | 
						|
MAX_REQUEST_SIZE = 1024  # bytes
 | 
						|
MIN_POLL_TIME = 0.01  # seconds. Minimum time to poll, in order to prevent
 | 
						|
                      # excessive looping because Python refuses to poll for
 | 
						|
                      # small timeouts.
 | 
						|
SEND_BUFFER_TIME = 0.5  # seconds
 | 
						|
TIMEOUT = 30  # seconds
 | 
						|
 | 
						|
 | 
						|
class Error(Exception):
 | 
						|
  pass
 | 
						|
 | 
						|
 | 
						|
class RequestTooLargeError(Error):
 | 
						|
  pass
 | 
						|
 | 
						|
 | 
						|
class ServeIndexError(Error):
 | 
						|
  pass
 | 
						|
 | 
						|
 | 
						|
class UnexpectedMethodError(Error):
 | 
						|
  pass
 | 
						|
 | 
						|
 | 
						|
class RequestParser(object):
 | 
						|
  """Parses an input buffer looking for HTTP GET requests."""
 | 
						|
 | 
						|
  global logfile
 | 
						|
 | 
						|
  LOOKING_FOR_GET = 1
 | 
						|
  READING_HEADERS = 2
 | 
						|
 | 
						|
  HEADER_RE = re.compile('([^:]+):(.*)\n')
 | 
						|
  REQUEST_RE = re.compile('([^ ]+) ([^ ]+) HTTP/(\d+)\.(\d+)\n')
 | 
						|
 | 
						|
  def __init__(self):
 | 
						|
    """Initializer."""
 | 
						|
    self._buffer = ""
 | 
						|
    self._pending_headers = {}
 | 
						|
    self._pending_request = ""
 | 
						|
    self._state = self.LOOKING_FOR_GET
 | 
						|
    self._were_all_requests_http_1_1 = True
 | 
						|
    self._valid_requests = []
 | 
						|
 | 
						|
  def ParseAdditionalData(self, data):
 | 
						|
    """Finds HTTP requests in |data|.
 | 
						|
 | 
						|
    Args:
 | 
						|
      data: (String) Newly received input data from the socket.
 | 
						|
 | 
						|
    Returns:
 | 
						|
      (List of Tuples)
 | 
						|
        (String) The request path.
 | 
						|
        (Map of String to String) The header name and value.
 | 
						|
 | 
						|
    Raises:
 | 
						|
      RequestTooLargeError: If the request exceeds MAX_REQUEST_SIZE.
 | 
						|
      UnexpectedMethodError: On a non-GET method.
 | 
						|
      Error: On a programming error.
 | 
						|
    """
 | 
						|
    logfile = open('log/server.input', 'a')
 | 
						|
    logfile.write(data)
 | 
						|
    logfile.close()
 | 
						|
    self._buffer += data.replace('\r', '')
 | 
						|
    should_continue_parsing = True
 | 
						|
    while should_continue_parsing:
 | 
						|
      if self._state == self.LOOKING_FOR_GET:
 | 
						|
        should_continue_parsing = self._DoLookForGet()
 | 
						|
      elif self._state == self.READING_HEADERS:
 | 
						|
        should_continue_parsing = self._DoReadHeader()
 | 
						|
      else:
 | 
						|
        raise Error('Unexpected state: ' + self._state)
 | 
						|
    if len(self._buffer) > MAX_REQUEST_SIZE:
 | 
						|
      raise RequestTooLargeError(
 | 
						|
          'Request is at least %d bytes' % len(self._buffer))
 | 
						|
    valid_requests = self._valid_requests
 | 
						|
    self._valid_requests = []
 | 
						|
    return valid_requests
 | 
						|
 | 
						|
  @property
 | 
						|
  def were_all_requests_http_1_1(self):
 | 
						|
    return self._were_all_requests_http_1_1
 | 
						|
 | 
						|
  def _DoLookForGet(self):
 | 
						|
    """Tries to parse an HTTTP request line.
 | 
						|
 | 
						|
    Returns:
 | 
						|
      (Boolean) True if a request was found.
 | 
						|
 | 
						|
    Raises:
 | 
						|
      UnexpectedMethodError: On a non-GET method.
 | 
						|
    """
 | 
						|
    m = self.REQUEST_RE.match(self._buffer)
 | 
						|
    if not m:
 | 
						|
      return False
 | 
						|
    method, path, http_major, http_minor = m.groups()
 | 
						|
 | 
						|
    if method != 'GET':
 | 
						|
      raise UnexpectedMethodError('Unexpected method: ' + method)
 | 
						|
    if path in ['/', '/index.htm', '/index.html']:
 | 
						|
      raise ServeIndexError()
 | 
						|
 | 
						|
    if http_major != '1' or http_minor != '1':
 | 
						|
      self._were_all_requests_http_1_1 = False
 | 
						|
 | 
						|
#    print method, path
 | 
						|
 | 
						|
    self._pending_request = path
 | 
						|
    self._buffer = self._buffer[m.end():]
 | 
						|
    self._state = self.READING_HEADERS
 | 
						|
    return True
 | 
						|
 | 
						|
  def _DoReadHeader(self):
 | 
						|
    """Tries to parse a HTTP header.
 | 
						|
 | 
						|
    Returns:
 | 
						|
      (Boolean) True if it found the end of the request or a HTTP header.
 | 
						|
    """
 | 
						|
    if self._buffer.startswith('\n'):
 | 
						|
      self._buffer = self._buffer[1:]
 | 
						|
      self._state = self.LOOKING_FOR_GET
 | 
						|
      self._valid_requests.append((self._pending_request,
 | 
						|
                                   self._pending_headers))
 | 
						|
      self._pending_headers = {}
 | 
						|
      self._pending_request = ""
 | 
						|
      return True
 | 
						|
 | 
						|
    m = self.HEADER_RE.match(self._buffer)
 | 
						|
    if not m:
 | 
						|
      return False
 | 
						|
 | 
						|
    header = m.group(1).lower()
 | 
						|
    value = m.group(2).strip().lower()
 | 
						|
    if header not in self._pending_headers:
 | 
						|
      self._pending_headers[header] = value
 | 
						|
    self._buffer = self._buffer[m.end():]
 | 
						|
    return True
 | 
						|
 | 
						|
 | 
						|
class ResponseBuilder(object):
 | 
						|
  """Builds HTTP responses for a list of accumulated requests."""
 | 
						|
 | 
						|
  def __init__(self):
 | 
						|
    """Initializer."""
 | 
						|
    self._max_pipeline_depth = 0
 | 
						|
    self._requested_paths = []
 | 
						|
    self._processed_end = False
 | 
						|
    self._were_all_requests_http_1_1 = True
 | 
						|
 | 
						|
  def QueueRequests(self, requested_paths, were_all_requests_http_1_1):
 | 
						|
    """Adds requests to the queue of requests.
 | 
						|
 | 
						|
    Args:
 | 
						|
      requested_paths: (List of Strings) Requested paths.
 | 
						|
    """
 | 
						|
    self._requested_paths.extend(requested_paths)
 | 
						|
    self._were_all_requests_http_1_1 = were_all_requests_http_1_1
 | 
						|
 | 
						|
  def Chunkify(self, data, chunksize):
 | 
						|
    """ Divides a string into chunks
 | 
						|
    """
 | 
						|
    return [hex(chunksize)[2:] + "\r\n" + data[i:i+chunksize] + "\r\n" for i in range(0, len(data), chunksize)]
 | 
						|
 | 
						|
  def BuildResponses(self):
 | 
						|
    """Converts the queue of requests into responses.
 | 
						|
 | 
						|
    Returns:
 | 
						|
      (String) Buffer containing all of the responses.
 | 
						|
    """
 | 
						|
    result = ""
 | 
						|
    self._max_pipeline_depth = max(self._max_pipeline_depth,
 | 
						|
                                   len(self._requested_paths))
 | 
						|
    for path, headers in self._requested_paths:
 | 
						|
      if path == '/verifiedserver':
 | 
						|
        body = "WE ROOLZ: {}\r\n".format(os.getpid());
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Server: Apache',
 | 
						|
                       'Content-Length: {}'.format(len(body)),
 | 
						|
                       'Cache-Control: no-store'], body)
 | 
						|
 | 
						|
      elif path == '/alphabet.txt':
 | 
						|
        body = 'abcdefghijklmnopqrstuvwxyz'
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Server: Apache',
 | 
						|
                       'Content-Length: 26',
 | 
						|
                       'Cache-Control: no-store'], body)
 | 
						|
 | 
						|
      elif path == '/reverse.txt':
 | 
						|
        body = 'zyxwvutsrqponmlkjihgfedcba'
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Content-Length: 26', 'Cache-Control: no-store'], body)
 | 
						|
 | 
						|
      elif path == '/chunked.txt':
 | 
						|
        body = ('7\r\nchunked\r\n'
 | 
						|
                '8\r\nencoding\r\n'
 | 
						|
                '2\r\nis\r\n'
 | 
						|
                '3\r\nfun\r\n'
 | 
						|
                '0\r\n\r\n')
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Transfer-Encoding: chunked', 'Cache-Control: no-store'],
 | 
						|
            body)
 | 
						|
 | 
						|
      elif path == '/cached.txt':
 | 
						|
        body = 'azbycxdwevfugthsirjqkplomn'
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Content-Length: 26', 'Cache-Control: max-age=60'], body)
 | 
						|
 | 
						|
      elif path == '/connection_close.txt':
 | 
						|
        body = 'azbycxdwevfugthsirjqkplomn'
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Content-Length: 26', 'Cache-Control: max-age=60', 'Connection: close'], body)
 | 
						|
        self._processed_end = True
 | 
						|
 | 
						|
      elif path == '/1k.txt':
 | 
						|
        body = '0123456789abcdef' * 64
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Server: Apache',
 | 
						|
                       'Content-Length: 1024',
 | 
						|
                       'Cache-Control: max-age=60'], body)
 | 
						|
 | 
						|
      elif path == '/10k.txt':
 | 
						|
        body = '0123456789abcdef' * 640
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Server: Apache',
 | 
						|
                       'Content-Length: 10240',
 | 
						|
                       'Cache-Control: max-age=60'], body)
 | 
						|
 | 
						|
      elif path == '/100k.txt':
 | 
						|
        body = '0123456789abcdef' * 6400
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK',
 | 
						|
            ['Server: Apache',
 | 
						|
             'Content-Length: 102400',
 | 
						|
             'Cache-Control: max-age=60'],
 | 
						|
            body)
 | 
						|
 | 
						|
      elif path == '/100k_chunked.txt':
 | 
						|
        body = self.Chunkify('0123456789abcdef' * 6400, 20480)
 | 
						|
        body.append('0\r\n\r\n')
 | 
						|
        body = ''.join(body)
 | 
						|
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK', ['Transfer-Encoding: chunked', 'Cache-Control: no-store'], body)
 | 
						|
 | 
						|
      elif path == '/stats.txt':
 | 
						|
        results = {
 | 
						|
            'max_pipeline_depth': self._max_pipeline_depth,
 | 
						|
            'were_all_requests_http_1_1': int(self._were_all_requests_http_1_1),
 | 
						|
        }
 | 
						|
        body = ','.join(['%s:%s' % (k, v) for k, v in results.items()])
 | 
						|
        result += self._BuildResponse(
 | 
						|
            '200 OK',
 | 
						|
            ['Content-Length: %s' % len(body), 'Cache-Control: no-store'], body)
 | 
						|
        self._processed_end = True
 | 
						|
 | 
						|
      else:
 | 
						|
        result += self._BuildResponse('404 Not Found', ['Content-Length: 7'], 'Go away')
 | 
						|
      if self._processed_end:
 | 
						|
          break
 | 
						|
    self._requested_paths = []
 | 
						|
    return result
 | 
						|
 | 
						|
  def WriteError(self, status, error):
 | 
						|
    """Returns an HTTP response for the specified error.
 | 
						|
 | 
						|
    Args:
 | 
						|
      status: (String) Response code and descrtion (e.g. "404 Not Found")
 | 
						|
 | 
						|
    Returns:
 | 
						|
      (String) Text of HTTP response.
 | 
						|
    """
 | 
						|
    return self._BuildResponse(
 | 
						|
        status, ['Connection: close', 'Content-Type: text/plain'], error)
 | 
						|
 | 
						|
  @property
 | 
						|
  def processed_end(self):
 | 
						|
    return self._processed_end
 | 
						|
 | 
						|
  def _BuildResponse(self, status, headers, body):
 | 
						|
    """Builds an HTTP response.
 | 
						|
 | 
						|
    Args:
 | 
						|
      status: (String) Response code and descrtion (e.g. "200 OK")
 | 
						|
      headers: (List of Strings) Headers (e.g. "Connection: close")
 | 
						|
      body: (String) Response body.
 | 
						|
 | 
						|
    Returns:
 | 
						|
      (String) Text of HTTP response.
 | 
						|
    """
 | 
						|
    return ('HTTP/1.1 %s\r\n'
 | 
						|
            '%s\r\n'
 | 
						|
            '\r\n'
 | 
						|
            '%s' % (status, '\r\n'.join(headers), body))
 | 
						|
 | 
						|
 | 
						|
class PipelineRequestHandler(socketserver.BaseRequestHandler):
 | 
						|
  """Called on an incoming TCP connection."""
 | 
						|
 | 
						|
  def _GetTimeUntilTimeout(self):
 | 
						|
    return self._start_time + TIMEOUT - time.time()
 | 
						|
 | 
						|
  def _GetTimeUntilNextSend(self):
 | 
						|
    if not self._last_queued_time:
 | 
						|
      return TIMEOUT
 | 
						|
    return self._last_queued_time + SEND_BUFFER_TIME - time.time()
 | 
						|
 | 
						|
  def handle(self):
 | 
						|
    self._request_parser = RequestParser()
 | 
						|
    self._response_builder = ResponseBuilder()
 | 
						|
    self._last_queued_time = 0
 | 
						|
    self._num_queued = 0
 | 
						|
    self._num_written = 0
 | 
						|
    self._send_buffer = ""
 | 
						|
    self._start_time = time.time()
 | 
						|
    try:
 | 
						|
      while not self._response_builder.processed_end or self._send_buffer:
 | 
						|
 | 
						|
        time_left = self._GetTimeUntilTimeout()
 | 
						|
        time_until_next_send = self._GetTimeUntilNextSend()
 | 
						|
        max_poll_time = min(time_left, time_until_next_send) + MIN_POLL_TIME
 | 
						|
 | 
						|
        rlist, wlist, xlist = [], [], []
 | 
						|
        fileno = self.request.fileno()
 | 
						|
        if max_poll_time > 0:
 | 
						|
          rlist.append(fileno)
 | 
						|
          if self._send_buffer:
 | 
						|
            wlist.append(fileno)
 | 
						|
          rlist, wlist, xlist = select.select(rlist, wlist, xlist, max_poll_time)
 | 
						|
 | 
						|
        if self._GetTimeUntilTimeout() <= 0:
 | 
						|
          return
 | 
						|
 | 
						|
        if self._GetTimeUntilNextSend() <= 0:
 | 
						|
          self._send_buffer += self._response_builder.BuildResponses()
 | 
						|
          self._num_written = self._num_queued
 | 
						|
          self._last_queued_time = 0
 | 
						|
 | 
						|
        if fileno in rlist:
 | 
						|
          self.request.setblocking(False)
 | 
						|
          new_data = self.request.recv(MAX_REQUEST_SIZE)
 | 
						|
          self.request.setblocking(True)
 | 
						|
          if not new_data:
 | 
						|
            return
 | 
						|
          new_requests = self._request_parser.ParseAdditionalData(new_data)
 | 
						|
          self._response_builder.QueueRequests(
 | 
						|
              new_requests, self._request_parser.were_all_requests_http_1_1)
 | 
						|
          self._num_queued += len(new_requests)
 | 
						|
          self._last_queued_time = time.time()
 | 
						|
        elif fileno in wlist:
 | 
						|
          num_bytes_sent = self.request.send(self._send_buffer[0:4096])
 | 
						|
          self._send_buffer = self._send_buffer[num_bytes_sent:]
 | 
						|
          time.sleep(0.05)
 | 
						|
 | 
						|
    except RequestTooLargeError as e:
 | 
						|
      self.request.send(self._response_builder.WriteError(
 | 
						|
          '413 Request Entity Too Large', e))
 | 
						|
      raise
 | 
						|
    except UnexpectedMethodError as e:
 | 
						|
      self.request.send(self._response_builder.WriteError(
 | 
						|
          '405 Method Not Allowed', e))
 | 
						|
      raise
 | 
						|
    except ServeIndexError:
 | 
						|
      self.request.send(self._response_builder.WriteError(
 | 
						|
          '200 OK', INFO_MESSAGE))
 | 
						|
    except Exception as e:
 | 
						|
      print(e)
 | 
						|
    self.request.close()
 | 
						|
 | 
						|
 | 
						|
class PipelineServer(socketserver.ThreadingMixIn, socketserver.TCPServer):
 | 
						|
  pass
 | 
						|
 | 
						|
 | 
						|
parser = argparse.ArgumentParser()
 | 
						|
parser.add_argument("--port", action="store", default=0,
 | 
						|
                  type=int, help="port to listen on")
 | 
						|
parser.add_argument("--verbose", action="store", default=0,
 | 
						|
                  type=int, help="verbose output")
 | 
						|
parser.add_argument("--pidfile", action="store", default=0,
 | 
						|
                  help="file name for the PID")
 | 
						|
parser.add_argument("--logfile", action="store", default=0,
 | 
						|
                  help="file name for the log")
 | 
						|
parser.add_argument("--srcdir", action="store", default=0,
 | 
						|
                  help="test directory")
 | 
						|
parser.add_argument("--id", action="store", default=0,
 | 
						|
                  help="server ID")
 | 
						|
parser.add_argument("--ipv4", action="store_true", default=0,
 | 
						|
                  help="IPv4 flag")
 | 
						|
args = parser.parse_args()
 | 
						|
 | 
						|
if args.pidfile:
 | 
						|
    pid = os.getpid()
 | 
						|
    f = open(args.pidfile, 'w')
 | 
						|
    f.write('{}'.format(pid))
 | 
						|
    f.close()
 | 
						|
 | 
						|
server = PipelineServer(('0.0.0.0', args.port), PipelineRequestHandler)
 | 
						|
server.allow_reuse_address = True
 | 
						|
server.serve_forever()
 |