| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161 | 
							- from __future__ import print_function
 
- import sys
 
- import os
 
- import timeit
 
- import math
 
- import argparse
 
- import fnmatch
 
- import json
 
- parser = argparse.ArgumentParser(description="Python protobuf benchmark")
 
- parser.add_argument("data_files", metavar="dataFile", nargs="+",
 
-                     help="testing data files.")
 
- parser.add_argument("--json", action="store_const", dest="json",
 
-                     const="yes", default="no",
 
-                     help="Whether to output json results")
 
- parser.add_argument("--behavior_prefix", dest="behavior_prefix",
 
-                     help="The output json format's behavior's name's prefix",
 
-                     default="")
 
- # BEGIN CPP GENERATED MESSAGE
 
- parser.add_argument("--cpp_generated", action="store_const",
 
-                     dest="cpp_generated", const="yes", default="no",
 
-                     help="Whether to link generated code library")
 
- # END CPP GENERATED MESSAGE
 
- args = parser.parse_args()
 
- # BEGIN CPP GENERATED MESSAGE
 
- # CPP generated code must be linked before importing the generated Python code
 
- # for the descriptor can be found in the pool
 
- if args.cpp_generated != "no":
 
-   sys.path.append( os.path.dirname( os.path.dirname( os.path.abspath(__file__) ) ) + "/.libs" )
 
-   import libbenchmark_messages
 
-   sys.path.append( os.path.dirname( os.path.dirname( os.path.abspath(__file__) ) ) + "/tmp" )
 
- # END CPP GENERATED MESSAGE
 
- import datasets.google_message1.proto2.benchmark_message1_proto2_pb2 as benchmark_message1_proto2_pb2
 
- import datasets.google_message1.proto3.benchmark_message1_proto3_pb2 as benchmark_message1_proto3_pb2
 
- import datasets.google_message2.benchmark_message2_pb2 as benchmark_message2_pb2
 
- import datasets.google_message3.benchmark_message3_pb2 as benchmark_message3_pb2
 
- import datasets.google_message4.benchmark_message4_pb2 as benchmark_message4_pb2
 
- import benchmarks_pb2 as benchmarks_pb2
 
- def run_one_test(filename):
 
-   data = open(filename, "rb").read()
 
-   benchmark_dataset = benchmarks_pb2.BenchmarkDataset()
 
-   benchmark_dataset.ParseFromString(data)
 
-   total_bytes = 0
 
-   for payload in benchmark_dataset.payload:
 
-     total_bytes += len(payload)
 
-   benchmark_util = Benchmark(full_iteration=len(benchmark_dataset.payload),
 
-                              module="py_benchmark",
 
-                              setup_method="init",
 
-                              total_bytes=total_bytes)
 
-   result={}
 
-   result["filename"] =  filename
 
-   result["message_name"] =  benchmark_dataset.message_name
 
-   result["benchmarks"] = {}
 
-   benchmark_util.set_test_method("parse_from_benchmark")
 
-   result["benchmarks"][args.behavior_prefix + "_parse_from_benchmark"] = \
 
-     benchmark_util.run_benchmark(setup_method_args='"%s"' % (filename))
 
-   benchmark_util.set_test_method("serialize_to_benchmark")
 
-   result["benchmarks"][args.behavior_prefix + "_serialize_to_benchmark"] = \
 
-     benchmark_util.run_benchmark(setup_method_args='"%s"' % (filename))
 
-   return result
 
- def init(filename):
 
-   global benchmark_dataset, message_class, message_list, counter, total_bytes
 
-   message_list=[]
 
-   counter = 0
 
-   total_bytes = 0
 
-   data = open(filename, "rb").read()
 
-   benchmark_dataset = benchmarks_pb2.BenchmarkDataset()
 
-   benchmark_dataset.ParseFromString(data)
 
-   if benchmark_dataset.message_name == "benchmarks.proto3.GoogleMessage1":
 
-     message_class = benchmark_message1_proto3_pb2.GoogleMessage1
 
-   elif benchmark_dataset.message_name == "benchmarks.proto2.GoogleMessage1":
 
-     message_class = benchmark_message1_proto2_pb2.GoogleMessage1
 
-   elif benchmark_dataset.message_name == "benchmarks.proto2.GoogleMessage2":
 
-     message_class = benchmark_message2_pb2.GoogleMessage2
 
-   elif benchmark_dataset.message_name == "benchmarks.google_message3.GoogleMessage3":
 
-     message_class = benchmark_message3_pb2.GoogleMessage3
 
-   elif benchmark_dataset.message_name == "benchmarks.google_message4.GoogleMessage4":
 
-     message_class = benchmark_message4_pb2.GoogleMessage4
 
-   else:
 
-     raise IOError("Message %s not found!" % (benchmark_dataset.message_name))
 
-   for one_payload in benchmark_dataset.payload:
 
-     temp = message_class()
 
-     temp.ParseFromString(one_payload)
 
-     message_list.append(temp)
 
-     total_bytes += len(one_payload)
 
- def parse_from_benchmark():
 
-   global counter, message_class, benchmark_dataset
 
-   m = message_class().ParseFromString(benchmark_dataset.payload[counter % len(benchmark_dataset.payload)])
 
-   counter = counter + 1
 
- def serialize_to_benchmark():
 
-   global counter, message_list, message_class
 
-   s = message_list[counter % len(benchmark_dataset.payload)].SerializeToString()
 
-   counter = counter + 1
 
- class Benchmark:
 
-   def __init__(self, module=None, test_method=None,
 
-                setup_method=None, total_bytes=None, full_iteration = 1):
 
-     self.full_iteration = full_iteration
 
-     self.module = module
 
-     self.test_method = test_method
 
-     self.setup_method = setup_method
 
-     self.total_bytes = total_bytes
 
-   def set_test_method(self, test_method):
 
-     self.test_method = test_method
 
-   def full_setup_code(self, setup_method_args=''):
 
-     setup_code = ""
 
-     setup_code += "from %s import %s\n" % (self.module, self.test_method)
 
-     setup_code += "from %s import %s\n" % (self.module, self.setup_method)
 
-     setup_code += "%s(%s)\n" % (self.setup_method, setup_method_args)
 
-     return setup_code
 
-   def dry_run(self, test_method_args='', setup_method_args=''):
 
-     return timeit.timeit(stmt="%s(%s)" % (self.test_method, test_method_args),
 
-                          setup=self.full_setup_code(setup_method_args),
 
-                          number=self.full_iteration);
 
-   def run_benchmark(self, test_method_args='', setup_method_args=''):
 
-     reps = self.full_iteration;
 
-     t = self.dry_run(test_method_args, setup_method_args);
 
-     if t < 3 :
 
-       reps = int(math.ceil(3 / t)) * self.full_iteration
 
-     if reps != self.full_iteration:
 
-         t = timeit.timeit(stmt="%s(%s)" % (self.test_method, test_method_args),
 
-                           setup=self.full_setup_code(setup_method_args),
 
-                           number=reps);
 
-     return self.total_bytes * 1.0 / 2 ** 20 / (1.0 * t / reps * self.full_iteration)
 
- if __name__ == "__main__":
 
-   results = []
 
-   for file in args.data_files:
 
-     results.append(run_one_test(file))
 
-   if args.json != "no":
 
-     print(json.dumps(results))
 
-   else:
 
-     for result in results:
 
-       print("Message %s of dataset file %s" % \
 
-           (result["message_name"], result["filename"]))
 
-       print("Average throughput for parse_from_benchmark: %.2f MB/s" % \
 
-           (result["benchmarks"][ \
 
-                       args.behavior_prefix + "_parse_from_benchmark"]))
 
-       print("Average throughput for serialize_to_benchmark: %.2f MB/s" % \
 
-           (result["benchmarks"][ \
 
-                       args.behavior_prefix + "_serialize_to_benchmark"]))
 
-       print("")
 
 
  |