Added dependencies locally
This commit is contained in:
0
deps/protobuf/benchmarks/python/__init__.py
vendored
Normal file
0
deps/protobuf/benchmarks/python/__init__.py
vendored
Normal file
161
deps/protobuf/benchmarks/python/py_benchmark.py
vendored
Normal file
161
deps/protobuf/benchmarks/python/py_benchmark.py
vendored
Normal file
@ -0,0 +1,161 @@
|
||||
from __future__ import print_function
|
||||
import sys
|
||||
import os
|
||||
import timeit
|
||||
import math
|
||||
import argparse
|
||||
import fnmatch
|
||||
import json
|
||||
|
||||
parser = argparse.ArgumentParser(description="Python protobuf benchmark")
|
||||
parser.add_argument("data_files", metavar="dataFile", nargs="+",
|
||||
help="testing data files.")
|
||||
parser.add_argument("--json", action="store_const", dest="json",
|
||||
const="yes", default="no",
|
||||
help="Whether to output json results")
|
||||
parser.add_argument("--behavior_prefix", dest="behavior_prefix",
|
||||
help="The output json format's behavior's name's prefix",
|
||||
default="")
|
||||
# BEGIN CPP GENERATED MESSAGE
|
||||
parser.add_argument("--cpp_generated", action="store_const",
|
||||
dest="cpp_generated", const="yes", default="no",
|
||||
help="Whether to link generated code library")
|
||||
# END CPP GENERATED MESSAGE
|
||||
args = parser.parse_args()
|
||||
# BEGIN CPP GENERATED MESSAGE
|
||||
# CPP generated code must be linked before importing the generated Python code
|
||||
# for the descriptor can be found in the pool
|
||||
if args.cpp_generated != "no":
|
||||
sys.path.append( os.path.dirname( os.path.dirname( os.path.abspath(__file__) ) ) + "/.libs" )
|
||||
import libbenchmark_messages
|
||||
sys.path.append( os.path.dirname( os.path.dirname( os.path.abspath(__file__) ) ) + "/tmp" )
|
||||
# END CPP GENERATED MESSAGE
|
||||
|
||||
|
||||
import datasets.google_message1.proto2.benchmark_message1_proto2_pb2 as benchmark_message1_proto2_pb2
|
||||
import datasets.google_message1.proto3.benchmark_message1_proto3_pb2 as benchmark_message1_proto3_pb2
|
||||
import datasets.google_message2.benchmark_message2_pb2 as benchmark_message2_pb2
|
||||
import datasets.google_message3.benchmark_message3_pb2 as benchmark_message3_pb2
|
||||
import datasets.google_message4.benchmark_message4_pb2 as benchmark_message4_pb2
|
||||
import benchmarks_pb2 as benchmarks_pb2
|
||||
|
||||
|
||||
def run_one_test(filename):
|
||||
data = open(filename, "rb").read()
|
||||
benchmark_dataset = benchmarks_pb2.BenchmarkDataset()
|
||||
benchmark_dataset.ParseFromString(data)
|
||||
total_bytes = 0
|
||||
for payload in benchmark_dataset.payload:
|
||||
total_bytes += len(payload)
|
||||
benchmark_util = Benchmark(full_iteration=len(benchmark_dataset.payload),
|
||||
module="py_benchmark",
|
||||
setup_method="init",
|
||||
total_bytes=total_bytes)
|
||||
result={}
|
||||
result["filename"] = filename
|
||||
result["message_name"] = benchmark_dataset.message_name
|
||||
result["benchmarks"] = {}
|
||||
benchmark_util.set_test_method("parse_from_benchmark")
|
||||
result["benchmarks"][args.behavior_prefix + "_parse_from_benchmark"] = \
|
||||
benchmark_util.run_benchmark(setup_method_args='"%s"' % (filename))
|
||||
benchmark_util.set_test_method("serialize_to_benchmark")
|
||||
result["benchmarks"][args.behavior_prefix + "_serialize_to_benchmark"] = \
|
||||
benchmark_util.run_benchmark(setup_method_args='"%s"' % (filename))
|
||||
return result
|
||||
|
||||
|
||||
def init(filename):
|
||||
global benchmark_dataset, message_class, message_list, counter, total_bytes
|
||||
message_list=[]
|
||||
counter = 0
|
||||
total_bytes = 0
|
||||
data = open(filename, "rb").read()
|
||||
benchmark_dataset = benchmarks_pb2.BenchmarkDataset()
|
||||
benchmark_dataset.ParseFromString(data)
|
||||
|
||||
if benchmark_dataset.message_name == "benchmarks.proto3.GoogleMessage1":
|
||||
message_class = benchmark_message1_proto3_pb2.GoogleMessage1
|
||||
elif benchmark_dataset.message_name == "benchmarks.proto2.GoogleMessage1":
|
||||
message_class = benchmark_message1_proto2_pb2.GoogleMessage1
|
||||
elif benchmark_dataset.message_name == "benchmarks.proto2.GoogleMessage2":
|
||||
message_class = benchmark_message2_pb2.GoogleMessage2
|
||||
elif benchmark_dataset.message_name == "benchmarks.google_message3.GoogleMessage3":
|
||||
message_class = benchmark_message3_pb2.GoogleMessage3
|
||||
elif benchmark_dataset.message_name == "benchmarks.google_message4.GoogleMessage4":
|
||||
message_class = benchmark_message4_pb2.GoogleMessage4
|
||||
else:
|
||||
raise IOError("Message %s not found!" % (benchmark_dataset.message_name))
|
||||
|
||||
for one_payload in benchmark_dataset.payload:
|
||||
temp = message_class()
|
||||
temp.ParseFromString(one_payload)
|
||||
message_list.append(temp)
|
||||
total_bytes += len(one_payload)
|
||||
|
||||
|
||||
def parse_from_benchmark():
|
||||
global counter, message_class, benchmark_dataset
|
||||
m = message_class().ParseFromString(benchmark_dataset.payload[counter % len(benchmark_dataset.payload)])
|
||||
counter = counter + 1
|
||||
|
||||
|
||||
def serialize_to_benchmark():
|
||||
global counter, message_list, message_class
|
||||
s = message_list[counter % len(benchmark_dataset.payload)].SerializeToString()
|
||||
counter = counter + 1
|
||||
|
||||
|
||||
class Benchmark:
|
||||
def __init__(self, module=None, test_method=None,
|
||||
setup_method=None, total_bytes=None, full_iteration = 1):
|
||||
self.full_iteration = full_iteration
|
||||
self.module = module
|
||||
self.test_method = test_method
|
||||
self.setup_method = setup_method
|
||||
self.total_bytes = total_bytes
|
||||
|
||||
def set_test_method(self, test_method):
|
||||
self.test_method = test_method
|
||||
|
||||
def full_setup_code(self, setup_method_args=''):
|
||||
setup_code = ""
|
||||
setup_code += "from %s import %s\n" % (self.module, self.test_method)
|
||||
setup_code += "from %s import %s\n" % (self.module, self.setup_method)
|
||||
setup_code += "%s(%s)\n" % (self.setup_method, setup_method_args)
|
||||
return setup_code
|
||||
|
||||
def dry_run(self, test_method_args='', setup_method_args=''):
|
||||
return timeit.timeit(stmt="%s(%s)" % (self.test_method, test_method_args),
|
||||
setup=self.full_setup_code(setup_method_args),
|
||||
number=self.full_iteration);
|
||||
|
||||
def run_benchmark(self, test_method_args='', setup_method_args=''):
|
||||
reps = self.full_iteration;
|
||||
t = self.dry_run(test_method_args, setup_method_args);
|
||||
if t < 3 :
|
||||
reps = int(math.ceil(3 / t)) * self.full_iteration
|
||||
if reps != self.full_iteration:
|
||||
t = timeit.timeit(stmt="%s(%s)" % (self.test_method, test_method_args),
|
||||
setup=self.full_setup_code(setup_method_args),
|
||||
number=reps);
|
||||
return self.total_bytes * 1.0 / 2 ** 20 / (1.0 * t / reps * self.full_iteration)
|
||||
|
||||
|
||||
if __name__ == "__main__":
|
||||
results = []
|
||||
for file in args.data_files:
|
||||
results.append(run_one_test(file))
|
||||
|
||||
if args.json != "no":
|
||||
print(json.dumps(results))
|
||||
else:
|
||||
for result in results:
|
||||
print("Message %s of dataset file %s" % \
|
||||
(result["message_name"], result["filename"]))
|
||||
print("Average throughput for parse_from_benchmark: %.2f MB/s" % \
|
||||
(result["benchmarks"][ \
|
||||
args.behavior_prefix + "_parse_from_benchmark"]))
|
||||
print("Average throughput for serialize_to_benchmark: %.2f MB/s" % \
|
||||
(result["benchmarks"][ \
|
||||
args.behavior_prefix + "_serialize_to_benchmark"]))
|
||||
print("")
|
31
deps/protobuf/benchmarks/python/python_benchmark_messages.cc
vendored
Normal file
31
deps/protobuf/benchmarks/python/python_benchmark_messages.cc
vendored
Normal file
@ -0,0 +1,31 @@
|
||||
#include <Python.h>
|
||||
|
||||
#include "benchmarks.pb.h"
|
||||
#include "datasets/google_message1/proto2/benchmark_message1_proto2.pb.h"
|
||||
#include "datasets/google_message1/proto3/benchmark_message1_proto3.pb.h"
|
||||
#include "datasets/google_message2/benchmark_message2.pb.h"
|
||||
#include "datasets/google_message3/benchmark_message3.pb.h"
|
||||
#include "datasets/google_message4/benchmark_message4.pb.h"
|
||||
|
||||
static struct PyModuleDef _module = {PyModuleDef_HEAD_INIT,
|
||||
"libbenchmark_messages",
|
||||
"Benchmark messages Python module",
|
||||
-1,
|
||||
NULL,
|
||||
NULL,
|
||||
NULL,
|
||||
NULL,
|
||||
NULL};
|
||||
|
||||
extern "C" {
|
||||
PyMODINIT_FUNC PyInit_libbenchmark_messages() {
|
||||
benchmarks::BenchmarkDataset().descriptor();
|
||||
benchmarks::proto3::GoogleMessage1().descriptor();
|
||||
benchmarks::proto2::GoogleMessage1().descriptor();
|
||||
benchmarks::proto2::GoogleMessage2().descriptor();
|
||||
benchmarks::google_message3::GoogleMessage3().descriptor();
|
||||
benchmarks::google_message4::GoogleMessage4().descriptor();
|
||||
|
||||
return PyModule_Create(&_module);
|
||||
}
|
||||
}
|
Reference in New Issue
Block a user