|  | @@ -0,0 +1,101 @@
 | 
	
		
			
				|  |  | +#!/usr/bin/env python
 | 
	
		
			
				|  |  | +# Copyright 2015 gRPC authors.
 | 
	
		
			
				|  |  | +#
 | 
	
		
			
				|  |  | +# Licensed under the Apache License, Version 2.0 (the "License");
 | 
	
		
			
				|  |  | +# you may not use this file except in compliance with the License.
 | 
	
		
			
				|  |  | +# You may obtain a copy of the License at
 | 
	
		
			
				|  |  | +#
 | 
	
		
			
				|  |  | +#     http://www.apache.org/licenses/LICENSE-2.0
 | 
	
		
			
				|  |  | +#
 | 
	
		
			
				|  |  | +# Unless required by applicable law or agreed to in writing, software
 | 
	
		
			
				|  |  | +# distributed under the License is distributed on an "AS IS" BASIS,
 | 
	
		
			
				|  |  | +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 | 
	
		
			
				|  |  | +# See the License for the specific language governing permissions and
 | 
	
		
			
				|  |  | +# limitations under the License.
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +"""Detect new flakes introduced in the last 24h hours with respect to the
 | 
	
		
			
				|  |  | +previous six days"""
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +from __future__ import absolute_import
 | 
	
		
			
				|  |  | +from __future__ import division
 | 
	
		
			
				|  |  | +from __future__ import print_function
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +import os
 | 
	
		
			
				|  |  | +import sys
 | 
	
		
			
				|  |  | +import logging
 | 
	
		
			
				|  |  | +logging.basicConfig(format='%(asctime)s %(message)s')
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +gcp_utils_dir = os.path.abspath(
 | 
	
		
			
				|  |  | +    os.path.join(os.path.dirname(__file__), '../gcp/utils'))
 | 
	
		
			
				|  |  | +sys.path.append(gcp_utils_dir)
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +import big_query_utils
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +def get_flaky_tests(days_lower_bound, days_upper_bound, limit=None):
 | 
	
		
			
				|  |  | +  """ period is one of "WEEK", "DAY", etc.
 | 
	
		
			
				|  |  | +  (see https://cloud.google.com/bigquery/docs/reference/standard-sql/functions-and-operators#date_add). """
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +  bq = big_query_utils.create_big_query()
 | 
	
		
			
				|  |  | +  query = """
 | 
	
		
			
				|  |  | +SELECT
 | 
	
		
			
				|  |  | +  filtered_test_name,
 | 
	
		
			
				|  |  | +  FIRST(timestamp),
 | 
	
		
			
				|  |  | +  FIRST(build_url),
 | 
	
		
			
				|  |  | +FROM (
 | 
	
		
			
				|  |  | +  SELECT
 | 
	
		
			
				|  |  | +    REGEXP_REPLACE(test_name, r'/\d+', '') AS filtered_test_name,
 | 
	
		
			
				|  |  | +    result,
 | 
	
		
			
				|  |  | +    build_url,
 | 
	
		
			
				|  |  | +    timestamp
 | 
	
		
			
				|  |  | +  FROM
 | 
	
		
			
				|  |  | +    [grpc-testing:jenkins_test_results.aggregate_results]
 | 
	
		
			
				|  |  | +  WHERE
 | 
	
		
			
				|  |  | +    timestamp >= DATE_ADD(CURRENT_DATE(), {days_lower_bound}, "DAY")
 | 
	
		
			
				|  |  | +    AND timestamp <= DATE_ADD(CURRENT_DATE(), {days_upper_bound}, "DAY")
 | 
	
		
			
				|  |  | +    AND NOT REGEXP_MATCH(job_name, '.*portability.*'))
 | 
	
		
			
				|  |  | +GROUP BY
 | 
	
		
			
				|  |  | +  filtered_test_name,
 | 
	
		
			
				|  |  | +  timestamp,
 | 
	
		
			
				|  |  | +  build_url
 | 
	
		
			
				|  |  | +HAVING
 | 
	
		
			
				|  |  | +  SUM(result != 'PASSED'
 | 
	
		
			
				|  |  | +    AND result != 'SKIPPED') > 0
 | 
	
		
			
				|  |  | +ORDER BY
 | 
	
		
			
				|  |  | +  timestamp ASC
 | 
	
		
			
				|  |  | +""".format(days_lower_bound=days_lower_bound, days_upper_bound=days_upper_bound)
 | 
	
		
			
				|  |  | +  if limit:
 | 
	
		
			
				|  |  | +    query += '\n LIMIT {}'.format(limit)
 | 
	
		
			
				|  |  | +  query_job = big_query_utils.sync_query_job(bq, 'grpc-testing', query)
 | 
	
		
			
				|  |  | +  page = bq.jobs().getQueryResults(
 | 
	
		
			
				|  |  | +      pageToken=None, **query_job['jobReference']).execute(num_retries=3)
 | 
	
		
			
				|  |  | +  testname_to_ts_url_pair = {row['f'][0]['v']: (row['f'][1]['v'], row['f'][2]['v']) for row in page['rows']}
 | 
	
		
			
				|  |  | +  return testname_to_ts_url_pair
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +def get_new_flakes():
 | 
	
		
			
				|  |  | +  last_week_sans_yesterday = get_flaky_tests(-7, -1)
 | 
	
		
			
				|  |  | +  last_24 = get_flaky_tests(-1, +1)
 | 
	
		
			
				|  |  | +  last_week_sans_yesterday_names = set(last_week_sans_yesterday.keys())
 | 
	
		
			
				|  |  | +  last_24_names = set(last_24.keys())
 | 
	
		
			
				|  |  | +  logging.debug('|last_week_sans_yesterday| =', len(last_week_sans_yesterday_names))
 | 
	
		
			
				|  |  | +  logging.debug('|last_24_names| =', len(last_24_names))
 | 
	
		
			
				|  |  | +  new_flakes = last_24_names - last_week_sans_yesterday_names
 | 
	
		
			
				|  |  | +  logging.debug('|new_flakes| = ', len(new_flakes))
 | 
	
		
			
				|  |  | +  return {k: last_24[k] for k in new_flakes}
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +def main():
 | 
	
		
			
				|  |  | +  import datetime
 | 
	
		
			
				|  |  | +  new_flakes = get_new_flakes()
 | 
	
		
			
				|  |  | +  if new_flakes:
 | 
	
		
			
				|  |  | +    print("Found {} new flakes:".format(len(new_flakes)))
 | 
	
		
			
				|  |  | +    for k, v in new_flakes.items():
 | 
	
		
			
				|  |  | +      ts = int(float(v[0]))
 | 
	
		
			
				|  |  | +      url = v[1]
 | 
	
		
			
				|  |  | +      human_ts = datetime.datetime.utcfromtimestamp(ts).strftime('%Y-%m-%d %H:%M:%S UTC')
 | 
	
		
			
				|  |  | +      print("Test: {}, Timestamp: {}, URL: {}\n".format(k, human_ts, url))
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +
 | 
	
		
			
				|  |  | +if __name__ == '__main__':
 | 
	
		
			
				|  |  | +  main()
 |