Skip to content

Commit 58ccebb

Browse files
committed
Changes to collect metrics
1 parent 8e00c68 commit 58ccebb

File tree

5 files changed

+49
-45
lines changed

5 files changed

+49
-45
lines changed

conf/collector/vector/collector-template.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -157,4 +157,4 @@ objects:
157157
maxUnavailable: 1
158158
type: RollingUpdate
159159
parameters:
160-
- name: image_name
160+
- name: image_name

conf/collector/vector/vector-clf.toml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -39,7 +39,7 @@
3939
route.app = '!(starts_with!(.kubernetes.pod_namespace,"kube") && starts_with!(.kubernetes.pod_namespace,"openshift") && .kubernetes.pod_namespace == "default")'
4040

4141
[sinks.stress]
42-
type="file"
42+
type = "file"
4343
inputs = ["ocp_sys.infra","ocp_sys.app"]
4444
encoding.codec = "ndjson"
45-
path = "/var/log/containers/stress.log"
45+
path = "/var/log/containers/stress.log"

conf/monitor/capture-statistics-template.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,8 +48,8 @@ objects:
4848
chmod 744 kubectl-top.sh ;
4949
echo "Collecting Statistics";
5050
OUTPUT_FILE=/var/log/containers/stress.log;
51+
sleep 60;
5152
echo -e "\nOutput log file is: $OUTPUT_FILE\n";
52-
touch $OUTPUT_FILE;
5353
./check-logs-sequence -rf ${output_format} -f $OUTPUT_FILE -c ${number_of_log_lines_between_reports} -l ${check_log_level} &
5454
echo "=== setup-end ===";
5555
while true; do

deploy_to_openshift.sh

Lines changed: 12 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ select_stress_profile() {
3030
heavy_containers_msg_per_sec=1000
3131
low_containers_msg_per_sec=10
3232
number_of_log_lines_between_reports=10;
33-
maximum_logfile_size=10485760;
33+
maximum_logfile_size=52428800; # default file size
3434

3535
case $stress_profile in
3636
"no-stress")
@@ -39,71 +39,55 @@ select_stress_profile() {
3939
number_low_stress_containers=0;
4040
low_containers_msg_per_sec=0;
4141
number_of_log_lines_between_reports=10;
42-
maximum_logfile_size=10485760;
42+
maximum_logfile_size=52428800;
4343
;;
4444
"very-light")
4545
number_heavy_stress_containers=0;
4646
heavy_containers_msg_per_sec=0;
4747
number_low_stress_containers=1;
4848
low_containers_msg_per_sec=10;
4949
number_of_log_lines_between_reports=100;
50-
maximum_logfile_size=10485760;
50+
maximum_logfile_size=52428800;
5151
;;
5252
"light")
5353
number_heavy_stress_containers=1;
5454
heavy_containers_msg_per_sec=100;
5555
number_low_stress_containers=2;
5656
low_containers_msg_per_sec=10;
5757
number_of_log_lines_between_reports=1000;
58-
maximum_logfile_size=1048576;
59-
;;
60-
"experiment")
61-
number_heavy_stress_containers=0;
62-
heavy_containers_msg_per_sec=0;
63-
number_low_stress_containers=20;
64-
low_containers_msg_per_sec=20000;
65-
number_of_log_lines_between_reports=100;
66-
maximum_logfile_size=10485760;
58+
maximum_logfile_size=52428800;
6759
;;
6860
"medium")
6961
number_heavy_stress_containers=2;
7062
heavy_containers_msg_per_sec=1000;
7163
number_low_stress_containers=10;
7264
low_containers_msg_per_sec=10;
7365
number_of_log_lines_between_reports=20000;
74-
maximum_logfile_size=1048576;
66+
maximum_logfile_size=52428800;
7567
;;
7668
"heavy")
7769
number_heavy_stress_containers=0;
7870
heavy_containers_msg_per_sec=0;
7971
number_low_stress_containers=10;
8072
low_containers_msg_per_sec=1500;
8173
number_of_log_lines_between_reports=200000;
82-
maximum_logfile_size=1048576;
83-
;;
84-
"very-heavy")
85-
number_heavy_stress_containers=0;
86-
heavy_containers_msg_per_sec=0;
87-
number_low_stress_containers=10;
88-
low_containers_msg_per_sec=3000;
89-
number_of_log_lines_between_reports=300000;
90-
maximum_logfile_size=1048576;
74+
maximum_logfile_size=52428800;
9175
;;
9276
"heavy-loss")
9377
number_heavy_stress_containers=2;
9478
heavy_containers_msg_per_sec=20000;
9579
number_low_stress_containers=8;
9680
low_containers_msg_per_sec=1500;
9781
number_of_log_lines_between_reports=200000;
98-
maximum_logfile_size=1048576;
82+
maximum_logfile_size=52428800;
9983
;;
10084
"very-heavy")
10185
number_heavy_stress_containers=10;
10286
heavy_containers_msg_per_sec=20000;
10387
number_low_stress_containers=10;
10488
low_containers_msg_per_sec=1500;
10589
number_of_log_lines_between_reports=1000000;
106-
maximum_logfile_size=1048576;
90+
maximum_logfile_size=52428800;
10791
;;
10892
*) show_usage
10993
;;
@@ -156,8 +140,8 @@ deploy() {
156140
delete_logstress_project_if_exists
157141
create_logstress_project
158142
set_credentials
159-
deploy_logstress $number_heavy_stress_containers $heavy_containers_msg_per_sec $number_low_stress_containers $low_containers_msg_per_sec $use_log_samples
160-
if $gowatcher ; then deploy_gologfilewatcher "$gologfilewatcher_image"; fi
143+
144+
161145
case "$collector" in
162146
'vector') deploy_log_collector_vector "$vector_image" "$vector_conf";;
163147
'fluentd') deploy_log_collector_fluentd "$fluentd_image" "$fluentd_conf_file" "$fluentd_pre";;
@@ -168,6 +152,8 @@ deploy() {
168152
deploy_log_collector_fluentbit "$fluentbit_image" conf/collector/fluentbit/dual/fluentbit.conf;;
169153
*) show_usage ;;
170154
esac
155+
deploy_logstress $number_heavy_stress_containers $heavy_containers_msg_per_sec $number_low_stress_containers $low_containers_msg_per_sec $use_log_samples
156+
if $gowatcher ; then deploy_gologfilewatcher "$gologfilewatcher_image"; fi
171157
if $gowatcher ; then expose_metrics_to_prometheus; fi
172158
deploy_capture_statistics $number_of_log_lines_between_reports "$output_format" "$report_interval"
173159
if $evacuate_node ; then evacuate_node_for_performance_tests; fi

print_performance_metrics.sh

Lines changed: 33 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -1,22 +1,40 @@
11
# This script collects metrics like log-per-sec, cpu-percentage, cpu-cores and memory required per iterations in running reg-ex rules
22
# in discovering log-levels.
3-
3+
# collector values could be fluentd, fluentbit
4+
export collector=$1
45
export capture_pod=`oc get pods | grep capture | cut -d" " -f1`
5-
export fluentd_pod=`oc get pods | grep fluentd | cut -d" " -f1`
6-
export iterations=`oc logs $capture_pod | grep " | ruby" | wc -l`
7-
8-
if [ "$iterations" -lt 10 ]; then echo "Total Iterations till now: $iterations Results will be printed after 10 iterations"; exit 1; fi
9-
10-
echo "Total No of iterations: $iterations"
6+
export collector_pod=`oc get pods | grep $collector | cut -d" " -f1`
7+
export iterations=`oc logs $capture_pod | grep "Top information on:" | wc -l`
118

12-
export LPS=`oc logs $capture_pod | grep -i "Total collected logs per sec:" | cut -d ":" -f2 | awk '{ SUM += $1} END { print SUM/NR }'`
13-
echo "Avg logs per sec/iter: $LPS"
9+
echo "$collector $collector_pod"
10+
# if [ "$iterations" -lt 10 ]; then echo "Total Iterations till now: $iterations Results will be printed after 10 iterations"; exit 1; fi
11+
echo "Total Iterations till now: $iterations"
1412

15-
export Cpu_Percentage=`oc logs $capture_pod | grep -i "| ruby" | cut -d "|" -f1 | awk '{ SUM += $1} END { print SUM/NR }'`
16-
echo "Avg cpu percentage/iter: $Cpu_Percentage"
13+
while : ; do
14+
iterations=`oc logs $capture_pod | grep "Top information on:" | wc -l`
15+
export total_time=`oc logs $capture_pod | grep "Time from start monitoring (in secs)" | cut -d ":" -f2 | tr -d ' ' | tail -1`
16+
echo "Total Iterations till now: $iterations"
17+
echo "Total time till now: $total_time"
18+
export current_LPS=`oc logs $capture_pod | grep -i "Total collected logs per sec:" | cut -d ":" -f2 | tr -d ' ' | tail -1`
19+
echo "Current LPS=$current_LPS"
20+
export LPS=`oc logs $capture_pod | grep -i "Total collected logs per sec:" | cut -d ":" -f2 | awk '{ SUM += $1} END { print SUM/NR }'`
21+
echo "Avg logs per sec/iter: $LPS"
1722

18-
export Cpu_Core=`oc logs $capture_pod | grep $fluentd_pod | awk '{print $2}' | cut -d 'm' -f1 | awk '{ SUM+= $1} END { print SUM/NR }'`
19-
echo "Avg cpu core/iter: $Cpu_Core"
23+
export current_cpu_core=`oc logs $capture_pod | grep $collector_pod | awk '{print $2}' | cut -d 'm' -f1 | tail -1`
24+
echo "Current CPU core=$current_cpu_core"
25+
export Cpu_Core=`oc logs $capture_pod | grep $collector_pod | awk '{print $2}' | cut -d 'm' -f1 | awk '{ SUM+= $1} END { print SUM/NR }'`
26+
echo "Avg cpu core/iter: $Cpu_Core"
27+
export current_memory=`oc logs $capture_pod | grep $collector_pod | awk '{print $3}' | cut -d 'M' -f1 | tail -1`
28+
echo "Current Memory=$current_memory"
29+
export Memory=`oc logs $capture_pod | grep $collector_pod | awk '{print $3}' | cut -d 'M' -f1 | awk '{ SUM+= $1} END { print SUM/NR }'`
30+
echo "Avg memory/iter: $Memory"
31+
export end_time=$(date +%s%N | cut -b1-13)
32+
echo "End time: $end_time"
33+
if [ "$total_time" -ge 1800 ]; then break; fi
34+
sleep 10
35+
done
2036

21-
export Memory=`oc logs $capture_pod | grep $fluentd_pod | awk '{print $3}' | cut -d 'M' -f1 | awk '{ SUM+= $1} END { print SUM/NR }'`
22-
echo "Avg memory/iter: $Memory"
37+
export end_time=$(date +%s%N | cut -b1-13)
38+
echo "Start time: $start_time"
39+
echo "End time: $end_time"
40+
echo "Exiting after $iterations"

0 commit comments

Comments
 (0)