α¬ tetrisology α’αα»αααααα·α
αα½α
α
α’αααΈβαααβααααΈβααΊβααα’βααααααα
βα
ααβααααβα
αΆααβα
Epigraphs α
ααΆααααααΎααααα αΆ
α’αααααααΌαααΆαααα―αααΆααααααα ααα» PostgreSQL αα
αα
α»ααααααααΈ AWS cloud ααΆαααααΆαααα
ααΆαααααΆαααΈα Linux αααα»ααααααααααα’αααα αα·αβαααβαα
βαααα»αβαααβαα·αβααααΆααβαα ααα»ααααβααΌαβαα·ααΆαβααΆβααΉαβααΆαβααΆαβαααααΆαβαααβααααα·α
α
ααααααααΆαααααΆαααααΎαα
αα
α»ααααααααΆαα―αααΆααααααα ααα»ααΊ 5 ααΆααΈα
α―αααΆααααααα ααα»αααα»α AWS ααααΌαααΆαααααα·ααααααΆααααααα
α§ααααααααααΆαααααΎ
ααΎααααΈααΆαααα―αααΆααααααα ααα»αα
αααΆαααΈα ααααααΈα bash ααααΌαααΆαααααΎαααα α
AWS API "
αααΆαααΆααααααα
- -db-instance-identifier: AWS instance name;
- --log-file-name: αααααααα―αααΆααααααα ααα»αααααΆααααααΎααα αα α»αααααα
- --max-itemα α ααα½αααα»αααααΆαα»αααααΆααααα‘αααααα·ααα αααα»αααααααααΆααααααααΆαααα ααααααααα―αααΆααααααΆαααΆαααα
- --starting-tokenα αα·αα·ααααααααΆα αΆααααααΎα
α αΎαααΆααΆαααα - αα·α
αα
ααΆααα½αα±ααα
αΆααα’αΆαααααααααααΆααααΆαααααα»ααααααΆα αα·αααΆααα»αααααΆαααα»αα’αα‘α»ααααααααααααΎααΆαα
αααα»αααΉααααααααΆαααα αΆααααΌαααΆααααααααΆααα½α
α αΎααααααΆαααααΈαα·ααααα
αΆαααααα ααα»αααα Google αα ααβαα·αβααΆαβααααΆαβαααααααααΆαβααΆαα½αβα‘αΎα α αΎαβαααα»αβαα·αβααΆαβααααβα
ααβαααααβααβα±ααβααΆαβαααΈβααααα
βααααααβαααβααα ααααααΆαααΆααααα ααΆααΆααΆαα αΆααααααΆαααααα’α
ααΆααααα αααΆααααΌαααΆαααααΆααα·α αα
α―αααΆααααααα ααα»α
α»ααααααααΆααααααΆααααΆα
αααΎαααααααααα’αααα ααΆαααααΆα ααα·α α―αααΆααααααα ααα»α’αΆα
ααααΌαααΆαααααΆαα’αααΈαα½αααΌα
αααα
ααΎααΆααααΉαα’αααα’αααΈα’αααΈαα½αα αΎαα¬αα
? ααΎ Tetris ααΆααααα’αααΈααΆαα½αααΆ? α αΎααααααΊααΆα’αααΈαααααΆααααΌαααααΎααΆαα½αααΆα
ααααα·αααΎααΎααααααααΎααααααΎααααα’αΆα
ααΎαααΆααα
ααααααα»αα―αααΆααααααΆααααΆααααΆα ααα·α (αααααΆααααΆαααΆαααα αααα»αααααΈαααα’αα»ααααΆαα±αααααααΆααααΆαααααααααΌα
ααααΆ) ααΎαααα½αααΆα ααααα Tetris αααααααΆαα
1) α―αααΆαααααΌαααΆαααΆαααααΆαααααα»α α αΎαααΆα―αααΆαα
α»ααααααα ααα ααααααααΊααααΆαααα αα―αααΆαα
α»ααααααα
2) α―αααΆαααααα ααα αααααΆααααΊααΌα
ααΆαααα αα―αααΆαα
α»ααααααα
3) α―αααΆαααΊααΆααΆαααααααα―αααΆααα»α αα·αααΆαααΆαααααα ααα αααααΆααααΊααΌα
ααΆαααα αααααα
αααααα―αααΆαα
α»ααααααα
4) α―αααΆαααΊααΆααΆαααααααα―αααΆααα»α αα·αααΆα―αααΆαα
α»ααααααα ααα αααααΆααααΆαααα αααααΆαααα αααααα
αααααα―αααΆαα
α»ααααααα
ααΆααα·α
αα
ααΊαααααΌααααα»αα
αα»ααααααα¬ααα Tetris αα
ααΎααααα·αααααΈαα½αα
αααα αΆαααααΎαα‘αΎααα ααααααααααΆααααα αΆ
1) ααΆααα·ααααα 2 ααααα
ααΆααΌαα
αα·αααΆααααα αΆαα·αααααα αααα αΆαααααααΆαααΈαααααα·ααααΆααααααααααα·ααΈααααΌαα
ααα ααααααΎααα’αααα»α
ααα»ααααααααα½αα±ααα
αΆααα’αΆααααααααΆααααααααα·α
α
ααΆα’αα»αα αα·αααΆααα·ααΈααααΎα’α»α αααα·ααααααΆααααΈααααΆααααααα
αΆααααααΎαααα
ααΌα αααα’αααααΆαααΉαα αΎαααΆαααααΎα βstarting-token ααααΌαααΆαααααΎααΎααααΈαααααΆαααααααααααααααΌαα αΆααααααΎαααΆαααααααααααα αααααΎααααααααααα String αααααΆααααααΆααααα·αααΎα’αααααααΆααΆααααααααααααα’α»α αααα·ααα ααΈαα»αααααα’αααα Next Token αααααΎαααΉααα·αααααΌαααΆααααααααα·α αΆαααΆααΆα’α»α αααα·αααα
ααΌα
ααααα αΎα α’αααααααΌαα’αΆαααΆααααΆαααα
ααααα·αααΎα’αααα’αΆααααα»αααααααα α
ααα½αααααΆαα’αΆαααΉαααΆααα·α
αα½α
ααα»ααααααα·ααΆαααΉαααΆαα’αα·ααααΆα
ααααα·αααΎα’αααα’αΆααααα»ααααααααΌα
α αααα»ααα
αα·α α
ααα½αααααΆαα’αΆαααΉαααΆαα’αα·ααααΆ ααα»ααααααα·ααΆαααΉαααΆααα·α
αα½α
αααα»αα
ααΌα
αααα ααΎααααΈααΆαααααααα
ααΆα
ααα αα·αααΎααααΈαααααααΆαααΌαα
αααααααααααΆα αααα»αααααΌααααααααααααααΆααα½ααααααα αααααΆα’αα»ααααΎααα
ααΌα
ααΆααΎα
ααααααααα·α
α
αααααΆααα§ααΆα ααα ααΌααα·α αΆαααΆααααΎαααΆαααααΆαααΆαααα―αααΆααααααα ααα»ααΆ 2 αααααααααΆααααα½ααααα»αα α ααα½αααααΆαα’αΆααα αααα»αααααΈααΆααααΈαααΊα’αΆαααααααΎααα ααααααα
1) αααα»ααααα»ααααααααΌα
α:
2) αααα»ααα
αααα»αααααααα:
ααΌα
ααααααΆαααααααααΆαααα’αααα»αααΊαα
αααααΆα.
ααα ααααααΎααΊαα·α
αα½α
ααα»αααααααα»αα’αα‘α»ααααααααΎαααΆαα’αΆα ααα αα’αΆα
ααααΌαααΆααααααΎα ααΎααααΈααΆαααααααα
ααα½αααααΆαα’αΆαα
ααΆαα½αααααααΌαααΆαααααααααΆαα ααΆαααα αΆααααΆαααααΎαααΎαααα αααααα’αααααΎαααααααααααα’αΆα α’αΆαααΆααα·αααΆααααααΌαααΆααααααααΆααα α‘αΎα α αΎαααΆαααΆαα±ααααΆαααΆααα·ααααΆ αα·αααΆααα·ααΆααααΈααααα αααααααααα αααα ααααΆααααα·α αααααααα
ααΆααα·αααααΆααΌαα ααααΆαα’αα»αααα
ααΆααΆαααααΆαααααααααΆαααααΎ
CREATE TABLE endpoint
(
id SERIAL ,
host text
);
TABLE database
(
id SERIAL ,
β¦
last_aws_log_time text ,
last_aws_nexttoken text ,
aws_max_item_size integer
);
last_aws_log_time β Π²ΡΠ΅ΠΌΠ΅Π½Π½Π°Ρ ΠΌΠ΅ΡΠΊΠ° ΠΏΠΎΡΠ»Π΅Π΄Π½Π΅Π³ΠΎ Π·Π°Π³ΡΡΠΆΠ΅Π½Π½ΠΎΠ³ΠΎ Π»ΠΎΠ³-ΡΠ°ΠΉΠ»Π° Π² ΡΠΎΡΠΌΠ°ΡΠ΅ YYYY-MM-DD-HH24.
last_aws_nexttoken β ΡΠ΅ΠΊΡΡΠΎΠ²Π°Ρ ΠΌΠ΅ΡΠΊΠ° ΠΏΠΎΡΠ»Π΅Π΄Π½Π΅ΠΉ Π·Π°Π³ΡΡΠΆΠ΅Π½Π½ΠΎΠΉ ΠΏΠΎΡΡΠΈΠΈ.
aws_max_item_size- ΡΠΌΠΏΠΈΡΠΈΡΠ΅ΡΠΊΠΈΠΌ ΠΏΡΡΠ΅ΠΌ, ΠΏΠΎΠ΄ΠΎΠ±ΡΠ°Π½Π½ΡΠΉ Π½Π°ΡΠ°Π»ΡΠ½ΡΠΉ ΡΠ°Π·ΠΌΠ΅Ρ ΠΏΠΎΡΡΠΈΠΈ.
α’αααααααααααΈαααα
ααΆααα_aws_piece.sh
#!/bin/bash
#########################################################
# download_aws_piece.sh
# downloan piece of log from AWS
# version HABR
let min_item_size=1024
let max_item_size=1048576
let growth_factor=3
let growth_counter=1
let growth_counter_max=3
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh:''STARTED'
AWS_LOG_TIME=$1
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh:AWS_LOG_TIME='$AWS_LOG_TIME
database_id=$2
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh:database_id='$database_id
RESULT_FILE=$3
endpoint=`psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE_DATABASE -A -t -c "select e.host from endpoint e join database d on e.id = d.endpoint_id where d.id = $database_id "`
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh:endpoint='$endpoint
db_instance=`echo $endpoint | awk -F"." '{print toupper($1)}'`
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh:db_instance='$db_instance
LOG_FILE=$RESULT_FILE'.tmp_log'
TMP_FILE=$LOG_FILE'.tmp'
TMP_MIDDLE=$LOG_FILE'.tmp_mid'
TMP_MIDDLE2=$LOG_FILE'.tmp_mid2'
current_aws_log_time=`psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -A -t -c "select last_aws_log_time from database where id = $database_id "`
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh:current_aws_log_time='$current_aws_log_time
if [[ $current_aws_log_time != $AWS_LOG_TIME ]];
then
is_new_log='1'
if ! psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -v ON_ERROR_STOP=1 -A -t -q -c "update database set last_aws_log_time = '$AWS_LOG_TIME' where id = $database_id "
then
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: FATAL_ERROR - update database set last_aws_log_time .'
exit 1
fi
else
is_new_log='0'
fi
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh:is_new_log='$is_new_log
let last_aws_max_item_size=`psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -A -t -c "select aws_max_item_size from database where id = $database_id "`
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: last_aws_max_item_size='$last_aws_max_item_size
let count=1
if [[ $is_new_log == '1' ]];
then
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: START DOWNLOADING OF NEW AWS LOG'
if ! aws rds download-db-log-file-portion
--max-items $last_aws_max_item_size
--region REGION
--db-instance-identifier $db_instance
--log-file-name error/postgresql.log.$AWS_LOG_TIME > $LOG_FILE
then
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: FATAL_ERROR - Could not get log from AWS .'
exit 2
fi
else
next_token=`psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -v ON_ERROR_STOP=1 -A -t -c "select last_aws_nexttoken from database where id = $database_id "`
if [[ $next_token == '' ]];
then
next_token='0'
fi
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: CONTINUE DOWNLOADING OF AWS LOG'
if ! aws rds download-db-log-file-portion
--max-items $last_aws_max_item_size
--starting-token $next_token
--region REGION
--db-instance-identifier $db_instance
--log-file-name error/postgresql.log.$AWS_LOG_TIME > $LOG_FILE
then
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: FATAL_ERROR - Could not get log from AWS .'
exit 3
fi
line_count=`cat $LOG_FILE | wc -l`
let lines=$line_count-1
tail -$lines $LOG_FILE > $TMP_MIDDLE
mv -f $TMP_MIDDLE $LOG_FILE
fi
next_token_str=`cat $LOG_FILE | grep NEXTTOKEN`
next_token=`echo $next_token_str | awk -F" " '{ print $2}' `
grep -v NEXTTOKEN $LOG_FILE > $TMP_FILE
if [[ $next_token == '' ]];
then
cp $TMP_FILE $RESULT_FILE
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: NEXTTOKEN NOT FOUND - FINISH '
rm $LOG_FILE
rm $TMP_FILE
rm $TMP_MIDDLE
rm $TMP_MIDDLE2
exit 0
else
psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -v ON_ERROR_STOP=1 -A -t -q -c "update database set last_aws_nexttoken = '$next_token' where id = $database_id "
fi
first_str=`tail -1 $TMP_FILE`
line_count=`cat $TMP_FILE | wc -l`
let lines=$line_count-1
head -$lines $TMP_FILE > $RESULT_FILE
###############################################
# MAIN CIRCLE
let count=2
while [[ $next_token != '' ]];
do
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: count='$count
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: START DOWNLOADING OF AWS LOG'
if ! aws rds download-db-log-file-portion
--max-items $last_aws_max_item_size
--starting-token $next_token
--region REGION
--db-instance-identifier $db_instance
--log-file-name error/postgresql.log.$AWS_LOG_TIME > $LOG_FILE
then
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: FATAL_ERROR - Could not get log from AWS .'
exit 4
fi
next_token_str=`cat $LOG_FILE | grep NEXTTOKEN`
next_token=`echo $next_token_str | awk -F" " '{ print $2}' `
TMP_FILE=$LOG_FILE'.tmp'
grep -v NEXTTOKEN $LOG_FILE > $TMP_FILE
last_str=`head -1 $TMP_FILE`
if [[ $next_token == '' ]];
then
concat_str=$first_str$last_str
echo $concat_str >> $RESULT_FILE
line_count=`cat $TMP_FILE | wc -l`
let lines=$line_count-1
tail -$lines $TMP_FILE >> $RESULT_FILE
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: NEXTTOKEN NOT FOUND - FINISH '
rm $LOG_FILE
rm $TMP_FILE
rm $TMP_MIDDLE
rm $TMP_MIDDLE2
exit 0
fi
if [[ $next_token != '' ]];
then
let growth_counter=$growth_counter+1
if [[ $growth_counter -gt $growth_counter_max ]];
then
let last_aws_max_item_size=$last_aws_max_item_size*$growth_factor
let growth_counter=1
fi
if [[ $last_aws_max_item_size -gt $max_item_size ]];
then
let last_aws_max_item_size=$max_item_size
fi
psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -A -t -q -c "update database set last_aws_nexttoken = '$next_token' where id = $database_id "
concat_str=$first_str$last_str
echo $concat_str >> $RESULT_FILE
line_count=`cat $TMP_FILE | wc -l`
let lines=$line_count-1
#############################
#Get middle of file
head -$lines $TMP_FILE > $TMP_MIDDLE
line_count=`cat $TMP_MIDDLE | wc -l`
let lines=$line_count-1
tail -$lines $TMP_MIDDLE > $TMP_MIDDLE2
cat $TMP_MIDDLE2 >> $RESULT_FILE
first_str=`tail -1 $TMP_FILE`
fi
let count=$count+1
done
#
#################################################################
exit 0
αααααααααααΈαααΆαα½αααΉαααΆααααααααα½αα ααα½αα
αααΆαααΆααααααβαααα αΌαβααααααΈαα
- ααααΆαααααααΆαααααααα―αααΆααααααα ααα»αααα»ααααααα YYYY-MM-DD-HH24: AWS_LOG_TIME=$1
- ααααααααΆααααΌαααααΆααα·ααααααα database_id=$2
- αααααα―αααΆααααααα ααα»ααααααααΌαααΆαα RESULT_FILE=$3
ααα½αααΆαααααΆαααααααΆααα―αααΆααααααα ααα»αααααΆααααα»αα α»ααααααα
current_aws_log_time=`psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -A -t -c "select last_aws_log_time from database where id = $database_id "`
ααααα·αααΎααααΆαααααααΆααα―αααΆααααααα ααα»αααααΆααααα»αα α»αααααααα·αααααΌαααααΆααΉααααΆαααΆαααααααααα αΌα αααα―αααΆααααααα ααα»ααααΈααααΌαααΆααααα»αα
if [[ $current_aws_log_time != $AWS_LOG_TIME ]];
then
is_new_log='1'
if ! psql -h ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -v ON_ERROR_STOP=1 -A -t -c "update database set last_aws_log_time = '$AWS_LOG_TIME' where id = $database_id "
then
echo '***download_aws_piece.sh -FATAL_ERROR - update database set last_aws_log_time .'
exit 1
fi
else
is_new_log='0'
fi
ααΎαααα½αααΆααααααααααααΆααααααΆαααααΆααααΈα―αααΆααααααΆαααΆαααα
next_token_str=`cat $LOG_FILE | grep NEXTTOKEN`
next_token=`echo $next_token_str | awk -F" " '{ print $2}' `
ααααααααααΆαααααααααααααααΎααΆαααααΆααααΆααααα ααααααΆαααΆαααα
αα
αααα»αααααα·ααα»α ααΎαααΆαααααααααα―αααΆα αααααααΆαααααααΆαααα
ααΆαααααΌα αα·ααααααΎαααα ααααααααα
ααααα·ααα»αααααΆαα
# MAIN CIRCLE
let count=2
while [[ $next_token != '' ]];
do
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: count='$count
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: START DOWNLOADING OF AWS LOG'
if ! aws rds download-db-log-file-portion
--max-items $last_aws_max_item_size
--starting-token $next_token
--region REGION
--db-instance-identifier $db_instance
--log-file-name error/postgresql.log.$AWS_LOG_TIME > $LOG_FILE
then
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: FATAL_ERROR - Could not get log from AWS .'
exit 4
fi
next_token_str=`cat $LOG_FILE | grep NEXTTOKEN`
next_token=`echo $next_token_str | awk -F" " '{ print $2}' `
TMP_FILE=$LOG_FILE'.tmp'
grep -v NEXTTOKEN $LOG_FILE > $TMP_FILE
last_str=`head -1 $TMP_FILE`
if [[ $next_token == '' ]];
then
concat_str=$first_str$last_str
echo $concat_str >> $RESULT_FILE
line_count=`cat $TMP_FILE | wc -l`
let lines=$line_count-1
tail -$lines $TMP_FILE >> $RESULT_FILE
echo $(date +%Y%m%d%H%M)': download_aws_piece.sh: NEXTTOKEN NOT FOUND - FINISH '
rm $LOG_FILE
rm $TMP_FILE
rm $TMP_MIDDLE
rm $TMP_MIDDLE2
exit 0
fi
if [[ $next_token != '' ]];
then
let growth_counter=$growth_counter+1
if [[ $growth_counter -gt $growth_counter_max ]];
then
let last_aws_max_item_size=$last_aws_max_item_size*$growth_factor
let growth_counter=1
fi
if [[ $last_aws_max_item_size -gt $max_item_size ]];
then
let last_aws_max_item_size=$max_item_size
fi
psql -h MONITOR_ENDPOINT.rds.amazonaws.com -U USER -d MONITOR_DATABASE -A -t -q -c "update database set last_aws_nexttoken = '$next_token' where id = $database_id "
concat_str=$first_str$last_str
echo $concat_str >> $RESULT_FILE
line_count=`cat $TMP_FILE | wc -l`
let lines=$line_count-1
#############################
#Get middle of file
head -$lines $TMP_FILE > $TMP_MIDDLE
line_count=`cat $TMP_MIDDLE | wc -l`
let lines=$line_count-1
tail -$lines $TMP_MIDDLE > $TMP_MIDDLE2
cat $TMP_MIDDLE2 >> $RESULT_FILE
first_str=`tail -1 $TMP_FILE`
fi
let count=$count+1
done
ααΆαβα’αααΈβαααααΆαα?
ααΌα
ααααααΆααα·α
αα
ααααα·ααααααααααΌα - "ααΆαααα―αααΆααααααα ααα»ααΈααα" ααααΌαααΆααααααααΆαα α’αααΈαααααααΌαααααΎααΆαα½ααααααα ααα»αααααΆαααΆααα?
ααααΌαα’αααααααΌααααα―αααΆααααααα ααα»α αΎααααααααααΎαα·αααααΆααααΈααΆα
ααΆααα·α
αα
αα·ααα·ααΆαααααΆααααα ααααααΈα bash ααΆαααααααα»αααααΎααΆαααΆαααΆααααΆαααα’α
upload_log_query.sh
#!/bin/bash
#########################################################
# upload_log_query.sh
# Upload table table from dowloaded aws file
# version HABR
###########################################################
echo 'TIMESTAMP:'$(date +%c)' Upload log_query table '
source_file=$1
echo 'source_file='$source_file
database_id=$2
echo 'database_id='$database_id
beginer=' '
first_line='1'
let "line_count=0"
sql_line=' '
sql_flag=' '
space=' '
cat $source_file | while read line
do
line="$space$line"
if [[ $first_line == "1" ]]; then
beginer=`echo $line | awk -F" " '{ print $1}' `
first_line='0'
fi
current_beginer=`echo $line | awk -F" " '{ print $1}' `
if [[ $current_beginer == $beginer ]]; then
if [[ $sql_flag == '1' ]]; then
sql_flag='0'
log_date=`echo $sql_line | awk -F" " '{ print $1}' `
log_time=`echo $sql_line | awk -F" " '{ print $2}' `
duration=`echo $sql_line | awk -F" " '{ print $5}' `
#replace ' to ''
sql_modline=`echo "$sql_line" | sed 's/'''/''''''/g'`
sql_line=' '
################
#PROCESSING OF THE SQL-SELECT IS HERE
if ! psql -h ENDPOINT.rds.amazonaws.com -U USER -d DATABASE -v ON_ERROR_STOP=1 -A -t -c "select log_query('$ip_port',$database_id , '$log_date' , '$log_time' , '$duration' , '$sql_modline' )"
then
echo 'FATAL_ERROR - log_query '
exit 1
fi
################
fi #if [[ $sql_flag == '1' ]]; then
let "line_count=line_count+1"
check=`echo $line | awk -F" " '{ print $8}' `
check_sql=${check^^}
#echo 'check_sql='$check_sql
if [[ $check_sql == 'SELECT' ]]; then
sql_flag='1'
sql_line="$sql_line$line"
ip_port=`echo $sql_line | awk -F":" '{ print $4}' `
fi
else
if [[ $sql_flag == '1' ]]; then
sql_line="$sql_line$line"
fi
fi #if [[ $current_beginer == $beginer ]]; then
done
α₯α‘αΌααααα’αααα’αΆα ααααΎααΆαααΆαα½αααααΎαααααΆαααααΎαααΎαααΈα―αααΆααααααα ααα»α
α αΎαα±ααΆαααΆαααααααααααΆα αααΎαααΆαααΎαα‘αΎαα
αααα½ααααααΆααααααααΌααααααΆαα»ααα ααααααααΆαα½αα ααΆααΆαααααΆααααααααΌαααΆαααααΎαααααΆααααΆααααα log_query
CREATE TABLE log_query
(
id SERIAL ,
queryid bigint ,
query_md5hash text not null ,
database_id integer not null ,
timepoint timestamp without time zone not null,
duration double precision not null ,
query text not null ,
explained_plan text[],
plan_md5hash text ,
explained_plan_wo_costs text[],
plan_hash_value text ,
baseline_id integer ,
ip text ,
port text
);
ALTER TABLE log_query ADD PRIMARY KEY (id);
ALTER TABLE log_query ADD CONSTRAINT queryid_timepoint_unique_key UNIQUE (queryid, timepoint );
ALTER TABLE log_query ADD CONSTRAINT query_md5hash_timepoint_unique_key UNIQUE (query_md5hash, timepoint );
CREATE INDEX log_query_timepoint_idx ON log_query (timepoint);
CREATE INDEX log_query_queryid_idx ON log_query (queryid);
ALTER TABLE log_query ADD CONSTRAINT database_id_fk FOREIGN KEY (database_id) REFERENCES database (id) ON DELETE CASCADE ;
ααααΎαααααΆααααααααΌαααΆαααααΎαααΆααα
αααα»α plpgsql αα»αααΆα "log_query"α
log_query.sql
--log_query.sql
--verison HABR
CREATE OR REPLACE FUNCTION log_query( ip_port text ,log_database_id integer , log_date text , log_time text , duration text , sql_line text ) RETURNS boolean AS $$
DECLARE
result boolean ;
log_timepoint timestamp without time zone ;
log_duration double precision ;
pos integer ;
log_query text ;
activity_string text ;
log_md5hash text ;
log_explain_plan text[] ;
log_planhash text ;
log_plan_wo_costs text[] ;
database_rec record ;
pg_stat_query text ;
test_log_query text ;
log_query_rec record;
found_flag boolean;
pg_stat_history_rec record ;
port_start integer ;
port_end integer ;
client_ip text ;
client_port text ;
log_queryid bigint ;
log_query_text text ;
pg_stat_query_text text ;
BEGIN
result = TRUE ;
RAISE NOTICE '***log_query';
port_start = position('(' in ip_port);
port_end = position(')' in ip_port);
client_ip = substring( ip_port from 1 for port_start-1 );
client_port = substring( ip_port from port_start+1 for port_end-port_start-1 );
SELECT e.host , d.name , d.owner_pwd
INTO database_rec
FROM database d JOIN endpoint e ON e.id = d.endpoint_id
WHERE d.id = log_database_id ;
log_timepoint = to_timestamp(log_date||' '||log_time,'YYYY-MM-DD HH24-MI-SS');
log_duration = duration:: double precision;
pos = position ('SELECT' in UPPER(sql_line) );
log_query = substring( sql_line from pos for LENGTH(sql_line));
log_query = regexp_replace(log_query,' +',' ','g');
log_query = regexp_replace(log_query,';+','','g');
log_query = trim(trailing ' ' from log_query);
log_md5hash = md5( log_query::text );
--Explain execution plan--
EXECUTE 'SELECT dblink_connect(''LINK1'',''host='||database_rec.host||' dbname='||database_rec.name||' user=DATABASE password='||database_rec.owner_pwd||' '')';
log_explain_plan = ARRAY ( SELECT * FROM dblink('LINK1', 'EXPLAIN '||log_query ) AS t (plan text) );
log_plan_wo_costs = ARRAY ( SELECT * FROM dblink('LINK1', 'EXPLAIN ( COSTS FALSE ) '||log_query ) AS t (plan text) );
PERFORM dblink_disconnect('LINK1');
--------------------------
BEGIN
INSERT INTO log_query
(
query_md5hash ,
database_id ,
timepoint ,
duration ,
query ,
explained_plan ,
plan_md5hash ,
explained_plan_wo_costs ,
plan_hash_value ,
ip ,
port
)
VALUES
(
log_md5hash ,
log_database_id ,
log_timepoint ,
log_duration ,
log_query ,
log_explain_plan ,
md5(log_explain_plan::text) ,
log_plan_wo_costs ,
md5(log_plan_wo_costs::text),
client_ip ,
client_port
);
activity_string = 'New query has logged '||
' database_id = '|| log_database_id ||
' query_md5hash='||log_md5hash||
' , timepoint = '||to_char(log_timepoint,'YYYYMMDD HH24:MI:SS');
RAISE NOTICE '%',activity_string;
PERFORM pg_log( log_database_id , 'log_query' , activity_string);
EXCEPTION
WHEN unique_violation THEN
RAISE NOTICE '*** unique_violation *** query already has logged';
END;
SELECT queryid
INTO log_queryid
FROM log_query
WHERE query_md5hash = log_md5hash AND
timepoint = log_timepoint;
IF log_queryid IS NOT NULL
THEN
RAISE NOTICE 'log_query with query_md5hash = % and timepoint = % has already has a QUERYID = %',log_md5hash,log_timepoint , log_queryid ;
RETURN result;
END IF;
------------------------------------------------
RAISE NOTICE 'Update queryid';
SELECT *
INTO log_query_rec
FROM log_query
WHERE query_md5hash = log_md5hash AND timepoint = log_timepoint ;
log_query_rec.query=regexp_replace(log_query_rec.query,';+','','g');
FOR pg_stat_history_rec IN
SELECT
queryid ,
query
FROM
pg_stat_db_queries
WHERE
database_id = log_database_id AND
queryid is not null
LOOP
pg_stat_query = pg_stat_history_rec.query ;
pg_stat_query=regexp_replace(pg_stat_query,'n+',' ','g');
pg_stat_query=regexp_replace(pg_stat_query,'t+',' ','g');
pg_stat_query=regexp_replace(pg_stat_query,' +',' ','g');
pg_stat_query=regexp_replace(pg_stat_query,'$.','%','g');
log_query_text = trim(trailing ' ' from log_query_rec.query);
pg_stat_query_text = pg_stat_query;
--SELECT log_query_rec.query like pg_stat_query INTO found_flag ;
IF (log_query_text LIKE pg_stat_query_text) THEN
found_flag = TRUE ;
ELSE
found_flag = FALSE ;
END IF;
IF found_flag THEN
UPDATE log_query SET queryid = pg_stat_history_rec.queryid WHERE query_md5hash = log_md5hash AND timepoint = log_timepoint ;
activity_string = ' updated queryid = '||pg_stat_history_rec.queryid||
' for log_query with id = '||log_query_rec.id
;
RAISE NOTICE '%',activity_string;
EXIT ;
END IF ;
END LOOP ;
RETURN result ;
END
$$ LANGUAGE plpgsql;
ααΆααΆαααααΆααααααααΌαααΆαααααΎααααΆααααα‘α»ααααααααΎαααΆα pg_stat_db_queriesαααααΆαααΌααααααααα½ααα
αα
α»ααααααααΈααΆααΆα pg_stat_history (ααΆαααααΎααααΆααααΆααΆαααααΌαααΆααα·αααααΆαα
ααΈααα β
TABLE pg_stat_db_queries
(
database_id integer,
queryid bigint ,
query text ,
max_time double precision
);
TABLE pg_stat_history
(
β¦
database_id integer ,
β¦
queryid bigint ,
β¦
max_time double precision ,
β¦
);
αα»αααΆαα’αα»ααααΆαα±ααα’αααα’αα»αααααααααααΆαααΆααααααααααα½αα ααα½ααααααΆααααααΎαααΆαααααΎααΈα―αααΆααααααα ααα»α αααααΊα
α±ααΆαααΈ 1 - αααααααα·ααααΆαααααα·ααααα·αααα½α
ααΆαααααααααααααΆααααΆαααααααΆααααΆαα
αΆααααααΎααααααααΆαα§ααααααα·α ααα»ααααΆαα’αα»ααααα ααΆααααΌα ααααααααααΈαααααααα·ααΆααααα - ααΎααΆαααΊααααΆαα
αΆααααααΎααα
αααααΆ?
αααααΆααααααααα
ααΆαααααα
αα»ααΆαααααΎαα ααα»ααααΆααααα
α αααα ααααΆααΆααααα»αααΌαααααΆααα·ααααααααΆαααΎαα‘αΎααααΆαααααΆαα α αΎαααααΎααΆααααΆαααα·αααΆαα’αααΈαααααααΌαααααΎααΆαα½αααΆααα
ααααααααΆαα»ααααΈαα
ααΆααΆα log_query
port_start = position('(' in ip_port);
port_end = position(')' in ip_port);
client_ip = substring( ip_port from 1 for port_start-1 );
client_port = substring( ip_port from port_start+1 for port_end-port_start-1 );
SELECT e.host , d.name , d.owner_pwd
INTO database_rec
FROM database d JOIN endpoint e ON e.id = d.endpoint_id
WHERE d.id = log_database_id ;
log_timepoint = to_timestamp(log_date||' '||log_time,'YYYY-MM-DD HH24-MI-SS');
log_duration = to_number(duration,'99999999999999999999D9999999999');
pos = position ('SELECT' in UPPER(sql_line) );
log_query = substring( sql_line from pos for LENGTH(sql_line));
log_query = regexp_replace(log_query,' +',' ','g');
log_query = regexp_replace(log_query,';+','','g');
log_query = trim(trailing ' ' from log_query);
RAISE NOTICE 'log_query=%',log_query ;
log_md5hash = md5( log_query::text );
--Explain execution plan--
EXECUTE 'SELECT dblink_connect(''LINK1'',''host='||database_rec.host||' dbname='||database_rec.name||' user=DATABASE password='||database_rec.owner_pwd||' '')';
log_explain_plan = ARRAY ( SELECT * FROM dblink('LINK1', 'EXPLAIN '||log_query ) AS t (plan text) );
log_plan_wo_costs = ARRAY ( SELECT * FROM dblink('LINK1', 'EXPLAIN ( COSTS FALSE ) '||log_query ) AS t (plan text) );
PERFORM dblink_disconnect('LINK1');
--------------------------
BEGIN
INSERT INTO log_query
(
query_md5hash ,
database_id ,
timepoint ,
duration ,
query ,
explained_plan ,
plan_md5hash ,
explained_plan_wo_costs ,
plan_hash_value ,
ip ,
port
)
VALUES
(
log_md5hash ,
log_database_id ,
log_timepoint ,
log_duration ,
log_query ,
log_explain_plan ,
md5(log_explain_plan::text) ,
log_plan_wo_costs ,
md5(log_plan_wo_costs::text),
client_ip ,
client_port
);
ααααααΆα #2 - αααααΆαα»ααααααΆαααααα·ααααα·αααα½α
αα α ααα»α ααα ααΆαααααΆαα-ααΆαααααααΊ-ααΆααααα ααααα·α’αΆα ααΉαααΎαα‘αΎαα "ααα»ααααααΆαααΆααααααααααααααααααααααα·αα½α α αΎα" ααΆα / α αΆα ααΆαα ααΈααα ααα»ααααααΎααΆαα’αααΈααΎαα‘αΎαααααα·αααΎαααααΆαααααα·ααααα·ααααΌαααΆααααααΆαα»ααααα»αα―αααΆααααααα ααα»ααΌα ααααΆ α αΎαααΎααααΈαααααΆαα»αααΆαααααΆααααΆααα·ααΆααααααα α’αααααααΌααααα―αααΆααααααα ααα»?
α’αααΈααααααα»αααααΌαααΆαααΊα
ααΈαα½αα αααααΆαα»ααααααΆαααααα·ααααα·αααα»αααΆααΆαααααΆααααΌαααααΆααα·αααααααααα½ααα·αα·αααα
ααΈααΈαα ααΎααααΈα’αΆα
αααααααααααααΆαααααα·ααααα·ααΆαα½αααααΆ ααΎααααΈααΎαααΎαααααΆααααΆαααααΆαααααα·ααααα·αααα½αααΆαααααΆααααααΌαα
ααΆαααααΎααΆαα½ααααΆαααΆααααααααααα·ααααα·ααΆααααΆααα ααΆαααα½αααΆα αα·ααααααΆαα»ααααααΆαααααα·ααααα·ααααααΆαααααααΎ EXPLAIN ααΊααΆαα·α
αα
ααΆααααα
ααΆαααααα
ααα αααααααΎαααααα EXPLAIN (COSTS FALSE) α’αααα’αΆα
ααα½αααΆαααααααααααααΆα αααααΉαααααΌαααααΎααΎααααΈααα½αααΆαααααα hash αααααααΆα αααααΉααα½ααααα»αααΆααα·ααΆαααΆαααααααααΆαααααααααααα·ααααΆαααααΆααααααΌααα
αααα»ααααααΆαααααα·ααααα·α
ααα½αααΆαααααΌαααααΆαααααα·ααααα·
--Explain execution plan--
EXECUTE 'SELECT dblink_connect(''LINK1'',''host='||database_rec.host||' dbname='||database_rec.name||' user=DATABASE password='||database_rec.owner_pwd||' '')';
log_explain_plan = ARRAY ( SELECT * FROM dblink('LINK1', 'EXPLAIN '||log_query ) AS t (plan text) );
log_plan_wo_costs = ARRAY ( SELECT * FROM dblink('LINK1', 'EXPLAIN ( COSTS FALSE ) '||log_query ) AS t (plan text) );
PERFORM dblink_disconnect('LINK1');
ααααααΆαααΈ 3 - ααΆαααααΎαααααα ααα»αααα½ααααααΆααααΆααααα½ααα·αα·ααα
αααααΆαβααααααβααΆαβα’αα»ααααβααααΌαβααΆαβαααααβαα
ααΆααααααααβαα·αβαα
βααΎβα’αααααβααααΎβαααβαα ααα»ααααβαα
βααΎβαααβαααααΆααβααααβααΆ α’αααβααααΌαβααααΆααβααααΎβααΈβα―αααΆαβαααααβα ααα»βααΆαα½αβααΉαβααααΎβαααβααΆαβαααααβαα
ααΆααααααααβααααααβααΆαβα’αα»ααααα
ααΆααΆααααααΎαααΆαα αααΆαα αα
ααΆααααΎααααΈα±ααααΆααααααααΆαα·αααααΆααααααΆαααΎαα‘αΎαααα§ααααααα·α ααα»ααααΆαααααααα
αα·ααΈααα αα
ααααααα§ααααααα·α ααα»ααααΆαα’αα»ααααααΎαα‘αΎααααααΆααααααααααΆααααααΎ ααΆααΉαααΆαααΆαααααΆαααα
ααΆααααααΎααΆααααΆαααα½ααααααΆαααααααααΆαααΆααααααααΆααααΆαα αα·ααααααααΆααααα·ααααα·αα·αααααΆαα αα·αααααααααααααΎα ααα½αααΆαααααααΆαααααααααααΎαααα·αααααΆαααα»αααααα pg_stat_statements - ααΆααααΌαααΆαα αΆαααΆααα
ααααααααααα½αααααααΎ αα·αααααΎαα
αα
α»ααααααααΆαααΆαα»αα
αααα»αααΆααΆα log_query
SELECT *
INTO log_query_rec
FROM log_query
WHERE query_md5hash = log_md5hash AND timepoint = log_timepoint ;
log_query_rec.query=regexp_replace(log_query_rec.query,';+','','g');
FOR pg_stat_history_rec IN
SELECT
queryid ,
query
FROM
pg_stat_db_queries
WHERE
database_id = log_database_id AND
queryid is not null
LOOP
pg_stat_query = pg_stat_history_rec.query ;
pg_stat_query=regexp_replace(pg_stat_query,'n+',' ','g');
pg_stat_query=regexp_replace(pg_stat_query,'t+',' ','g');
pg_stat_query=regexp_replace(pg_stat_query,' +',' ','g');
pg_stat_query=regexp_replace(pg_stat_query,'$.','%','g');
log_query_text = trim(trailing ' ' from log_query_rec.query);
pg_stat_query_text = pg_stat_query;
--SELECT log_query_rec.query like pg_stat_query INTO found_flag ;
IF (log_query_text LIKE pg_stat_query_text) THEN
found_flag = TRUE ;
ELSE
found_flag = FALSE ;
END IF;
IF found_flag THEN
UPDATE log_query SET queryid = pg_stat_history_rec.queryid WHERE query_md5hash = log_md5hash AND timepoint = log_timepoint ;
activity_string = ' updated queryid = '||pg_stat_history_rec.queryid||
' for log_query with id = '||log_query_rec.id
;
RAISE NOTICE '%',activity_string;
EXIT ;
END IF ;
END LOOP ;
Afterword
αα
αα
ααααααααααΆααα·αααααΆαα
ααΈαααα»αααΆαααααΎααααααα·ααΈαα
αααα»α
αααααΈααΆααΆααΆααα·ααααααααΆααααα·αααααΆαααααα½ααααααααα»αααΆααΉαα αΆαααΆα ααααα»αααΆαααααΎααΆααααααααααααΎαααα½ααααααααΆααααααΆααααΆαααααΎαααΎααα·αααααΆααααααΌαααα αααααααααααααΆαααΆαααα αααα αΆβαα βαα·αβααΆααβααααΌαβααΆαβαααβααααΆαβαα βα‘αΎαβααβαααα»αβααααΈβααΌαα α ααΆαααα ααααΆαα½αα±ααα αΆααα’αΆααααααα
ααα»ααααααΆααΆααΏααα»αααααΆααΆαααααα»α...
ααααα: www.habr.com