Compare commits
39 Commits
Author | SHA1 | Date | |
---|---|---|---|
5573024fd9 | |||
a58f914c40 | |||
4ec9690981 | |||
6af509c2f6 | |||
22eba69526 | |||
a2855edd47 | |||
a2f720855d | |||
cbb9ff7a23 | |||
84b5bbe325 | |||
a42254cf95
|
|||
3d05ea5d28
|
|||
52217c5251
|
|||
022d7a8926 | |||
698f926376 | |||
f74ad50b94 | |||
0c03d9f94e
|
|||
eca5affd53
|
|||
6236673d28
|
|||
4213dc7329
|
|||
2d3eab0db8
|
|||
73b55b05c4
|
|||
aa74c02498 | |||
6fd2bd0863 | |||
2c78fba3a6 | |||
86b883569f | |||
ba86a08632 | |||
0b61a18eb1 | |||
1418603007 | |||
9926c89ef2 | |||
fc6f407a52 | |||
0e9cb0a7f8 | |||
e3b2ea704d | |||
8bd4a4b695 | |||
89f3cbb5d1 | |||
37c4a373b7 | |||
408cff442c | |||
67b88aa2a1 | |||
7b0238b4a5 | |||
16771227bb |
@ -13,14 +13,44 @@ steps:
|
||||
dockerfile: Dockerfile
|
||||
when:
|
||||
- event: [push, tag]
|
||||
scan_image:
|
||||
image: aquasec/trivy
|
||||
|
||||
scan:
|
||||
image: quay.io/wollud1969/woodpecker-helper:0.5.1
|
||||
environment:
|
||||
TRIVY_TOKEN:
|
||||
from_secret: trivy_token
|
||||
TRIVY_URL:
|
||||
from_secret: trivy_url
|
||||
DTRACK_API_KEY:
|
||||
from_secret: dtrack_api_key
|
||||
DTRACK_API_URL:
|
||||
from_secret: dtrack_api_url
|
||||
commands:
|
||||
- TRIVY_DISABLE_VEX_NOTICE=1 trivy image $FORGE_NAME/$CI_REPO:$CI_COMMIT_SHA --quiet --exit-code 1
|
||||
- HOME=/home/`id -nu`
|
||||
- TAG="${CI_COMMIT_TAG:-$CI_COMMIT_SHA}"
|
||||
- |
|
||||
trivy image \
|
||||
--server $TRIVY_URL \
|
||||
--token $TRIVY_TOKEN \
|
||||
--format cyclonedx \
|
||||
--scanners license \
|
||||
--output /tmp/sbom.xml \
|
||||
$FORGE_NAME/$CI_REPO:$TAG
|
||||
- cat /tmp/sbom.xml
|
||||
- |
|
||||
curl -X "POST" \
|
||||
-H "Content-Type: multipart/form-data" \
|
||||
-H "X-Api-Key: $DTRACK_API_KEY" \
|
||||
-F "autoCreate=true" \
|
||||
-F "projectName=$CI_REPO" \
|
||||
-F "projectVersion=$TAG" \
|
||||
-F "bom=@/tmp/sbom.xml"\
|
||||
"$DTRACK_API_URL/api/v1/bom"
|
||||
when:
|
||||
- event: [push, tag]
|
||||
|
||||
deploy:
|
||||
image: quay.io/wollud1969/k8s-admin-helper:0.1.3
|
||||
image: quay.io/wollud1969/k8s-admin-helper:0.2.1
|
||||
environment:
|
||||
KUBE_CONFIG_CONTENT:
|
||||
from_secret: kube_config
|
||||
|
@ -1,4 +1,4 @@
|
||||
FROM python:3.12-alpine3.21
|
||||
FROM python:3.11-alpine3.21
|
||||
|
||||
ENV REDIS_URL=""
|
||||
ENV SECRET_KEY=""
|
||||
|
@ -5,7 +5,7 @@ metadata:
|
||||
labels:
|
||||
app: pv-stats
|
||||
annotations:
|
||||
secret.reloader.stakater.com/reload=pv-stats
|
||||
secret.reloader.stakater.com/reload: pv-stats
|
||||
spec:
|
||||
replicas: 1
|
||||
selector:
|
||||
|
@ -19,10 +19,11 @@ kubectl create namespace $NAMESPACE \
|
||||
-o yaml | \
|
||||
kubectl -f - apply
|
||||
|
||||
SECRETS_FILE=`mktemp`
|
||||
gpg --decrypt --passphrase $GPG_PASSPHRASE --yes --batch --homedir /tmp/.gnupg --output $SECRETS_FILE secrets.asc
|
||||
. $SECRETS_FILE
|
||||
rm $SECRETS_FILE
|
||||
# SECRETS_FILE=`mktemp`
|
||||
# gpg --decrypt --passphrase $GPG_PASSPHRASE --yes --batch --homedir /tmp/.gnupg --output $SECRETS_FILE secrets.asc
|
||||
# . $SECRETS_FILE
|
||||
# rm $SECRETS_FILE
|
||||
eval "`cat secrets.asc | /usr/local/bin/decrypt-secrets.sh`"
|
||||
|
||||
kubectl create secret generic pv-stats \
|
||||
--dry-run=client \
|
||||
|
123
src/run.py
123
src/run.py
@ -7,9 +7,11 @@ import redis
|
||||
import json
|
||||
import os
|
||||
import plotly.express as px
|
||||
import plotly.graph_objects as po
|
||||
import pandas as pd
|
||||
import psycopg
|
||||
import sqlalchemy
|
||||
import time
|
||||
|
||||
try:
|
||||
redis_url = os.environ['REDIS_URL']
|
||||
@ -47,13 +49,21 @@ def token_debug():
|
||||
@oidc.require_login
|
||||
def index():
|
||||
try:
|
||||
stepX_time = time.time()
|
||||
dbh = psycopg.connect()
|
||||
engine = sqlalchemy.create_engine("postgresql+psycopg://", creator=lambda: dbh)
|
||||
|
||||
step0_time = time.time()
|
||||
df = pd.read_sql("SELECT month, cast(year AS varchar), current_energy AS value FROM pv_energy_by_month", con=engine)
|
||||
step1_time = time.time()
|
||||
duration1 = step1_time - step0_time
|
||||
logger.info(f"{duration1=}")
|
||||
fig_1 = px.bar(df, x='month', y='value', color='year', barmode='group')
|
||||
step2_time = time.time()
|
||||
duration2 = step2_time - step1_time
|
||||
logger.info(f"{duration2=}")
|
||||
fig_1.update_layout(
|
||||
title="Jahreswerte Exportierte Energie",
|
||||
title=f"Jahreswerte Exportierte Energie {duration1:.3f}, {duration2:.3f}",
|
||||
xaxis_title="",
|
||||
yaxis_title="",
|
||||
legend_title="Jahr",
|
||||
@ -64,27 +74,44 @@ def index():
|
||||
),
|
||||
yaxis=dict(ticksuffix=" kWh")
|
||||
)
|
||||
graph_html_1 = fig_1.to_html(full_html=False, default_height='33%')
|
||||
graph_html_1 = fig_1.to_html(full_html=False, default_height='30%')
|
||||
|
||||
step3_time = time.time()
|
||||
df = pd.read_sql("SELECT time_bucket('5 minutes', time) AS bucket, AVG(power) AS avg_power FROM pv_power_v WHERE time >= date_trunc('day', now()) - '1 day'::interval AND time < date_trunc('day', now()) GROUP BY bucket ORDER BY bucket", con=engine)
|
||||
step4_time = time.time()
|
||||
duration3 = step4_time - step3_time
|
||||
logger.info(f"{duration3=}")
|
||||
fig_2 = px.line(df, x='bucket', y='avg_power')
|
||||
step5_time = time.time()
|
||||
duration4 = step5_time - step4_time
|
||||
logger.info(f"{duration4=}")
|
||||
fig_2.update_layout(
|
||||
xaxis_title="",
|
||||
yaxis_title="",
|
||||
title="Export gestern",
|
||||
title=f"Export gestern {duration3:.3f}, {duration4:.3f}",
|
||||
yaxis=dict(ticksuffix=" W")
|
||||
)
|
||||
graph_html_2 = fig_2.to_html(full_html=False, default_height='33%')
|
||||
graph_html_2 = fig_2.to_html(full_html=False, default_height='30%')
|
||||
|
||||
step6_time = time.time()
|
||||
df = pd.read_sql("SELECT time_bucket('5 minutes', time) AS bucket, AVG(power) AS avg_power FROM pv_power_v WHERE time >= date_trunc('day', now()) AND time < date_trunc('day', now()) + '1 day'::interval GROUP BY bucket ORDER BY bucket", con=engine)
|
||||
step7_time = time.time()
|
||||
duration5 = step7_time - step6_time
|
||||
logger.info(f"{duration5=}")
|
||||
fig_3 = px.line(df, x='bucket', y='avg_power')
|
||||
step8_time = time.time()
|
||||
duration6 = step8_time - step7_time
|
||||
logger.info(f"{duration6=}")
|
||||
fig_3.update_layout(
|
||||
xaxis_title="",
|
||||
yaxis_title="",
|
||||
title="Export heute",
|
||||
title=f"Export heute {duration5:.3f}, {duration6:.3f}",
|
||||
yaxis=dict(ticksuffix=" W")
|
||||
)
|
||||
graph_html_3 = fig_3.to_html(full_html=False, default_height='33%')
|
||||
graph_html_3 = fig_3.to_html(full_html=False, default_height='30%')
|
||||
stepZ_time = time.time()
|
||||
duration7 = stepZ_time - stepX_time
|
||||
logger.info(f"{duration7=}")
|
||||
|
||||
return render_template_string(f"""
|
||||
<html>
|
||||
@ -95,6 +122,23 @@ def index():
|
||||
{graph_html_1}
|
||||
{graph_html_2}
|
||||
{graph_html_3}
|
||||
<div style="height:9vh; background-color:lightgrey; font-family: Courier, Consolas, monospace;">
|
||||
<table style="border-collapse: collapse;">
|
||||
<style>
|
||||
td.smallsep {{ padding-right: 10px }}
|
||||
td.largesep {{ padding-right: 30px }}
|
||||
</style>
|
||||
<tr>
|
||||
<td class="smallsep">Query 1:</td><td class="largesep"> {duration1:.3f} s</td><td class="smallsep">Graph 1:</td><td> {duration2:.3f} s</td>
|
||||
</tr><tr>
|
||||
<td class="smallsep">Query 2:</td><td class="largesep"> {duration3:.3f} s</td><td class="smallsep">Graph 2:</td><td> {duration4:.3f} s</td>
|
||||
</tr><tr>
|
||||
<td class="smallsep">Query 3:</td><td class="largesep"> {duration5:.3f} s</td><td class="smallsep">Graph 3:</td><td> {duration6:.3f} s</td>
|
||||
</tr><tr>
|
||||
<td class="smallsep">Total:</td><td> {duration7:.3f} s</td><td></td><td></td>
|
||||
</tr>
|
||||
</table>
|
||||
</div>
|
||||
</body>
|
||||
</html>
|
||||
""")
|
||||
@ -105,6 +149,73 @@ def index():
|
||||
dbh.close()
|
||||
|
||||
|
||||
@app.route('/ntpserver')
|
||||
def ntpserver():
|
||||
try:
|
||||
dbh = psycopg.connect()
|
||||
engine = sqlalchemy.create_engine("postgresql+psycopg://", creator=lambda: dbh)
|
||||
|
||||
query = """
|
||||
select time_bucket('5 minutes', time) as bucket,
|
||||
device,
|
||||
avg(cast(values->'rootdisp'->>'value' as float)) as rootdisp,
|
||||
max(cast(values->'stratum'->>'value' as int)) as stratum
|
||||
from measurements
|
||||
where time >= date_trunc('day', now()) AND time < date_trunc('day', now()) + '1 day'::interval and
|
||||
application = 'TSM' and attributes->>'Label' = 'david'
|
||||
group by bucket, device
|
||||
order by bucket, device
|
||||
"""
|
||||
|
||||
df = pd.read_sql(query, con=engine)
|
||||
|
||||
fig = po.Figure()
|
||||
fig.add_trace(po.Scatter(x=df['bucket'], y=df['rootdisp'], mode='lines', name='Root Dispersion', yaxis='y1', line=dict(color='red')))
|
||||
fig.add_trace(po.Scatter(x=df['bucket'], y=df['stratum'], mode='lines', name='Stratum', yaxis='y2', line=dict(color='blue')))
|
||||
|
||||
fig.update_layout(
|
||||
title='NTP Server Numbers',
|
||||
|
||||
# Linke Y-Achse
|
||||
yaxis=dict(
|
||||
title='Root Dispersion',
|
||||
ticksuffix=' ms'
|
||||
),
|
||||
|
||||
# Rechte Y-Achse
|
||||
yaxis2=dict(
|
||||
title='Stratum',
|
||||
overlaying='y', # Legt die zweite Y-Achse über die erste
|
||||
side='right', # Setzt sie auf die rechte Seite
|
||||
tickmode='linear', # Stellt sicher, dass die Ticks in festen Intervallen sind
|
||||
dtick=1, # Zeigt nur ganzzahlige Ticks
|
||||
),
|
||||
|
||||
legend=dict(x=0.05, y=1) # Position der Legende
|
||||
)
|
||||
|
||||
|
||||
graph_html = fig.to_html(full_html=False, default_height='30%')
|
||||
|
||||
return render_template_string(f"""
|
||||
<html>
|
||||
<head>
|
||||
<title>NTP Server Numbers</title>
|
||||
</head>
|
||||
<body>
|
||||
{graph_html}
|
||||
</body>
|
||||
</html>
|
||||
""")
|
||||
except Exception as e:
|
||||
raise Exception(f"Error when querying NTP server values: {e}")
|
||||
finally:
|
||||
if dbh is not None:
|
||||
dbh.close()
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
if __name__ == '__main__':
|
||||
|
Reference in New Issue
Block a user