docs(day2): Added edge cpu talk notes
This commit is contained in:
parent
11e3866f01
commit
84abb7e1b9
@ -5,12 +5,17 @@ title: Cloud Native Summit Munich 2025
|
||||
|
||||
All about the things I did and sessions I attended at Cloud Native Summit 2025 in Munich.
|
||||
|
||||
This current version is probably full of typos - will fix later. This is what typing the notes blindly in real time get's you.
|
||||
This current version is probably full of typos - might fix later (prbly won't tbh). This is what typing the notes blindly in real time get's you.
|
||||
|
||||
## How did I get there?
|
||||
|
||||
I attended Cloud Native Rejekts and KubeCon + CloudNativeCon Europe 2025 in London and some of the atendees reccomended checking out CNS Munich for another event in the same spirit as Cloud Native Rejekts.
|
||||
After a short talk with my boss, I there by my employer [DATEV eG](https://datev.de) alongside two of my coworkers.
|
||||
After a short talk with my boss, I got sent there by my employer [DATEV eG](https://datev.de) alongside two of my coworkers.
|
||||
|
||||
## And how was it.
|
||||
|
||||
I'd say that attending CNS Munich 2025 was worth it. The event is pretty close to my place of employment (2hrs by car or train) and relatively small in size (400 attendees). The talks varied a bit - the first day had a bunch of interesting talks but the second day indulged in ai-related talks (and they were not quite my cup of tea). This might me fine for others but I've heard enogh about ai use cases for the coming years at the last events i attended (or just reddit).
|
||||
Maybe disributing the ai-talks over the two days - while always providing an interesting alternative - might be the right move for next time.
|
||||
|
||||
## And how does this website get it's content
|
||||
|
||||
|
56
content/day2/03_k3s-gpu.md
Normal file
56
content/day2/03_k3s-gpu.md
Normal file
@ -0,0 +1,56 @@
|
||||
---
|
||||
title: "Brains on the edge - running ai workloads with k3s and gpu nodes"
|
||||
weight: 3
|
||||
tags:
|
||||
- ai
|
||||
- gpu
|
||||
---
|
||||
|
||||
<!-- {{% button href="https://youtu.be/rkteV6Mzjfs" style="warning" icon="video" %}}Watch talk on YouTube{{% /button %}} -->
|
||||
<!-- {{% button href="https://docs.google.com/presentation/d/1nEK0CVC_yQgIDqwsdh-PRihB6dc9RyT-" style="tip" icon="person-chalkboard" %}}Slides{{% /button %}} -->
|
||||
|
||||
I decided not to note down the usual "typical challenges on the edge" slides (about 10 mins of the talk)
|
||||
|
||||
## Baseline
|
||||
|
||||
- Edge can be split up: Near Edge, Far Edge, Device Edge
|
||||
- They use k3s for all edge clusters
|
||||
|
||||
## Prerequisites
|
||||
|
||||
- Software: GPU Driver, Container Toolkit, Device Plugin
|
||||
- Hardware: NVIDIA GPU with a supported distro
|
||||
- Runtime: Not all runtimes support GPUs (containerd and CRI-O do)
|
||||
|
||||
## Architecture
|
||||
|
||||
```mermaid
|
||||
graph LR
|
||||
subgraph Edge
|
||||
MQTT
|
||||
Kafka
|
||||
Analytics
|
||||
|
||||
MQTT-->|Publish collected sensor data|Kafka
|
||||
Kafka-->|Provide data to run|Analytics
|
||||
end
|
||||
subgraph Azure
|
||||
Storage
|
||||
Monitoring
|
||||
MLFlow
|
||||
|
||||
Storage-->|Provide long term analytics|MLFlow
|
||||
end
|
||||
|
||||
Analytics<-->|Sync models|MLFlow
|
||||
Kafka-->|Save to long term|Storage
|
||||
Monitoring-.->|Observe|Storage
|
||||
Monitoring-.->|Observe|MLFlow
|
||||
```
|
||||
|
||||
## Q&A
|
||||
|
||||
- Did you use the nvidia gpu operator: Yes
|
||||
- Which runtime did you use: ContainerD via K3S
|
||||
- Why k3s over k0s: Because we used it
|
||||
- Were you power limited: Nope, the edge was on a large ship
|
Loading…
x
Reference in New Issue
Block a user