40 lines
1.5 KiB
Markdown
40 lines
1.5 KiB
Markdown
---
|
|
title: AI Keynote discussion
|
|
weight: 2
|
|
tags:
|
|
- keynote
|
|
- ai
|
|
- panel
|
|
---
|
|
|
|
A podium discussion (somewhat scripted) lead by Pryanka
|
|
|
|
## Guests
|
|
|
|
* Tim from Mistral
|
|
* Paige from Google AI
|
|
* Jeff founder of OLLAMA
|
|
|
|
## Discussion
|
|
|
|
* What do you use as the base of dev for ollama
|
|
* Jeff: The concepts from docker, git, kubernetes
|
|
* How is the balance between ai engi and ai ops
|
|
* Jeff: The classic dev vs ops devide, many ML-Engi don't think about
|
|
* Paige: Yessir
|
|
* How does infra keep up with the fast research
|
|
* Paige: Well, they don't - but they do their best and Cloudnative is cool
|
|
* Jeff: Well we're not google, but kubernetes is the saviour
|
|
* What are scaling constraints
|
|
* Jeff: Currently sizing of models is still in it's infancy
|
|
* Jeff: There will be more specific hardware and someone will have to support it
|
|
* Paige: Sizing also depends on latency needs (code autocompletion vs performance optimization)
|
|
* Paige: Optimization of smaller models
|
|
* What technologies need to be open source licensed
|
|
* Jeff: The model b/c access and trust
|
|
* Tim: The models and base execution environemtn -> Vendor agnosticism
|
|
* Paige: Yes and remixes are really imporant for development
|
|
* Anything else
|
|
* Jeff: How do we bring our awesome tools (monitoring, logging, security) to the new AI world
|
|
* Paige: Currently many people just use paid apis to abstract the infra, but we need this stuff selfhostable
|
|
* Tim: I don'T want to know about the hardware, the whole infra side should be done by the cloudnative teams to let ML-Engi to just be ML-Engi |