Job
- Level
- Senior
- Job Feld
- Software, Data
- Anstellung
- Vollzeit
- Vertragsart
- Unbefristetes Dienstverhältnis
- Ort
- Wien
- Arbeitsmodell
- Onsite
Job Zusammenfassung
In dieser Rolle entwickelst du robuste ML-Services, integrierst Modelle in unsere Monorepo, automatisierst CI/CD-Pipelines und sorgst für Service-Zuverlässigkeit durch Monitoring sowie Optimierung.
Job Technologien
Deine Rolle im Team
You’ll be the bridge between research and production.
Partnering closely with researchers, you’ll ensure experimental code is production ready, integrate models into our monorepo, build shared libraries and services, and create the tooling and processes that let multiple model variants ship safely and quickly.
Your work shortens the research-to-user loop, reduces duplication, and ensures our ML features are reliable, observable, and easy for other teams to adopt.
At the moment, this role is focused on:
- Research-to-Production Pipeline: Hardening experimental models (containerisation, tests, CI/CD), making them deployable for real users.
- Library development: Converting experiments into well-factored libraries with clear APIs, dependency hygiene, and versioning—so teams can import rather than copy-paste.
- Multi-Variant & Parallel Execution: Enabling multiple model variants to run in parallel (for canaries, A/Bs, and rollbacks) across our image-generation and related codebases.
- Developer Experience & Documentation: Creating templates, examples, and guidance; offering supportive, high-signal communication so others can adopt libraries confidently.
- Reliability, Observability & Cost: Instrumenting services with metrics/logging/tracing, setting SLIs/SLOs, and optimising inference performance and spend.
Primary Responsibilities:
- Productionise research models: refactor, test, containerise, and integrate them into the monorepo for scalable reuse.
- Build and maintain inference services, SDKs, and shared libraries that standardise pre/post-processing and interfaces across variants.
- Create multi-variant runners and rollout frameworks (feature flags, canaries, A/B testing, automated rollbacks).
- Establish CI/CD workflows, artifact management, and reproducible builds for ML services and model assets.
- Add robust observability (dashboards, alerts) and reliability practices (load tests, chaos/resiliency checks).
- Optimise inference (batching, caching, quantisation/compilation, hardware utilisation) to reduce latency and cost.
- Partner with researchers and product engineers via code reviews, pair sessions, and clear documentation to accelerate adoption.
- Drive good engineering hygiene in the research codebase: testing strategy, dependency management, and de-duplication across multiple model variants.
Unsere Erwartungen an dich
Qualifikationen
You’re probably a match if you:
- Have strong software engineering fundamentals and excellent Python skills; you’re comfortable turning notebooks and prototypes into production-grade services.
- Have shipped ML systems in production (containers, APIs, CI/CD), ideally within a monorepo environment.
- Can read research code and refactor it into clean abstractions with stable, well-documented interfaces.
- Understand service reliability and observability (metrics, tracing, logging) and how they apply to ML systems.
- Communicate clearly and empathetically—especially when guiding others to adopt libraries and best practices.
Nice to Have:
- Familiarity with model-serving/optimisation tooling (e.g., ONNX, TorchScript, Triton, quantisation).
- Background with multimodal/image generation stacks or LLM-adjacent tooling (not the core focus, but helpful).
- Knowledge of MLOps practices (model registries, artifact stores, dependency/version management).
Erfahrung
- Bring cloud experience (AWS a plus) without needing to be a deep specialist.
- Java experience (or JVM ecosystem) for services that integrate with ML components.
- Experience with experimentation platforms (feature flags, A/B testing) and safe rollout strategies.
Benefits
Gesundheit, Fitness & Fun
- 🤫Ruheräume
- ⚽️Tischkicker o. Ä.
- 🎮Gaming Room
- 🧠Psychische Gesundheitsv.
- 🚲Fahrradabstellplatz
- 🎳Team Events
- 🙂Gesundheitsförderung
Work-Life-Integration
Essen & Trinken
Mehr Netto
Job Standorte
Themen mit denen du dich im Job beschäftigst
Das ist dein Arbeitgeber
Canva Austria GmbH.
Wien
Empowering the world to design by Visual AI: By making complicated tech simple, Kaleido strives to enable individuals and businesses of all sizes to benefit from the recent advances in visual AI. Our tools simplify and accelerate workflows, foster creativity, and enable others to create new products. Since 2021, we are part of Canva. Our mission at Kaleido as part of Canva is to empower the world to design and since Canva launching in 2013, we have grown exponentially, amassing over 100 million monthly active users across 190 different countries and a team of over 3,000 people… and the best bit is that we’ve only achieved 1% of what we know we’re capable of.
Description
- Unternehmensgröße
- 50-249 Employees
- Gründungsjahr
- 2018
- Sprachen
- Englisch
- Unternehmenstyp
- Startup
- Arbeitsmodell
- Hybrid, Onsite
- Branche
- Internet, IT, Telekom