Ship It! ā€“ Episode #87

Why we switched to serverless containers

with Florian Forster, co-founder & CEO of Zitadel.com

All Episodes

Last September, at the šŸ‡ØšŸ‡­ Swiss Cloud Native Day, Florian Forster, co-founder & CEO of ZITADEL, talked about why they switched to serverless containers. ZITADEL has a really interesting workload that is both CPU intensive and latency sensitive. On top of this, their users are global, and traffic is bursty. Florian talks about how they evaluated AWS, GCP & Azure before they settled on the platform that met their requirements.

Featuring

Sponsors

Fastly ā€“ Our bandwidth partner. Fastly powers fast, secure, and scalable digital experiences. Move beyond your content delivery network to their powerful edge cloud platform. Learn more at fastly.com

Fly.io ā€“ The home of Changelog.com ā€” Deploy your apps and databases close to your users. In minutes you can run your Ruby, Go, Node, Deno, Python, or Elixir app (and databases!) all over the world. No ops required. Learn more at fly.io/changelog and check out the speedrun in their docs.

The Changelog ā€“ Conversations with the hackers, leaders, and innovators of the software world

Notes & Links

šŸ“ Edit Notes

Chapters

1 00:00 Intro
2 01:14 Florian Forster
3 05:02 The Kubernetes dance
4 07:49 "Person for everything no one else takes care of"
5 09:11 Zitadel
6 12:43 Double-dip traffic
7 15:27 Zitadel's tech stack
8 18:04 Terraform
9 19:18 Sponsor: The Changelog
10 21:03 For the cloud service?
11 28:39 Why Cloud Run?
12 33:14 Pops and regions
13 38:49 What's the CDN like?
14 40:31 Zitadel's price model
15 43:40 Zitadel Cloud
16 48:28 The "Oh f**k" moments
17 52:02 How to know there's a problem
18 54:27 Florian's 2023
19 58:31 Upcoming for the cloud
20 1:02:08 Florian's one key take away
21 1:07:06 Wrap up
22 1:07:29 Outro

Transcript

šŸ“ Edit Transcript

Changelog

Play the audio to listen along while you enjoy the transcript. šŸŽ§

One of the talks that I didnā€™t have time to watch in-personā€¦ I think it was while I was giving my talk - this was Cloud Native Day in 2022 - was Florianā€™s talk, ā€œWhy we switched to serverless containers.ā€ I have re-watched it on YouTube, twice. Not because I was preparing for this, but because I really liked it, you know? So Florian, welcome to Ship It.

Well, thank you for having me.

What is the story behind the talk? What made you want to give it?

It basically boils down to we have been huge fans of Kubernetes so far. And I still really like the ongoing effort behind Kubernetes. But while running a cloud service that should be kind of scalable, we hit some limits around the quickness of scalability in Kubernetes, and we rethought our stack to better adhere to that, and to kind of get better scaling efforts and better cost profiles as well. Because thatā€™s also a side. So the economical side was also kind of a driver to that.

Yeah, yeah. So when you say serverless containers, what do you mean by that?

Well, we labeled it that way because we thought, ā€œOkay, Citadel is being built like a Go binary, crash it into a container, you can run it basically everywhere.ā€ And thatā€™s what our customers do. They use normally Docker Compose, or Kubernetes. And we wanted to keep that availability around, so that we actually eat our own dogfood, and not just create something new for our internal purpose, but instead rely on the same things we tell our customers. And so serverless containers is kind of the definition - in our head, itā€™s basically a platform where you can run some kind of OCI image. And you could call it native, you can call it AWS Fargate, you can call it Google Cloud Run, whatever fits your poison. I mean, it could even be like fly.io, or something like that. Basically, plug in a container and it should be scalably being run. Thatā€™s kind of the definition we made right there.

Okay. Okay. So the container part - that is very clear. But why serverless?

Because we donā€™t want to handle and tackle the effort of running the underlying infrastructure in general. Because, yeah, if coming from Kubernetes, you are well aware you need to handle nodes to some degree, even with things like GKE Autopilot, there is still an abstract concept of compute nodes behind it. And the whole, letā€™s say undertaking with Cloud Run and AWS Fargate and all of those - you basically plug in your container and tell it ā€œPlease run it from 0 to 1000ā€, and you donā€™t need to care whether itā€™s one or 100,000 servers beneath. Thatā€™s why we call it serverless to some degree, even though the word is absolutely ā€“ itā€™s wrong. [laughs] I donā€™t like the word serverless, to be specific.

Okay. So if you could use a different word, what would you use, instead of serverless? What do you think describes it better?

You could call it container as a service, to some degree, I thinkā€¦ Although thatā€™s kind of plagued by the HashiCorps, Nomad, and Kubernetes, and Docker Swarmsā€¦ Because people tend to think that a managed Kubernetes offering its container as a service. And I donā€™t like that nuance, but thatā€™s a personal opinionā€¦ Because I still, to some degree, see the underlying infrastructure. And yeah, serverless kind of reflects well on - yeah, there is no server to be managedā€¦ But still, you can use your funny old OCI image and just throw it in and it will keep it working.

Yeah, yeah. So just by going on this very brief conversation, my impression is that youā€™ve been doing this Kubernetes dance for quite some time. Right?

Yesā€¦ [laughs]

Youā€™re not in this for a few months and decided you donā€™t like itā€¦ Or youā€™ve tried it and you said, ā€œYou know what? This is not for us.ā€ Youā€™ve been using it for a while, because you have a very deep knowledge about the ins and outs of how this works. So tell us a little bit about that.

Well, I mean, the first contact points with Kubernetes were around where the kube-up script was still there, by kubeup.sh.

Wowā€¦ Okay. I donā€™t remember that thingā€¦ So thatā€™s a really long time. Okay.

[05:43] Yeah. It was way pre Renshaw, pre K3sā€¦ Yeah, even OpenShift was like the classic OpenShift at that point in time. Yeah, at the company I worked at that time we had a need for an orchestration system to run some things in a scalable fashion, and we started poking into the Kubernetes ecosystem, because everybody was kind of hypedā€¦ ā€œYeah, yeah, Google is behind it. Borg is being replaced by Kubernetes.ā€ Everybody was talking in that way, and I thought itā€™s worth looking into it. And I liked some thoughts, but still, until today, it kind of gets more enterprisy, and thatā€™s a great thing if you are an enterprise. But if youā€™re a startup, itā€™s too much abstraction, too much things to care for. It feels not like a hands-off operation. I mean, even just running observability in Kubernetes is like, ā€œYeah, pick your poison.ā€ There is like 20 different ways of doing things, and why should I even have to need to care for that too much if I just want to run like one container? Please run that container for me. And so yeah, thatā€™s where my personal change from Kubernetes to something else started. But still, I totally like what they do on their end, even though a lot of complexity is involved.

So I think we finally did it. Weā€™ve found the person with 20 years of Kubernetes experience, right? Everyone thinks they donā€™t exist. Weā€™ve just found him. [laughter]

I definitely aged like 20 years around Kubernetesā€¦

Because of it? I see. Okay, okay.

Yeah, definitely. I mean, at one point, funnily enough, we decided to switch to Tectonic, like from CoreOS; it was at that point still aroundā€¦ With their self-hosted Kubernetes control plane, like running the kubectl plane in containersā€¦ It was a great thing to do. And as well Etcd. But yeah, it had some nifty tweaks and problemsā€¦ And you aged quite well in that case.

Okay, okay. So what is it that you do now? What is your role?

Iā€™m not sure what the translation into English is, but letā€™s call it person for everything that nobody else takes care ofā€¦

I see. Janitor? No. Handyman? Noā€¦ [laughter]

I label myself as CEO, though still I feel more like a CTO-ish, head of DevOps-ish kind of guy, butā€¦

Whatever. Yeah. Whatever needs to happen, you know, youā€™re there.

Exactly. I do a wide range on the business side of things, the overall vision on how we want to shape Citadel, and also the things - how can we ease some stress in our operations part of the gameā€¦ So normally, well-opinionated about many things, even though Iā€™m not all the times able to talk in the whole depth.

Because time, right? And time constraints. You canā€™t be everywhere, doing everything all the time, at 120%, right? You have to pick your battles.

Exactly. I mean, if you asked me whether I like Go generics or something like that, I need to resort to the answer ā€œI have not experienced them yet, because I havenā€™t had time to look into it.ā€

Okayā€¦

My engineers will tell you a different story, but Iā€™m not opinionated on that end.

So what does Citadel do? And maybe Iā€™m mispronouncing it, but Iā€™m sure that you will give us the official pronunciation.

I mean, we call it Citadel as well, so itā€™s totally fine. And the logo kind of tries to reflect the word origin, because it comes from the French way of building fortresses, with the star designā€¦ The place where you commonly see this is in Copenhagen, for example; thereā€™s still like the fortress with the star design. That would be called Citadel, but in French, but nobody cares. So Citadel in English.

[09:45] What we basically try to do is we want to bring the greatness of zero, like a classic closed source proprietary cloud service, with the greatness of key cloaks, run yourself capabilities, and we want to combine them into one nice, tidy package, so that basically everybody who has at least a heart in engineering can solve some of the problems around the identity game in general. So that includes things like ā€œI want to have a login site, I want to have authorization, I want to have single sign on, with different IDPs, I want to have tokens that I can send through the worldā€¦ā€ And everything like that. So basically, you could call it a turnkey solution to solve user management and authentication in general.

Okay. Okay. So thereā€™s one word that comes to mind when you see that. Actually two. CPU and hashing. [laughter] So from all that description, all Iā€™m thinking is ā€œI hope your CPU is fast, and your hashes are even faster.ā€ But donā€™t skimp on the cycles. Right? Like, make sure you do enough iterations. Okay? No cheating, please. Okay.

Yeah, I mean, it comes down to hashing. We rely on Bcrypt normally, and yes, we need many CPU cyclesā€¦ But also, thereā€™s like a second thought, that if you do signing tokens - so thatā€™s also quite exhaustive for CPUs nowadays. And so yes, we use a lot of CPU to make that happen.

But actually, the stress is somewhat alleviated down to the future, because the pass key concept and [unintelligible 00:11:30.03] in general, since that relies on public-private key cryptography, it reduces quiet well on our end the amount of CPU we need to use, because RSA signature verification normally takes around one millisecond. So itā€™s not that much of a stress. But hashing, Bcrypt, password with 10 to 12 of iterations might be more unrivaled of like 800 milliseconds to 1,000 milliseconds, around that [unintelligible 00:11:57.27] if you run like two to four CPUs.

Yeah, yeah. Okay. Do you use GPUs for any of this?

No. At the moment, no. I mean, it comes down to a certain degree that our cloud provider does not really allow for thatā€¦ And on-premise environments also are oftentimes really restricted in the kind of GPU they have aroundā€¦ And so we try and avoid for Citadel too much of integration and depth there. So we still rely on CPUs. But there might be a time where we want to retrain some analytics model with machine learning, and so we are looking into that space. But itā€™s not yet there, to be honest.

So most of the workloads that you run, I imagine theyā€™re CPU-bound, because especially of the hashing. But I also think network is involved. So you need to have a good network. Iā€™m gonna say good network - not high throughput, but low-latency network. Can you predict these workloads? I mean, people sign in whenever they need to sign in; they donā€™t sign in all the time. Right? So can you tell us a little bit about the workloads specifically that you run?

Yeah, we have a common pattern called double-dip traffic. So we often see like traffic starting like seven oā€™clock, until twelve, and then from one to around seven in the evening. So thatā€™s the major traffic phase. But if you spread that across the globe, you see certain regions ingest traffic during certain times, but globally, overarching, itā€™s basically a big white noise line. Itā€™s like more flat if you watch all the regions, but if you really do a region, you really see like double-dips happening all the time.

[13:45] And yeah, having a network that can be easily and fast scalable, and really provides low latency to our customers, because they use our API, our login page, is kind of important to the service quality. And so we internally set our goal to ā€œOkay, letā€™s try and keep latency below 250 milliseconds for 80% of the calls.ā€ Thatā€™s kind of the SLO weā€™re aiming at, but thatā€™s not all the times possible. I mean, down under - yeah, you get bad latency. You get strange latency if you go for South Americaā€¦ So yeah. Normally, it does not ā€“ it comes down not to the problem being Citadel, who you can easily run at the edge, but rather how can you move data along the journey, and thatā€™s kind of the thing that holds you back most.

Yeah, thatā€™s right. Especially if youā€™re in like SAML, or something like that, where you yourselves need to talk to another provider, which itself may have variable latency at specific periods of timeā€¦ And then you can only guarantee what you get, basically. You canā€™t make (I donā€™t know) GitHub login, or Twitter login, or whatever youā€™re using, faster than it is.

Yes. Itā€™s bound to external services. And in our case, for the storage layer we mainly rely on CockroachDB, since they can move around and shift data closer to data centers that they are actually being used. But still, the data is the most heavy thing right there if you want to reduce latency.

Okay. What does your tech stack look like? Like, what do you have? What is Citadel running on?

Basically, we use internally Cloud Run for our offering, so Google Cloud Run, because we can easily scale during traffic peaks, as well as CPU peaks, because Google will scale Cloud Run containers as well based on CPU usage. That fits our narrative quite well. And also, the ability to scale to zero in not frequently used regions is really a thing we really like.

Below that, we use CockroachDB dedicated, as like their managed service offering of Cockroach. In the past, we actually did run Cockroach on our own, but we figured the amount of money and time we need to invest basically is being taken by Docker if they run it; itā€™s even economically more value, because they include their license quite nifty into their cloud offerings, soā€¦ Basically, you have multiple reasons to do that.

And upfront, we use Google CDN and Google Global Load Balancer with Cloud Armor to mitigate some of the risks around DDoS, and rate limits, and stuff, because we see malicious behaviors all the timeā€¦ And yeah, the stack is really lean in that regardā€¦ Because the only thing we can append to that from an operational perspective is like we use Datadog to some degree for observability purpose, and also Google Suiteā€¦ And weā€™re kind of torn in between, because - yeah, why should I send logfiles to Datadog if they already have ā€“ I already have them in Googleā€™s cloud offering. Why should I pay twice for it? But there are some catches in Cockroachā€™s offering that you only can use Datadog to monitor the database, and so weā€™ve kind of bound to Datadog. So itā€™s not so funny in between state thereā€¦

Interesting.

ā€¦because we try to reduce our amount of third-party processes whenever possible on that end.

Thatā€™s kind of the stack we use.

I like that. Itā€™s simple, but it sounds modern, and itā€™s complicated in the areas where itā€™s sometimes by design; as you mentioned, this coupling between CockroachDB the dedicated one, which is managed for you, so youā€™re consuming it as a service, and thereā€™s a coupling to Datadog. And Iā€™m sure that choice was made for whatever reasonsā€¦

Yesā€¦ [laughs]

Okay. So thereā€™s Google Cloud, thereā€™s CockroachDB, thereā€™s Datadog. Anything else?

[18:10] We use TerraForm to provision infrastructure, and we use GitHub Actions mainly to do that. We have still some stuff in TerraForm Cloud, but weā€™re constantly migrating into GitHub Actions and private repositories, because it fits better with our flow. 80% of the company knows how to deal with Git, and so we ā€“ we did a lot of classic GitOps in the past, with like Argo and Flux and all of those toolsā€¦ We feel comfortable with Git, and so we try to shift as many things that we can into into that space. So from a source control perspective - yeah, itā€™s GitHub with us. But thatā€™s about it on that end. Thereā€™s more chat tools, and things around, or auxiliary services, but the core operations really revolves around that stuff.

[19:03] to [21:02]

How much of this do you use for the actual Citadel software as a service, where itā€™s all managed for you? I think youā€™re calling it the cloud now?

[unintelligible 00:21:11.10] versus building the binaries that you make available to users? Is it like most of this is for the cloud service, or what does the combination look like?

I think 80% is for the cloud service, because the whole CI side of the story comes down to GitHub Actions, some Docker files, and some custom-built shell scripts, because protoc and gRPC tends to be a little bit nifty how to build clients with gRPC. So that basically comes down to that from a CI perspective. And we can basically run all our components we have - like, we have a management GUI in Angular, we have like a cloud GUI in Next.js, we have like the backend of Citadel, which is written in Goā€¦ And we can cram everything into GitHub and GitHub Actions, into a dockerfile, basically. And thatā€™s a process that is constantly evolving to reduce drag for additional committersā€¦ Because it kind of is ā€“ if you donā€™t make that understandable, itā€™s a high entry bar to commit things.

Yeah, of course.

[22:23] Yeah, and you need to reduce the drag. I think an honorable mention on that end is like Cypress for each retest stuffā€¦ Because we test our management GUI with Cypress, which basically passes on to our API, so we can do that end to end from our perspectiveā€¦ Yeah, and then thereā€™s like automated things, like Dependabot, and CodeQL, and static analysis tools to make the security rightā€¦ But thatā€™s about it, because we think the tool drag - I call it often this tool dragā€¦ I donā€™t like having too many tools around to do the same job, because you lack focus if you do that.

Of course. Of course, that just makes sense. Yeah, that cognitive overload of having ten ways of doing the same thing, based on the part of the company that youā€™re working inā€¦ Okay, okay. So how many engineers are there for us to better understand the scale of people that touch this code, and work on this code, and help maintain it? So thereā€™s the company, and then the community, because thatā€™s also an important element.

Yeah. Currently, itā€™s like eight dedicated engineers, like software engineers from our end, working on Citadelā€™s code. Some of them also working on things like the Helm chart, and stuff, because itā€™s mainly our engineering staff who does thatā€¦ And then we have like 20, 25-ish external contributors across the field who do a variety of tasks. Because we have also some separated packages from Citadel. Like the OpenID connect library, for example, is not in Citadel; we use it on our end, but we needed it to create, for example, a Go library for OpenID connect, and we wanted to have it certified, so we needed to do like an extra leg on that and. And we have separate maintainers on that end from Citadel, because many people are using [unintelligible 00:24:17.17] without our knowledge, so to say.

So itā€™s multiple projects, multiple contributors, but the stack beneath is kind of the same. And with us itā€™s basically the eight engineers who work on that, even though the company now has roughly 15 employees since we started to work more and more on technical writing things, API documentationā€¦ Because otherwise, itā€™s not nice to work withā€“

Oh yes, the docs. Oh, yeah. Tell me about it. Yeah. Okay, okay. [laughter] Oh, yeah.

Docs is like the conflicting major pain point all the times. And I donā€™t say that lightly, but I think that you can use that on any project - as soon as itā€™s kind of open source, itā€™s hard to get good quality, consistent reading docsā€¦ But thatā€™s a topic - for example, we want to invest quite heavily over the coming few months, because we really feel like if you want to engage with developers and engineers, you need to have proper documentation. Otherwise, it feels like the impediment is too high.

Yeah, for sure. And I canā€™t think of a better way to show the users that you care, you care about your product. I mean, very few would dig into the code and say, ā€œWow, this is amazing code. Iā€™m going to use this product, because itā€™s amazing.ā€ How often does it happen?

But put like a nice docs site out there, easy to search, easy to understand, with good flows, and people go ā€œWow, they really spent some timeā€, even though the code may not be that good, but the docs are, and then the perception is ā€œThis is amazing.ā€

[26:02] Yeah. I mean, I have recently heard the term ā€œContent is oxygen for the communityā€, and I think thatā€™s also applicable to the documentation side of things, because itā€™s not only outreach content, but also, and rather important, the documentation side of things. Because even if you write like the best blog about something, at one point you will link to your documentation, and if that link is not nicely, tidy being done - yeah, it breaks the experience. And so if I need to point out one specific thing we need to improve over the coming months, itā€™s really docs. They need to have a clear flow. ā€œWhere should I start? From where can I go to what?ā€ So they need to reflect the user journey, basically, and thatā€™s kind of the biggest rework we will do, is restructure docs to better appeal for that.

Yeah. Okay, that makes a lot of sense. Yeah, for sure. For sure. Iā€™m just taken aback by how simply youā€™ve put that. Itā€™s a very high problem, right? And it all boils down to this. So if you donā€™t get this, then forget everything else, about like making them easy to understand. I mean, thatā€™s important, but what are the flows? Where do you enter? What are the entry points? Where do you drop off? What happens next? What is the follow-up? What is the story that you get when you go to the docs? And if you just get referenced ā€“ by the way, thereā€™s so many types of actuallyā€¦ Four, as far as I can remember; thereā€™s guidelines, thereā€™s references, thereā€™sā€¦

Examples, oftentimesā€¦

Examples, exactly.

API documentationā€¦

Oh, yeah.

In our case, we split out the whole self-hosting part, because in our cloud itā€™s not applicable. But if you want to run it on your own, you need like production checklists, examples to deploy it to x, y, z, how to configure all the nifty details, how to configure a CDN, how to configure TLSā€¦ I mean, thereā€™s a whole array of topics just for the self-hosting stuff. And so you kind of need to figure out that flow, and it will take you a lot of time to do that. But if you figure it out, it will get beautiful. The only thing you can break at that point is basically to style how you write content, in what kind of language, and thatā€™s especially difficult if you have non-native speakers and engineers; they tend to write different documentations, as dev rels, or as content marketing guys or gals, because itā€™s just a different way on thinking of itā€¦ Thatā€™s the second thing to get right there.

For sure. For sure. Okay, so Iā€™d like us to come back now to your talk, because one thing which I really liked in that video, and in your presentation, is you talked about why you havenā€™t chosen Microsoft, and why you havenā€™t chosen AWS. Itā€™s not like, you know, ā€œWe havenā€™t even looked there.ā€ You did try them, you did consider them, but there were certain things which wouldnā€™t work for you. So tell us a little bit about that, how you ended up with Cloud Run, which I donā€™t think it was your first choice, but basically, you ended up there because of your requirements.

Yeah, the first and most prominent thing that struck us was kind of having end to end HTTP support. Because we provide gRPC APIs to our customers, and they need HTTP/2. And while verifying that with all the different offerings, it was kind of hard to either get proper documentation, whether they support it or not, or they oftentimes only supported it from their CDN to the customerā€™s site, but not in the upstreamā€¦ And so that was kind of really like bogging us down on that end.

[29:51] I mean, we could have chosen the route and say ā€œOkay, we do not offer gRPC, but only gRPC web and REST, because we supply that as well. But we really wanted to have the HTTP capabilities, because we think at one point there is a unique opportunity to be taken to use streams and stuff for identity-related things. So if something changes, we can notify you immediately, which can be an interesting way of thinking of it. And it reduces latency quite a lot. I mean, our own measurement states thatā€™s reduced seven milliseconds each call purely down to JSON serializationā€¦ Which is not like a bad thing, but itā€™s seven milliseconds.

Yeah. It adds up, right? Seven milliseconds here, three milliseconds thereā€¦ Before you know it, itā€™s a minute.

No, really, if you have microservice architectures ā€“ I mean, if you have like five services cascaded, and every time they call like our introspect endpoint to validate some tokens, it adds up. Itā€™s 50 milliseconds only serialization at that point. But that was just the decision they made there, and really, the major breaking point was the HTTP stuff. It was just too confused, too not clearly stated. We started poking around, we saw that eventually you can make it work, to some degree. For example, Azure - whatā€™s it called now? Azure App Container Instances, I thinkā€¦ They now use Envoy as proxy, so you now get HTTP/2ā€¦ But as soon as you want to use the web application firewall - yeah, well, youā€™re in a bad spot, because that thing does only support HTTP/2 to the front, and not to the back. And the CDN as well. So it always did come down to friction on that end. And so yeah, we choose Google Cloud Run, exactly. Thatā€™s one of the major reasons we chose Cloud Run, even though we donā€™t like some of the limitations with Cloud Run. There are some which we donā€™t like.

Tell us about them.

I mean, I still to the day not fully understand why Google Cloud Run needs to use their internal artifact registryā€¦ Like, you need to push your images into Googleā€™s registry, and from there on out, you can fetch it in Cloud Run. I donā€™t know why that decision was being made. There might be a technical reason to that. I mean, you could argue on availability; that would be a ground. But I donā€™t like that fact. And the other thing that really is kind of a bad thing is if you want to use VPCs, you need to use the VPC connector, which now can be edited. I guess they released that like one month back, or something like that, but you need to have basically VMs in the background that handle connectivity from your Cloud Run to your VPCs. And since we use Cockroach, that traffic passes through a VPC, and we use like [unintelligible 00:32:47.18] gateways, so we pass anyway traffic through that, because we want to have control over what traffic leaves our site, and stuff. And that VPC connector thing is always and so like, yeah. Itā€™s there, I donā€™t like it, because it scales not down. Itā€™s just scaling up, and then I have like 10 VMs running in the background, doing nothing. But yeah, itā€™s kind of a thing I donā€™t like. But other than that, itā€™s a great tool.

Thatā€™s right. One thing that maybe we havenā€™t done as a good job to convey this is that your service is global. So when youā€™re saying VPC connect, you donā€™t mean just in one region or in one zone; you mean across the whole world. So how many pops do you have worldwide, where Citadel runs?

We have like a core pop region. Itā€™s like three regions we run constantly, that comes down because we run our storage there as well. And sometimes if we see different traffic profiles, we start regions without storage to them, just to get some of the business logic closer to customers. So that can range normally from three to nine regions during normal operations. But since Googleā€™s internal network is quite efficient from our perspective, and their connectivity is really great, we donā€™t need to spread it to more regions than that normally.

[34:13] We did some internal experiments, we built like a small TerraForm function where you basically can throw in a list of regions you want to deploy, and it will basically deploy to 26 regions in like one to two minutesā€¦

That works really well. But you get strange problems if you do that, because sometimes you want to have like hot regions, because your application is anyway running, it can serve traffic quite easilyā€¦ If you have to cold-start a region, it always takes a few milliseconds to do that. And itā€™s not like a big thing, but it can influence customersā€™ view on your service. Because if you hit the login page and it takes like two seconds to get everything spinning, and database connectivity set up, and everything, it gives you some drag. And so weā€™re trying to keep hot regions, as we call them.

Yeah, thatā€™s crazy. Like, you say that two seconds is slow for a whole region to come up. Itā€™s like, ā€œWhat?!ā€ [laughter] Like, try booting something; it will take more than two seconds. Anything, really. Wow, okayā€¦

I feel itā€™s like engineering ethos that you might at some time over-engineer certain thingsā€¦ But still, it feels right to do that, because itā€™s more easy to just scale up an existing region and throw some more traffic into that. You can easily steer that around. The thing you most of the times will miss out is basically 50 milliseconds. There is edge cases with different regionsā€¦ So if you live like in Australia - yes, we donā€™t have like an immediate region in your vicinityā€¦ But yeah, thatā€™s a matter of - if you have enough traffic, you will actually open a region at one point, and then you try to keep it hot as long as you can. And I always think the classic engineering decision that comes down to that is the same thing that Cloudflare and Fastly are constantly arguing aroundā€¦ I mean, the last time I checked, Cloudflare was still ā€œLetā€™s build small regions across many placesā€, and Fastly was like, ā€œNo, letā€™s build like huge data centers, with huge compute to them.ā€ Itā€™s a matter of what your service needs to decide that, and we decided that ā€œYeah, two seconds feels bad.ā€

Yeah. But itā€™s interestingā€¦ Sometimes routing to a region which is further away, sending a request to that region can be faster based on your workload. And even though your workloads are super-optimized, for something to be up and ready in two seconds, thatā€™s just crazy. Try doing that with Javaā€¦ [laughter] Right? Or something which is like slow to start. Not picking on Java, but itā€™s known for slow starts, which is why you wouldnā€™t stop it. And that would be like a no-starter; like, you canā€™t even consider that for what youā€™re doing. Maybe GraalVM does things better. But JVM, itā€™s slow by design, because it runs optimizationsā€¦ I mean, a lot of things need to happen. Again, nothing wrong with that, but not suitable for this workload.

I mean, itā€™s even small things involved into getting like fast startup latenciesā€¦ One big drive ofā€“ we hypothesize - we donā€™t really have evidence, but we hypothesize - is that the image sizes of your containers influence that quite a lot. Even though I think Google does quite a lot of magic in caching things in Cloud Run to scale it quickly. But nonetheless, we see that bigger images take more time. So for example, we have a documentation page built with Docusaurus, and we normally use Netlify to deploy those things, and we are now currently testing to move that to Cloud Run as well. And that container is approximately 500 megs in size, because of the images and stuff, and it takes more time to start. So it takes like three to four seconds just to get like the node server started, even though everything is pre-built, so itā€™s not like we are compiling stuff on the fly. Itā€™s really like start the Node server with static assets.

[38:14] Okay, okay. Yeah, youā€™re right, youā€™re right; that can make a big difference as well. So a few seconds is not bad, right? Especially if we have like a blue/green style of deployment, where ā€“ and I know that Cloud Run supports that. So youā€™re not taking the live version down, and thatā€™s okayā€¦ Something that needs dynamic requests - for that itā€™s a bit more difficult, right? Because it needs to service it, it needs to keep the connection openā€¦ Thereā€™s like a couple of things happening, rather than just static ones, where you can just cache them, use a CDN. So you said that you are using a CDN, right?

Okay. Okay. How is that like? How are you finding the Google CDN? Because I havenā€™t used it in anger; I mean, only small projectsā€¦ How does that work in practice?

I actually quite like it. One of the things we like the most is that you can cache assets across multiple domains. So for example, each of our customer has their own domain name, and our management GUI is built with Angular, so we have a lot of static assets to thatā€¦ And if one customer accesses that data in one region, we can cache it for basically every customer. And thatā€™s nice, because you basically can ignore the name as the host name, and instead just cash the files. And thatā€™s a feature I have not easily found in CloudFlare or Fastlyā€™s offering. I mean, you can always make it work to some degree, but that was basically ā€“ yeah, just input a validation rule into the Google CDN and it will take care after all that stuff.

And the overall strategy with Googleā€™s pricing is more beneficial to our end, because we basically only pay usage, and we are not feature-locked. And with the Cloudflares, Fastlys, and everybody, you basically are always feature-locked until you get to their enterprise offering. And at that point - yeah, the cost is quite steep to be paid at that endā€¦ Even though they have great offerings, but it feels wrong to spend so much money on a CDN, when itā€™s just ā€“ itā€™s basically caching some static assets. I mean, itā€™s not doing the heavy-lifting, itā€™s more quality of life improvement, I would call it.

I was reading something about this as well, enterprise features, feature locking, things like that. It was your blog, where he said you charged for requests, right?

[laughs] Yes.

See, I have done a bit of research; not too much, but I did notice that, where you mentioned thatā€¦ And to me, that is very reasonable. You donā€™t have to upgrade to higher price tiers just to unlock certain features. I mean, why? Does it cost you more? I mean in development time sure, but you donā€™t get the full experience of the service. Per seats - again, that pricing can work in some cases. So what influenced your decision to do that? I thought that was very interesting, for all the right reasons.

We thought long and hard about our pricing, so many times in the pastā€¦ We even had like a feature locked model, closely to what Cloudflare doesā€¦ And what does not reflect well in the security area is if you want to provide your customers with a security service, you should give them the means to have a secure application, and not to tell them, ā€œNo, if you want to have 2-factor, you need to pay extraā€, because that kind of defeats the purpose of having an external specialized system of handling the security in the first placeā€¦

And the second thing there is like if you price by identities, customers will stop creating identities at one point, if they can choose. And we wanted to remove that sensation by telling them, ā€œHey, store as many things as you like, do as many things as you like. The only thing we want to have from you in return is we need to be able to finance our infrastructure to some degree, and so it comes down to pay us for the usage.ā€

[42:20] Thatā€™s really what it boils down toā€¦ Because it feels like a nice trade-off, even though - and I can be honest on that, and itā€™s during sales meetings - it can sometimes be a problem or impediment, because people still think in users. ā€œI want to have like a million users. I want to have a price for a million users.ā€ And I mean, we have a lot of data, we can do cost estimations for that; itā€™s not like a big problem. We even do over H deals where we say ā€œOkay, weā€™ll give you like 10 million requests a month for a price XYZ, and we will use it for 12 months, and we will check after five months how it was workingā€, because we want to reduce that friction out of the equation. But itā€™s just a matter of different strategy, and we are committed to that end, because it feels like the right thing to do, even though it has some challenges.

Yeah, for sure. For sure. Yeah, I mean, to me, that sounds a more sensible approach, a more honest approach, a more open approach. Everything is out there for you to use. Thereā€™s like one requirement that we are able to support all your requests, and we are able to give you the quality of service that we know what you wantā€¦ And these are the SLOs, and thatā€™s what it will cost. Okay. How long have you been running the cloud service for, your cloud service? Six months, 12 months?

Now, looking at the date, itā€™s like seven.

Seven months. Okay.

Yeah. The thing we call Citadel Cloud now is now seven months in ageā€¦ But we had a service we called Citadel v1, with kind of a different sensation to it, with the old pricing I just mentionedā€¦ And that was started in mid-2021. But we learned so many things across that journey that we needed to reconsider some of the things, like pricing, deployment strategy, locations where we deploy, because customers actually care sometimes about thatā€¦ The API needed to be reshaped to a degreeā€¦ And so yeah, it comes down to an evolution of Citadel, like from version one to version two, and our cloud service changed as well. And so the new service is basically seven months in age right now.

Okay. What are the things that worked well for the cloud service in the last seven months? Good decisions, that proved to be good in practice?

I think itā€™s not only directly the cloud service, but the overall change in our messaging, what we actually want to sell, and why we recommend that you use our cloud service - that message is being picked up better since like the seven last months. So thatā€™s a thing we constantly improved. So many people now use the free offering we have, because it provides already a lot of value, and we are even considering increasing the amount of things we give you, the amount of requests as well, to get developers an easygoing free tier that they can actually start building software. Because nobody really likes to run things. And I think thatā€™s the most ā€“ letā€™s call it the biggest change I experienced so far in behavioral thingsā€¦ Because everybodyā€™s always shouting, ā€œI want to run system XYZ on my own hardwareā€, but in the end everybody turns to some kind of free hosted offering, because everybody just knows ā€œOh, no, I donā€™t want to take care of backups. Oh no, it just runs. Oh no, I need to start it again.ā€ So ā€“

ā€œUpgrades? Again?! Oh, my goodnessā€¦ā€ When did you last upgrade your phone? Serious question.

My phoneā€¦ Iā€™m quite pedantic, so I will catch up on releases in one to two days. [laughter]

[46:14] Okay, thatā€™s a great answerā€¦ But for me, the updates just happen, right? I mean, unless itā€™s like a major update, your applications on your phone - they just update. Itā€™s not a problem that people think about, or should think about. So if you run it yourself, guess what? You have to think about that. And then you say, ā€œOh, dammit, I want this free auto update feature. Why donā€™t you give it to me?ā€ Okay, well, itā€™s not that easy. I mean, the easiest auto-update - like, delete it, and then deploy it again. And then youā€™re good. But people donā€™t want that. So the point being is, you want the service, because this stuff should be seamless, and someone needs to put in the effort for it to be seamless every single time.

Itā€™s hands-off. We really call it hands-off. I mean, we take care of the TLS stuff, we will take care of updates, of backups, of rate limits, of malicious trafficā€¦ Everything is just handled for you, and thatā€™s a value I think is going great with the communityā€¦ Even though Citadelā€™s open source version is really like ā€“ there is like 99% of the things that we have in our cloud service is in the open source code, and then you can run it on your own. You can even get like an enterprise license with us to get support, and stuff. So we really encourage you to do that. But the main reason we still encourage doing that is we see many customers having special requirements on data residency, or data accessā€¦ And we always tell them ā€œWe donā€™t want to do like a managed service offering for you guys, because it feels wrong. Because we still have access.ā€ And if your reason is, ā€œNobody else should have accessā€, well then you need to run the system on your own.

Yeah. Yeah, that makes sense. I forgot, youā€™re a security, right? [laughter] And security has this very important requirement. No, sorry, I have to run this. I mean, I understandā€¦ I can pay you to run it for me, but it has to be in the specific locations, with these restrictions, andā€¦ Yeah, that makes sense.

Yeah. So thatā€™s the thing I think went great with the cloud service. So free tier is definitely a thing we will reiterate on, even though you could run it on your own. I mean, itā€™s not like there is like a feature gap, or something like that.

Okay. What about the things that you wish you knew, before we [48:29]? [laughter] The ā€œOh, f**k!ā€ moments. And that will be bleeped, butā€¦

[laughs] Yeah, there are many. Honestly, there are so many. Iā€™m going to choose to focus on the operation side of things for the first moment. Itā€™s really like, donā€™t try and build many funny things, even though there are great open source tools around, and everything is ready. Just try and relax a little and use ready-made services in the beginning. Because we thought ā€œOkay, letā€™s run our own Kubernetes, or GCPā€, and stuff. ā€œLetā€™s get more control of it.ā€ Yeah, it was going great, but the added cost and the added slowness you have while maintaining our own Kubernetes and stuff - itā€™s not worth it. So thatā€™s really ā€“ just use turnkey services to begin with. And at one point, be ready to make decisions to change that stuff to more enterprisy side of thingsā€¦ Because [unintelligible 00:49:36.04] application to Netlify, and Vercel, and calls it a dayā€¦ But I think thatā€™s only worth for the start. At one point you want to get more control, more flexibilityā€¦ You want to create rate limits, you want private IPs, you want to have like the enterprisy thingsā€¦ And you get that way easier if you start focusing on using like Google Cloud, AWS, Azure, whatever is your poison, basicallyā€¦ Just use an infrastructure provider for that.

[50:10] While reflecting back, I feel like that would have helped us decrease some of the drag along operation efforts. And as well, donā€™t run Cockroach on your own; just use the cloud service. Why not? You must really have valid and specific requirements that do not allow you to do that before you make that decision. So thatā€™s really a big thing on that end.

Other things? Yeah, our startup lesson is like ā€œDonā€™t assume things. Always validate things. Talk to your customers, talk to potential customers whether a feature is really needed.ā€ And we built, for example, too many features in Citadel. We assumed too many things in the past, and so we now strip some of the things that nobody actually needs. And you need to have like data to make that judgment, but other than that, it really comes down to ā€œCheck first whether somebody needs somethingā€, and not only one guy, but also multiple guys, and then build it, and not the other way aroundā€¦ Because you get a lot of dead code that you need to maintain, and it can have bugsā€¦ And yeah, so thatā€™s really a lesson weā€™re mentioning.

Yeah. Itā€™s a good thing that you are pruning some of this stuff. Because what usually happens - you never touch it. You add it, but then you never touch it again. And that too often ends up in some very big messes, that no one wants to touch, everā€¦ And then things just die like that, you know?

I mean, as soon as you see somebody raising a concern, or a bug over something, and you think, ā€œIs that feature even being used?ā€ And you canā€™t really validate to yourself, ā€œYes, thatā€™s actively being used.ā€ You should invoke a discussion if you want to rather remove it, because just one person is using it. Itā€™s just maintenance effort.

Iā€™m wondering, on the operational side of your cloud service, how do you know when something doesnā€™t work as expected? How do you know that there is a problem, basically, is what Iā€™m asking?

Letā€™s say I am a huge fan of using existing data to get a sensation whether somethingā€™s healthy or notā€¦ By which I mean we try and avoid active health checks; like, we donā€™t ping Citadel from the outside world and figure out whether a service is available or not. Iā€™d rather use the logfiles and throw them into some kind of analytics engine to figure out ā€œOkay, how many status codes do we have, in what variety? How many calls? Whatā€™s the latency?ā€ Because that gives you a broad understanding of the health of your service across regions. Because if an error rate is growing, and thatā€™s being tied into a region, youā€™ll normally know itā€™s more an infrastructure-related problem. If the error rate is growing across the globe, it might be more on the storage side of things, because thatā€™s a global thingā€¦ And so you get quite fast a broad sensation of how things are.

And the other hugely important thing to us is like having traces, like open tracing. We use Google Cloud Tracing for that stuff, to get a sensation on whether releases change things, or smallish changes, or A/B tests alsoā€¦ Because sometimes we create a new branch in our open source repository, and refactor some of the code, for example, get user by IDs being revoked with a new SQL statement, and then we deploy it to production to get some traffic to it, and see how latency shifts around.

So it really comes down to observability based on the data you have around, but not checking it actively, because that gives you a wrong sensation. Because youā€™re always checking happy paths there. I mean ā€“

[54:03] Yeah, true. Yeah, thatā€™s a good point.

Yeah, you can check a login page. I mean, filling in fields, pressing a button. It will work. But what happens if itā€™s not working because the user has strange extensions in the browser, because they have strange proxies in their environment, because mobile connections are resetā€¦ So it does not reflect the natural world.

Yeah, thatā€™s a good point. Okay. So weā€™re still at the beginning of 2023ā€¦ What are the things that youā€™re looking forward to? ā€¦for Citadel, for your cloud service, for what you do, for the space, for the wider security space. You can answer this whichever way you want. Thatā€™s why Iā€™m giving you a wide berth. Pick whatever resonates with you the most.

We have identified some needs for our customers who want to have more control over their general user experienceā€¦ Because currently, people use the hosted login page from Citadel, which can be white-labeled, and you can basically change everything, and customize itā€¦ But still, itā€™s like the [unintelligible 00:55:11.09] for you. Even though itā€™s a good approach, because we can include all the security policies into it, and verifications, and stuffā€¦ But there is a huge demand in the developer space for having like only authentication APIs. So they basically can send us a username, password, and we will respond by true or false. So people want to create their own login experiences, as well as their own register experiences. And thatā€™s a thing we will tackle in the next few months, in the next coming months. We want to extend our APIs so that people can build their own login pages.

And during that phase, we will also change our login, because now itā€™s built with Go, and Go HTML templates, and thatā€™s kind of not so beautiful as it could be. There might be a point where we change that to Next.js, to get like a more SDK-y approach, so that we kind of build our login with Next.js, and we will provide you the template that you can clone and create your own login without our intervention.

So Citadel might become more of a headless identity system in one place, where we just provide you the means to have different components that you can deploy, or if you donā€™t want them, you can get rid of them. So thatā€™s kind of a natural evolution path we see there.

And the other big thing we will change in 2023 is we will extend our actions concept more. Basically, actions are ā€“ you can invoke customized JavaScript codes in Citadel at certain points. Like, if a user is created, you can call a CRM system with some JavaScript code, you can fetch some information from there, and the whole action engine will be reworked that we can allow for more flexibility. So you basically can think of it like a GitHub Action; you can subscribe to events, and then execute something. And thatā€™s a thing we encourage quite heavily, even though it has a steep cost to be paid in regard of runtime security. I mean, running foreign code in a system is always not so funny to doā€¦

Oh, yeahā€¦

We do a lot of pen testing, and testing in generalā€¦ And thatā€™s the reason why you canā€™t really run it everywhere currently in Citadel, because we want to reduce the threat surface, to get our bearings whether the engine works, and everythingā€¦ So that will be a subject to be changed in 2023. So those are the two biggest, prominent points I think you will see from us on that end.

There is more underlying stuff, especially around machine learning and things, because we ā€“ I mean, Citadel is built in an event sourcing concept, so we have a lot of data availableā€¦ And we want to give our customers the option to train data, threat prevention models with their own data, to compensate for signal-based risksā€¦ I mean, thatā€™s kind of a little bit on the academic side of things, and we are working with research partners on that end. But bringing value from the data is a huge thing that we wanted to provide our customers.

[58:28] Yeah, okay. Okay. What about the cloud? Anything for the cloud that you have planned?

Some thingsā€¦ [laughs]

Some things. Great. Thatā€™s a great answer. We can move on. All good.

Iā€™m not sure much I should already disclose. [laughs]

No, no, all good. Letā€™s move on. Not a problem. All good. [laughs]

Letā€™s seeā€¦ A thing I can definitely disclose right there is the ā€“ we are strongly considering opening additional regions, because we see now where traffic is originating from, and we are considering to expand our footprint on that endā€¦ And also a thing that will land in our cloud eventually is a feature I did not disclose yet. Itā€™s basically an event API. Itā€™s a simple API; you can basically fetch everything that changed in Citadel, with the whole body to it, like first name changed to whatever, because that gives the developer a great way of backpressure processing of things that changed, so they get a proper change track of everything.

And that thing - I mean, the event API will land at one point in our cloud service, but that needs some rigorous testing to be sure that all the inner fencings of our cloud service work out well, not that customers see the wrong data, and stuff. So thatā€™s a thing. And we are experimenting with different CPU profiles in our cloud service to reduce some of the latency we see, especially during hashing operationsā€¦ itā€™s one small [unintelligible 00:59:55.17] But there is like a limitation in Cloud Runā€™s offering; you canā€™t have really high CPU containers without high memory, and we donā€™t need so much memory. Normally, Citadel uses 500 megs of memory; we donā€™t use more. The rest is being handled by the storage layer and caching.

So yeah, thatā€™s a thing that needs untanglingā€¦ So either we can use resources better, or we can somehow influence to have more CPUs in there, to get better latency. Yeah, thatā€™s an ongoing experiment. We always try to wiggle out some things.

Yeah. Okay. Well, that sounds like a very good plan for 2023. Letā€™s see how much of that actually happensā€¦ As we know, the plans are the best.

[laughs] Yeah, yeah,

Everythingā€™s gonna be amazing, and then reality hits, and then you realize that half the stuff you thought would work will be impossible. So thatā€™s my favorite - like, get it out there, see what sticks, see what doesnā€™t, keep making those changes, those improvements, drop what doesnā€™t make senseā€¦ Whatever.

Itā€™s really testing things. And there is also some discussions around a reiteration of the pricing. For example, a thing we are currently testing, as well as deciding, is whether we want to give away a domain for free, for example. Because currently, if you want to have like a production domain with Citadel Cloud, you need to pay 25 bucks a month. But we are strongly considering whether we want to provide that domain name for free to developers, because it reduces drag along their journey. I mean, they want to start poking around, they want to use itā€¦ And if our cost attached to that is not so high, itā€™s no real problem. And since Google changed some of their offering in the TLS space, you can get like customer certificates quite easily, without a huge cost profile. So a certificate would cost you only like 10 cents per month, per customerā€¦ So it feels like the right way to do it, but itā€™s not yet a done deal.

[01:02:06.03] Okay, that sounds exciting. Not as exciting as your takeaway, because Iā€™m sure that you had the time to think as we were discussing this, about the one key takeaway that you want listeners that got as far as this in the conversation with us. So honestly, this was very eye-opening for me, to see how maturely youā€™re thinking about some very hard problems, like distributing code globally, latency, and the scale that youā€™re thinkingā€¦ Youā€™re saying a few seconds is too slow. And to me, itā€™s like ā€œWhoa, what?ā€ Sometimes requests take longer than a few seconds, because it happens, because maybe youā€™re on a phone, or youā€™re in a watch, or whatever the case may be. And then your CPU isnā€™t as fast, or the cellular doesnā€™t work as well as you may think.

So to me, from what Iā€™m hearing is operationally, youā€™re very advanced. And youā€™ve tried a couple of things, and youā€™ve seen a bunch of things that didnā€™t work out very well in practice, even though the promise is there, and the marketing is working well for certain thingsā€¦ So you have a lot of like ā€“ I think street-wise, itā€™s called; youā€™re street-wise. Youā€™ve been out there, youā€™ve tried a few things, and you know what works and what doesnā€™t for you. So, again, for the listeners that made it thus far with us - and thank you all for sticking - what would you like them to take away from our conversation?

I think the most important thought - in my space, to be specific - is donā€™t think of the authentication system as two input fields and a button. Because thatā€™s plainly under-estimating the amount of effort in depth that goes into such topics. And so I would encourage every listener to always think thoroughly across the reasoning, whether you want to use just a framework in your application, or just create your own login thingā€¦ Because there is a huge attached cost to that in regard of operational security. Because you need to maintain it, you need to pen test it, you need to prove the security of the processes, and all that - letā€™s call it dirty plumbing work, to make it happen.

And so I really encourage everybody, please use some kind of turnkey solution thatā€™s battle-testedā€¦ Even though if you use a framework from your language-specific thing, donā€™t build it your own; you will hurt yourself at one point. And you donā€™t need to take my word for it, but a general agreement in the industry is that itā€™s better to have somebody deeply committed to the topic of authentication and authorization, to have them working on that, and not you. Just use something, and build great features and great products. So thatā€™s really my one thing I really want to get across to everybody.

It just shows that your head is where your heart is, right? ā€¦which is authentication, authorizationā€¦ I mean, if you build a company around it, you really have to believe it that thingā€¦ So youā€™re committed through and through. And that continues to be top of your mind, which is important.

Yeah, I mean, itā€™s such a huge ā€“ authentication and identity space, in general, it has a huge depthness to it. And it always feels like you can easily do that. But as soon as you start poking into the space, you will see there is like a huge amount of time flowing into it. I mean, the OAuth threat framework is like 60 pages. I recently did something for the Swiss government, which will come out in a few monthsā€¦ Itā€™s like 120 pages on just things you should consider when building something like thatā€¦ And it starts with things like XSS, and CSPs, andā€¦ You need to care for that. Itā€™s just ā€“ the depth is the problem, basically.

Yeah. Okay. So it may seem simple when you consume it as a service, but thereā€™s a lot that goes into it. And if you think you can do it - sure, but there are dragons thereā€¦ So at least just be aware of them. Just donā€™t get eaten without knowing what youā€™re getting yourself into.

Yes. I mean, you can always pokeā€¦ If you want to make funny games, just go to a login page where you have a user, throw in your username, throw in your password, and if the response is coming back faster than 500 milliseconds, I can tell you something is broken already, because no solid password hashing algorithm will return a resulting that fast. Otherwise, they run huge CPUs. I mean, I call it the Xeons and stuff. Otherwise, you canā€™t get that hashing through so fast. But yeah, thatā€™s just a sensation. Thatā€™s a thing I always, I poke around login pages. [laughs]

Well, Florian, thank you very much for joining us today. It was an absolute pleasure hosting you. Thank you. It was a great conversation, and I look forward to the next one.

Thank you. Likewise. I really liked your questions. And you see, Iā€™m still laughingā€¦

Yeah, exactly. So Iā€™ve done something right. Thank you very much for that. Until next time. See you.

Changelog

Our transcripts are open source on GitHub. Improvements are welcome. šŸ’š

Player art
  0:00 / 0:00