Discussion about this post

User's avatar
Thomas's avatar

Everything here makes sense, great summary. The only question I have concerns the controllers. My understanding is that they are constantly evaluating incoming traffic and then optimizing the system in various ways. I wonder how are these controllers able to compile and process so much data concurrently? I'd be interested in knowing if the paper includes any information about how any of the controllers are able to do this. I'm wondering if it's based on aggregate data being constantly produced into view models, some sort of append only log data, Kafka, or maybe just best effort estimations based on periodic sampling (since at that huge of a scale, sampling could statistically provide decently accurate approximations of the underlying reality)?

Expand full comment
Ryan Peterman's avatar

> XFaaS has achieved a daily average CPU utilization of 66%, much better than the industry average.

What are the industry averages?

> These configs are cached in the components themselves, so if the central controllers fail, the system keeps running just fine (but it cannot be reconfigured).

I liked all the details about how they made the system resilient. Thanks for writing this article!

Expand full comment
9 more comments...

No posts