podcasts
Collection
Collection for podcasts
•
531 items
•
Updated
•
2
speaker
stringclasses 3
values | text
stringlengths 2
4.15k
|
---|---|
A | Welcome to Bankless, where we explore the frontier of Internet money and Internet finance. And today on bankless, we explore the frontier of composability of Ethereum's roll up centric roadmap. Today on the show we have Justin Drake as my technical co moderator as we both interview Ben Fish, the CEO of Espresso Systems. Espresso is building a shared sequencer that operates at a higher plane than the Ethereum layer, one that allows for shared execution amongst the Ethereum roll up landscape, the frontier that Espresso is building in is becoming more and more relevant as the fragmentation of Ethereum's roll ups become more and more obvious. Yet there are a bunch of solutions being built in parallel, and Espresso unlocks a lot of them. A lot of innovation is allowed to come into the world of composability across Ethereum's roll ups. At the beginning of this episode, I really make sure that we drill down into the foundations of this conversation, the 100 level, 200 level basics, before I kind of let Justin Drake take the reins and get into some of the deeper ends of the more technical nuances of shared sequencing. So I think this episode is accessible to all skill levels. If you are at the beginning of your arc of understanding Ethereum, you're going to catch a vibe. And then if you are deep into the weeds, I think you're going to really enjoy the second half of this episode. I think the exploration of composability off across Ethereum's roll ups are, is just beginning on bankless. And this is our second step into the world of composability here, our second episode with Justin. We will get into further conversations with other players in this arena as well in the future. But before we get into this conversation with Justin and Ben, first a moment to talk about some of these fantastic sponsors that make the show possible. Bankless nation, ladies and gentlemen. In my left corner we have Mister Moon math himself, the blockchain brainiac, the ethereum evangelist Justin the juggernaut Drake. Justin, welcome to the show. |
B | David, thanks for having me again. |
A | And coming in for his first time on bankless, we got Ben Fish, the Sultan of sequencing, the espresso emperor, the cross roll up connoisseur, the blockchain barista Ben, welcome to Bankless. |
C | Pleasure to be here. David, thank you for having me. |
A | So where the fragments of Ethereum threatened to turn this digital landscape into a maze of confusion, two men stand tall, illuminating the way to come save us from the labyrinth of Ethereum's roll up centric roadmap today on the show, we're going to be talking about shared sequencing and what this means for Ethereum's roll ups, what it means for composability across Ethereum's roll ups, and trying to get down into the deep end of some of the nuances that are going to ultimately turn into engineering questions and then ultimately turn into production products for the Ethereum ecosystem. And so we got Justin Drake not as a guest, but as my technical co moderator for this episode. And I think we're going to just make sure that we kind of touch on some of the easy things, the easy subjects, before we get into the deep end here. But Justin, maybe I could just ask you before we really get started here, what would you say are your goals for this episode? What knowledge do you want to get out of Ben? What should listeners be prepped for or be thinking about as we progress further into this conversation? |
B | Right. I mean, I'd like to push forward the discussion around shared sequencing and base sequencing. Specifically, I think we're in this interesting phase shift moment where people are starting to pay attention to the problems of fragmentation and also the solutions are starting to pop up. There's a bunch of misconceptions, or maybe I should say preconceptions, that it would be good to clarify. And I also feel that there's this wave of momentum of teams like Ben's team that are potentially making a big pivot. And I think Ben might have a big announcement to make here during the podcast. |
A | Interesting. Yeah. Ben, maybe share some of your similar thoughts or goals or just interests in the, the knowledge and content and conversations that you want to broadcast. What goals do you have for this episode? |
C | Yeah, similar to Justin, I would like to address head on shared sequencing and what it is and why I'm very bullish on it, why I think that it's very important for the evolution of the roll of space and for the scalability of Ethereum while preserving its unity and defragmenting what has happened to Ethereum with the current scaling of Ethereum through many different fragmented roll ups that are not united. I also want to talk about, as Justin alluded to, we've got very excited about based sequencing, the idea of using involving the l one itself, Ethereum l one, in as much a way as possible in shared sequencing. And we've actually found a way to merge the narrative of espresso with that of base sequencing. And that's been a very exciting and new development for us that I've been talking a little bit about, but haven't talked about publicly yet until this podcast. And so I could get into a bit of how we've made those adjustments at espresso and how rollups running on espresso can actually be based rollups. |
A | Okay, I love that. And I want to put a pin in that because I want to work towards that point. But I think we need to establish a foundation, some knowledge before we get into what the significance of that is. So let's see how fast we can kind of speed run through the 100 level and 200 level aspects of this conversation before we get into the deep end. So this just start at the 101 level sequencing. What is it? And then what does it mean to share it? So what's, what's sequencing at all? And then what does it mean to add sharing onto that? Ben, we want to take this one. |
C | Yes. So sequencing today for a roll up is the act of determining the transactions that will actually go into the roll up. And currently most roll ups today have a server that collects these transactions and orders them, gives confirmations to users that these are the transactions that are going to be included. And then ever so often it settles them on Ethereum by posting them to a smart contract. And the smart contracts today can only be updated by this sequencer. So they give in total authority to this central server to determine which transactions are going to update the roll up, with the exception of transactions that can be forced included. And there are other details we don't need to get into right now. Maybe the first thing to talk about then with sequencing is, well, what is decentralized sequencing? It's simply taking that and making it not one server that's in charge of sequencing, but many different servers. So having some kind of decentralized protocol whereby it's not just one party that can determine the transactions that get included, maybe it's rotating set of parties, maybe it's a byzantine fault tolerant protocol. And shared sequencing is the idea that multiple roll ups can share a common mechanism for determining this ordering. And I'll give one slightly zoomed out take though on shared sequencing that I don't think is a common way of describing it, but it's definitely the way that I think about it. I think of shared sequencing as not quite, not necessarily sharing a common party or protocol for determining the ordering, but it's more about sharing a mechanism, sharing a marketplace whereby rollups can effectively sell their block space to third parties who are bidding on the joint block space and may value it more and can create a surplus value. But rollups really can determine by the slot whether to sell that block space or not. If there is value to be had in parties jointly sequencing for multiple rollups at once, then the outcome will be that there are these third parties that will be simultaneously purchasing the block space from multiple rollups and producing these blocks simultaneously and creating a surplus value that gets redistributed to rollups. I think that last point gets a little bit deeper, and I'm going to put a pin in that because I know we're trying to go over basics one on one of what shared sequencing is, but I'll stop there. |
A | Maybe one way to articulate the purpose or advantage of shared sequencing, and this is just my understanding. So check me on this, Ben, is that when you have one rollup, you have the economics of that one rollup. And then when you have a sequencer that is spanning multiple rollups, aka a shared sequencer, we one roll up can have more opportunities for double coincidence of wants with the transactions in one roll up as it relates to another set of transactions in another roll up. And so a shared sequencer is kind of like a matchmaker between the double coincidence of wants of the economies of two rollups. And when it behooves both roll ups, a shared sequencer can kind of make those connections happen. Is that like a simplified way to understand this? |
C | Absolutely, yeah. It enables faster bridging between rollups. There are certain things that you can't do today, like maybe it's good. This isn't what you said is entirely correct, but I think just giving a few quick examples might help make it concrete. Let's say that I want to buy an NFT on one roll up, but I want to pay for it with cash on a different roll up. I might want to make these two transactions atomic without even bridging assets over. Could I buy an NFT on Zora with, I don't know, cash that I have on Zksync? What if I want to take advantage of arbitrage between amms on two different roll ups? What if I want to fund a transaction on one roll up using funds that I have on the other? Let's say I want to take a loan on one roll up and use it to fund a transaction that I have on another, and then even repay the loan. So do some kind of flash loan, for example. All of these things are not possible with the status quo. Insured sequencing is not the same as being in the same execution environment like Ethereum, but it can get us under certain kinds of economic guarantees very, very close to that. |
B | Yeah, I like the marketplace analogy, because it's all about creating value that otherwise couldn't have been created. So there are certain instances where we want this really low friction form of interaction, which we call synchronous composability, and that provides more value than the counterpart, which would be asynchronous composability. And it turns out that on l one, every roll up benefits from synchronous composability. So this is the luxury that we've gotten used to. And once you don't have shared sequencing, when you have this siloed sequencing, then suddenly you fall back to this lower form of composability. And I think what Ben is trying to do is bubble up this value creation. And when there are opportunities to have synchronous composability, try and capture this additional value creation and potentially even give it back to the roll ups. So it's a net positive for everyone. Not only are we creating value, but this value goes back to the roll ups. And there's no loss of the perspective of an individual rollup. |
A | Maybe the ethereum layer one, the common settlement layer for all rollups, maybe one way to perceive it is like a minimum viable composability for all rollups. At the very least, rollups are composable, albeit slowly, albeit extremely asynchronously, through the layer one. And that is our threshold, that's our floor, that's our foundation of level of composability. And with additional levels of innovation, shared sequencing being one of them, maybe an intense is also relevant here. We can actually kind of raise the floor of what composability looks like. And maybe that's kind of like where we are in Ethereum's arc right now is like we've got this minimum level of composability, and now the webbing between some of the roll ups are starting to get fleshed out, which is like what this conversation is here. Ben, what would you like to add to this? |
C | I'll say it like this. Ethereum is a shared sequencer. We already have a shared sequencer, right? Ethereum is a shared sequencer not only for all applications that run on Ethereum, it is a shared sequencer for rollups. It is not as good a shared sequencer as we want right now. |
A | It's not optimized to be a shared sequencer. |
C | Yeah, well, it's a shared sequencer that happens on a delay, right? And I think that it's talking about Ethereum as a shared. The difference between shared settlement layers and shared sequencers I think is greater than people might think. So I think all this discussion about shared sequencing is about improving on the baseline that we have, which is that rollups share Ethereum as a settlement layer, and how can we get more out of that? How can we get atomic execution promises? How can we get just more value out of shared infrastructure than what we already have today? |
B | David, I really like your point around the minimum foundation that Ethereum provides, which is asynchronous composability. And the way, I guess a good metaphor is countries. Let's say you have two different countries. At a minimum, what you can do is transport from one country to another. If you get a visa or if you show your passport and you can also do trade, but there's some friction, you might have to pay import duties or whatever it is. And then there's something a little tighter, a little more close knit, which is something like the United States or Europe, which is some sort of coalition, where within those countries or within those states, you have full freedom of movement and full freedom of commerce. And in some sense chains. Opting into a shared sequencer is all about maximizing this freedom of movement of assets and value, and also maximizing the freedom of commerce. And I really like the analogy that you and Ryan put forward, which is these, this united chains of Ethereum were creating a super low friction environment for the different chains to scale Ethereum with very, very little trade offs. |
A | Okay, so this all sounds great. Fantastic. Why don't we have this right now? Like why it seems so simple. Just like share the sequencers, just get it done. Why can't we have it? What are some of the obstacles, Ben, that you say you're running into a espresso? |
C | Right? |
A | What are just the net obstacles for the ecosystem as to why this is difficult? Like why can't we have this? Why can't we have this? |
C | Yeah, good question. So I think that in part there's a, you know, there's a technical challenge. There's also a more social challenge, a coordination challenge. Let me touch on the technical challenge first. So actually the original concept of rollups was just to share Ethereum as a sequencer. Like the original idea for rollups was you don't actually have a sequencer, you just allow the smart contract to collect transactions. And so effectively, the l one proposer is basically the shared sequencer for all rollups. And just as a bit of a historical note, the sequencer was introduced as something that could be used for a performance benefit. So by having the smart contract only allow itself to be updated by this one party, you can allow that party to collect many transactions, and because it knows that it's the only entity that can update this contract, it could collect many transactions at high throughput. It could compress those transactions so it can get more compression out of this and thus cheaper fees. It can also, if you trust it, give users very fast finality guarantees. So it can say, if you trust me, then believe me, this is the transaction. You don't need to wait for it to settle on Ethereum. This is the transaction, just trust me and I'll eventually post it to Ethereum. So that has all kinds of performance benefits and UX benefits. If you're going to go back to getting all rollups to share a sequencer, then you need, from a technical perspective to preserve some of these performance benefits that we got from centralized sequencing. And so that has been the main, I guess, technical challenge at Espresso. And even now, where we've figured out how to allow roll ups to be based and to basically involve the, allow the, whoever is proposing the next Ethereum block to act as a shared proposer for the roll ups as well. We've figured out technically how to integrate the, I would say the meat of what we have, the technical meat of what we have at Espresso, which is a very fast, high throughput consensus protocol called hotshot, in a way that allows these roll ups to enjoy much of the performance benefits that they have today while still sort of sharing a proposer who is also the proposer of the l one. I don't want to get too much into the weeds on that, but I just want to say that, look, that's the sketch of where the technical challenges come from. And then the social challenge is convincing roll ups that this is a good idea, that they are going to get surplus value from nothing having so much fragmentation, but enabling stronger interactions with other roll ups and also with the l one itself. That coordination challenge is difficult because rollups evolved in a world where there wasn't shared sequencing. And many of these roll up projects have considered doing shared sequencing within their own ecosystems, like optimism with the super chain, for example. But the idea of there being one Ethereum shared sequencer that expands beyond any individual roll up ecosystem is new. And even if it's somewhat where we started, it's not where we ended up, and we were trying to get back to that original state. |
B | Yeah. So if I were to summarize basically what you're saying is that the sequencer is this coordinator which provides various services it can do optimal sequencing and deduplication and compression, so it's a performance benefit, but it also provides UX benefits. One of them, as you said, is this finality, if you trust me, also known as pre confirmations. And then another very useful service is MEV protection, essentially implementing an encrypted mempool using a centralized sequencer. As a user, I send my transaction encrypted to the centralized sequencer. No one else sees it, and so I can't get sandwiched. And I guess what you're saying here is that we've made this mental unlock, which is very new recently, where we can think of the l one proposers as providing these exact same services that centralized sequences provided, and at the same time have all sorts of benefits. Benefit number one is we have this credibly neutral platform. Benefit number two is that we inherit a lot of the security of a fill room. And then benefit number three is that now we can think of true global and universal shared sequencing, as opposed to having ecosystem by ecosystem shared sequencing. |
A | Justin, where would you like to take things from here? I have one set of questions that maybe we will get to, but it's more about just one of the last that Ben said, which is like at the very start, no one wants to use shared sequencing. At the very end, everyone wants to use shared sequencing, the end being when everyone else is already using shared sequencing. But I don't know if I want to hop there first. Justin, what rabbit hole in this whole world do you want to go down first? |
B | Right? I mean, I do want to go through various innovations that Ben and his team are bringing to the table. One of the big ones that we've already alluded to is MEV redistribution. But I really like what you said, that here, shared sequencing is all about network effects. And so if you're one sole roll up, what is the incentive for me to join a network? With one node, there's no network effects, and so bootstrapping these initial network effects can be difficult. But once everyone else has a shared sequencer, then it almost becomes a no brainer for you to come in because you enjoy these network effects. One of the things that I do want to highlight economically is that there's this idea of MEV and do roll ups lose it or do they get it kicked back to them? But there's also this idea of execution fees, whereby by connecting to the network you actually increase your total amount of execution fees. I guess I want to go to Ben and ask him, do you agree that there's these network effects at play. And what would be your strategy to try and kickstart things and solve the cold start problem? |
C | Yeah, Justin, I agree. There are absolutely network effects, and the value of a shared sequencer increases quadratically in the number of roll ups that are joining it. I wouldn't say, however, that nobody wants to use a shared sequencer at the beginning. I think everyone talks about decentralizing their sequencer, and if you're going to decentralize your sequencer, you might as well use a decentralized shared sequencer that, you know, other roll ups are going to join, as long as it doesn't have any downsides. So the idea that it solves an immediate problem for you now, even although if you're the first one to the party, you still needed a solution for decentralized sequencing, and you might as well go to the party that everyone else is going to show up at. The other thing I will say, wait. |
A | Ben, before you move on, I'd like to actually unpack that just a little bit more. Decentralizing a sequencer right now, the base case for Ethereum roll ups is we have centralized sequencers. We get some benefits from that. But ideally, we would like to go from the current state, where arbitrum or optimism have one single sequencer, and they kind of want to split it, maybe into three different sequencers that do some sort of round robin thing, some sort of setup that gives them some new properties, mainly liveness protection. So if one sequencer goes down, the other two out of the three are always still in the network. And so the network has some sort of robustness in terms of liveness. That's the reason why internally, optimism, or arbitrum, would enjoy decentralizing the sequencers. Maybe there's some other benefits as well. Mainly, it's a big liveness thing. And what you're saying is like, well, they want this. They do want this. It's not this. This is a known quantity. What you're saying is like, well, there happens to be this specific sequencer that could also be one of these, you know, of the three arbitrary number that I just picked, of the three, they could just choose this one, which also is a shared sequencer. So they are decentralizing their sequencer if they just pick one that's also shared. Maybe arbitrum and optimism are actually using espresso as one of their sequencers, and so there's not as much of a cold start problem. If they are already looking to decentralize their sequencer, they'll just, you know, why not also show, pick the sequencer, which is having a party where other people will show up, is a shelling point for where people will show up and we can start to get some of those double, triple, quadruple coincidences of wants, depending on how many chains decide to elect to use this particular shared sequencer. That's kind of what you were saying, right? |
C | That is kind of what I was saying, yes. The other thing to note is with based sequencing, you automatically are sharing a sequencer with Ethereum, with the L one, with the EVM. So by involving the L one proposer in the sequencing protocol, all sequencing protocols basically have a rotation of who can propose the next block, who's able to propose the next block, all the roll ups that are on this sequencing protocol. And if that can be the L one, same as the proposer for the next ethereum block, then suddenly this proposer can start to give what we call preconfirmations about what it will do. If I'm a user and I say I only want my transactions to execute on optimism, or arbitrum or Zksync or Starquip or whatever roll up. Tycho. If the price of uniswap on the l one is x, then I can get some kind of promise from the proposer that this will happen. And if it doesn't happen, the proposal will be slashed. Or another way of thinking about this is insurance. I can buy an insurance policy from the proposer that will cover my losses. If this thing that I want to happen, this intent that I have as a user, is not satisfied, who is able to actually sell this insurance? Not just anyone on the street, the proposer, and specifically proposers who have the ability to affect the outcomes of both of these roll ups simultaneously. So even if there's just one roll up, the first roll ups, let's say, I think it will be Tyco. So if Tyco is the first roll up that becomes a based roll up, then it's the first roll up that will enjoy this additional composability with the l one, and its users will enjoy that. I think that's a benefit that anyone can have by becoming a base roll up. And the wonderful thing is that if they join it, and other people want to join it too, for this reason, then again, the benefit of being part of the same party that everyone else is joining is going to only increase quadratically with the number of parties that join the. |
B | If I were to summarize, Ben, I guess you're saying two things. One is that a lot of these teams already on their roadmaps want to decentralize their sequencer, and so they might as well have someone else do all the work. The roll up teams are experts in virtual machines. They know how to design the virtual machines and design fraud proofs around them and snark proofs around them. And they might also branch out in terms of tokenomics and things like that. But sequencing is not what they excel at. Even the simplest form of sequencing, which is centralized sequencing, centralized sequences go down all the time and they're not DevOps experts. The other thing that you're saying is that even if there's just one base roll up, that's already a network with a network effect. And the reason is that there's another node which is the l one, and it's not just any plain old node, it's the damn l one, which has half a trillion dollars of TVL. And so from day one you enjoy huge amounts of network effects with base sequencing specifically because you have this network with two nodes, one of which is the dominant one. |
C | Yes. Yeah, I'll soften just the first remark, because my personal view is that, and I know a lot of the leaders of these other projects, I think all these projects have brilliant people and they're all actually very, very capable of building advanced technical solutions across many different types of domains, including building decentralized sequencing, et cetera. But we know that specialization is also important. And are you going to want to maintain a consensus protocol, a decentralized sequencer, in addition to being the best at building your ZKVM or your optimistic VM down the line, that may not be what you want to focus on. I think that rollups are better off focusing on making their stacks the best stack as an execution environment. But I think that the greater reason is just around the network effect that we're going to have with the benefit of rollups joining a shared sequencer. |
B | Right. So one of the big concerns that rollup have when they're first exposed to this idea of shared sequencing is that they have to give up MEV as a source of income. And I have my own thesis, which is that MeV is going to go down by at least an order of magnitude, if not two orders of magnitude relative to this other thing, which is execution revenue and execution fees. But let's assume that I'm wrong for a moment and that MEV is a very significant part of roll up income. You've been working on something super interesting called MeV redistribution. Can you talk about what it is and how do you achieve it? Technically? |
C | Yes. So I'll go back to what I said at the beginning, which is that I think of a shared sequencer, if built appropriately, and this is the way that we're building it at espresso. I think of a shared sequencer as a marketplace whereby chains are, roll ups are selling proposal rights by the slot to others, whether they're the proposers of the l one blocks or other third parties who are creating even more value by offering pre conformations and cross chain liquidity and atomicity and all kinds of things and more. And so what do, I mean? Why is this a marketplace in the way that we've designed it? Well, let's consider an example. Okay? Let's say that there are just two roll ups. |
B | Roll up a, roll up b, arbitrum and base. |
C | Fine. Arbitrum and base. Right? So, hypothetically, arbitrum and base. |
B | I. |
C | Then, if they're the only two rollups on this shared sequencer, what's going to happen is that third parties are going to bid for the right to be the proposer for the next blocks of base and arbitrum. And there's different kinds of bids that we'll collect will allow people to bid individually on proposing the next block of arbitrum individually on proposing the next block of base, and to bid on the pair. So the bid on the right to produce simultaneously the next block of both arbitrum and base. So let's say that the highest bid that we get for arbitrum individually is, I don't know, five. Let's say that the highest bid that we get for base individually is three. And what we get, we get a bid to produce both of them together for ten. Well, ten is greater than five plus three. So we have the most economic value created by allocating the bundle to whoever bid ten for the right to produce the bundle. But because we got the highest bid for arbitrary bid five, and the highest bid for base at three, that's what they would have got if they were just auctioning off the right to propose their next slot on their own. So we will give arbitrum five, we will give base three, and the surplus of two can be divided proportionally. It could be burned. We will redistribute it in some way so that they're getting even more than they would conceptually on their own. So that's the basic idea. How do you estimate the contribution levels of each roll up to the surplus value? Or how do you make sure that they're making at least as much as they would make on their own. Another way to think about this, though, is that each roll up has the opportunity to set a reservation price. That's why it's really that the roll ups themselves are selling their block space. What does it mean to set a reservation price? Well, arbitrum can bid on itself. It can bid five, it could bid six if it wants, it could bid ten. If it bids ten, then it won't end up in a bundle with base. It will basically exclude itself from the potential surplus value that can be created if it's true value. |
A | Is that like a tariff? Ben, if we're using the nation state United chains of Ethereum metaphor, is that like a tariff? Like you must pay at least this. |
C | Not a tariff will. I mean, it's, I don't think it's exactly a tariff. It's more, I mean, in the language of auction of auctions, or, you know, the economic term would be a reservation price because, or reserve. Reserve price because it's saying that the seller is only willing to sell regardless above this price. What would happen if arbitrum bids ten on itself? Well, because the true value of arbitrum is actually five, and nobody values the bundle more than 13, which would be the sum of ten plus three, the highest bid on basis three, then. Then Arbitrum's bid will win itself. It will be excluded from a bundle. It's the same as if arbitrum were just its own centralized sequencer. It guarantees that it will only make more than its reserve price if it wants its reserve price to be ten, will only make more than ten by participating in this shared economic mechanism. And I think this is just this example with two parties, is here to illustrate that it's very much by the slot. Rollups are selling the proposal rights for their next slot to other parties who are now going to hopefully buy up the rights for many roll ups at once and create surplus value from that. It gets more complicated. Only go from two to three. You don't want a situation where one roll up can bid excessively high on itself, or set a very high reserve price and ruin it for everyone else. So the solution, in order to generalize it from two to three, where it gets a little bit more complicated, and I won't go into the details right here, but we will have a blog post on this, is to ensure that if one roll up that's a very high reserve price for itself. Every other roll up that sets a correct reserve price that honestly reflects its true value will enjoy the benefit of being in a bundle with each other. Right. So one person can't cause others to be excluded from the bundle. One roll up can cause itself to be excluded from the bundle only. |
B | So if I were to highlight one specific point, what you're saying is that there's a mechanism whereby at the very least, the roll ups are getting the value, the MEV, that they would have received if they were completely isolated from a sequencing standpoint. And potentially they're getting more because there's excess MEV that's generated from the additional value of synchronous composability. |
C | That's right. I also want to highlight again what you said earlier, which is that this has nothing to do with the gas fees that rollups charge today. Rollups are making hundreds of millions in sequencing revenue today. That's not from MeV. That's all from execution fees today mostly. And shared sequencers don't touch any of that revenue because even though that's called sequencing revenue today, it's really separate from sequencing revenue. Sequencers make it. MEV sequencers don't make money on transactions that are executed by the individual rollups. They do today only because it's all conflated into this one. |
A | All integrated. Right. |
C | It's all integrated. Right. But shared sequencers separate that. So the transactions are still paying fees in the roll up itself. They don't. We wouldn't pay it to the shared sequencer. The execution fees that they're paying, the gas fees that they're paying within optimism, or arbitram or Zksync or whatever. Whichever roll ups decide to be part of a shared sequencer would be going to the roll ups themselves. And the roll ups can decide what to do with those. Right. The value that's being captured by a shared sequencer is mev. And MEV is not just front running. We still want to build shared sequencers that prevent harmful forms of meV. You can also think about pre conformations as generating mev. So the tips that users are willing to pay for atomic execution across two different roll ups, what users are willing to pay to take advantage of arbitrage between amms and different rollups. There's all kinds of things that people are, would be willing to pay for that goes into what we give the umbrella term mev to, and that's what's being captured and then redistributed. |
A | I want to talk about this reserve price just to try and get more of a mental model around it. So if arbitrum sets a reserve price of ten or two or 100 is the significance of that reserve price just arbitrum or any other layer two, raising or lowering the threshold of how much they need to be paid in order to be a part participating in this marketplace, this double, triple quote, quadruple coincidence of wants. So if somebody is increasing the number, saying like, that's them interpreting how valuable that their transactions are, and they're setting a threshold for how much value they need to retain in order for even to be included in this marketplace, correct? |
C | Yep. Yeah. |
A | Okay. So like maybe a tariff is like kind of a rough, rough analogy, rough metaphor. But I think where I was going with that is like this is the sovereignty of one particular chain, making sure that they are receiving what they perceive to be their due amount of value before they are okay with being interoperable with other chains. That's kind of like where I was going with that metaphor. Maybe that. Maybe that doesn't stand. Maybe you don't like that. |
C | I like it. Yeah. I tend to be over precise. So I think I. Yeah, I do like the analogy. |
A | Okay, cool. |
B | While we're on the topic of sovereignty, is there any loss of sovereignty that might happen beyond economic sovereignty? So you've talked about how economically it does make sense to connect to this network, but is there something else that maybe is lost when we move to shared sequencing? My personal perspective is that there is no loss of sovereignty. And this is something that roll ups, when they're first exposed to this concept, are worried about. Really. There's so many things that rollups can do. They can set their tokenomics, they can choose what their governance is. They can choose all the details around their virtual machine. They can choose how they do biz dev, they can choose how they allocate their treasury to public goods. They can do all of these things and nothing is really compromised. But I'm curious, from your perspective, is there something that they really lose from opting into shared sequencing? |
C | Yeah, I mean, my honest perspective that this is, this is net positive. I'll tell you some of the concerns that I'm hearing that I see, and I think, I think they're coming from. These concerns need to be, I think, mitigated and considered seriously. I think one of the concerns that I hear is that, you know, shared sequencing could create a network effect whereby at some point leaving the shared sequencer could be very harmful for that roll up. If everyone is already used to the idea that being part of a shared sequencer has all these benefits, and then now by leaving it, you. So if shared sequencing does have the ability to create this strong network effect, then maybe chains could lose sovereignty to some degree because they are creating a reality where there's a strong network effect around something that's not just them, right? I mean, the way that I think about this is it's making Ethereum more valuable, and ethereum is not an individual role, right? So it could have that effect that if the downsides to being on a different roll up, for example, are now lower because you're part of. Because Ethereum is a shared sequencer for all rollups, and it could make it less likely that there's one roll up that wins. For example, if you believed that you were going to be the only roll up, and that will never be any other roll up, then maybe it is better for you not to participate in shared sequencing. I just don't think that that is a realistic outcome. I don't think there's going to be one roll up. I think there's at least going to be multiple. And I think that it's better for roll ups to focus on being the best rollup stack and playing to their unique advantages. I think that there's many unique advantages of these different types of ecosystems. For example, ZK rollups have a very unique advantage of being able to enable all kinds of bridging and synchronous composability that can happen on top of a shared sequencer. You can take a look at ag layer, the proposal from polygon, whereby proofs can be used to basically pass messages between chains. That requires a degree of coordination that can come from shared sequencing. So these things are very nicely composable with each other. Zksync has this beautiful idea of roll ups sharing a contract. So sharing basically a bridge, so that you can bridge assets easily from one roll up to the other without going through the l one. But this is also complementary to shared sequencing. So I think by rollups focusing on their own unique advantages and building the best stacks and getting customers, because they're the portal to customers, they're the portal to chains. Rather than trying to just be the only roll up out, there is a better move. But I just want to acknowledge the concern that I'm hearing, and I think it's a valid, it's coming from a valid place. I think the other thing that I hear as a risk or downside of shared sequencing is that it will create a hypercentralized builder market. I think that the concern is really not over shared sequencing, even though people think it's over shared sequencing, but rather the concern is over what we actually want, which is if you make. If you make ethereum more valuable and it's really acting more like one chain, then there will naturally be parties that rise to influential roles through competition. There will be some party that's able to create the most economic value across all these different roll ups that are now more interoperable with each other. And if the roll ups are more isolated from each other and those interactions were not possible, then, you know, maybe there wouldn't be one party that rises to that role. But I think we should find ways to address this more directly. Rather than shying away from basically creating more economic value overall for everyone, we should address those concerns. How do you avoid having parties gain, individually, gain a ton of influence? How do you avoid censorship? And I think that we are addressing those through the design, at least, of the espresso shared sequencer. I know that's something that Justin wanted to touch on a bit as well, so I'll pause there. |
B | Right, so I guess on the first point that you brought up, you're saying that it's very unlikely that we're going to have just one single roll up. They're going to win everything. The way that I think about it is there's going to be a very rich long tail of virtual machines. There's going to be virtual machines that are specialized for gaming, those that are specialized for trading, those that are specialized for whatever it is. And in some sense, shared sequencing is all about diversity. It's all about saying we won't have one single virtual machine that will win everything. It's not just about diversity of virtual machine, but it's also diversity of the tokenomics and the public goods funding and all there is with the package of a roll up. But the other thing that you're saying, I guess, is that the l one will always exist, right? |
C | Yes, yes. |
B | And so there's going to be some amount of assets that might never migrate from l one. Think of Ens, for example. The root of truth, of Ens might always be the l one, just for historical reasons and because of maximum security. The same might be said for large whales. They might always choose, just have their treasuries or whatever it is on l one in some sense. We had this really awkward plan before where we would migrate everything asynchronously from the l one and the l two, and in the process break network effects. But now we have this in some sense better plan, which is that we can smoothly migrate assets from l one to l two and everything stays synchronous. And we won't be breaking network effects now on that point of composability with Dr. One, there was some sort of a technical shift from Espresso's perspective. And I alluded to this at the very beginning, that there might be some sort of a pivot or an announcement, and I want to give you the opportunity to voice that. |
C | Yeah, so I mentioned it at the beginning, but originally we were thinking of espresso as just a replacement for centralized sequencers, a shared decentralized sequencer that would have participation from ETH restakers through Eigen layer. And we thought that this was different than the narrative that you were describing as base rollups. But I think that as we got to understand base rollups more, and as the concept of base roll ups evolved to also try to introduce things like pre conformation layers on top of base roll ups, and as we came to defining from a higher level perspective, what are the things that we really are trying to get from base sequencing. Better security, liveness inherited from the l one, and most importantly, composability with the l one, so that the l one can have synchronous interactions, ethereum itself, the EVM, the l one can have synchronous interactions with roll ups, not just roll ups between each other. We realize that those are properties that we can achieve while retaining the unique advantages of the espresso design. And so that's what we've done now with the design. In hindsight, it only required some subtle changes, but I think that the impact is massive. And basically what we're doing now is, in addition to having this high throughput consensus protocol that can be run by e three stakers, we also allow the proposer to be the l one proposer. Now, the way that we do this is we, after running this auction that I described, whereby rollups are basically allowing third parties to bid on joint blocks based from the roll ups, we give the l one proposer of the next l one block, what I like to call a right of first refusal. So the l one proposer has the option to purchase proposal rights from any individual roll up or even purchase the winning bundle that comes out of this auction. So let's say the outcome of the auction is a bundle of all the roll ups that's going to be produced, or a collection of different bundles. So some partition of the roll ups into bundles. The l one proposer can come and buy the rights for all those bundles at the winning bid price. So it still has to pay for it. We still get this, mev redistribution. So we still capture the MEV in the protocol. Some of it goes, of course, to the l one proposer, mainly the surplus value that the l one proposer is now creating by enabling more interactions with the l one itself. And yet we retain also the. And I described this in a post in greater detail. I think we'll be going a bit too much into the weeds to describe how this is done, but we still retain the fast finality that comes from the espresso sequencer. But it becomes this conditional finality conditioned on published l one blocks. So rollups now know that certain transactions are final unless some published ethereum l one block reorgs. |
B | Okay, let's talk about this fast finality, because that's a whole new topic. But I just want to very briefly summarize what you just said, which is that each l two individually generates some value. But once we consider the l two s together, there's this excess value that's created, and that's all well and good, but now, in addition to the l two s, the pure l two s, the l one proposer can join the game and create even more value on top of that. And we're basically giving the option for every l one proposer to play this game and add even more value when they can. Now, the other thing that you alluded to is this idea of fast finalities. |
A | Justin, can I also just kind of do my attempt at kind of explaining the topology of the network that I'm seeing here? So we have the base ethereum chain kind of thinking of this as just like the planet. And then we have some ball at the very center when it's got all of its validators. And then we have the vertical layer twos spawning off of the base chain. And so multiple layer twos are horizontally scaling. We have layer twos individually, which are vertically scaling. And then we have this espresso middleware layer between the layer twos, which is like this higher level plane that is producing this double triple coinc, you know, quadruple coincidence of wants between two, three, four different roll ups. And so, like, kind of how we were talking about ethereum, is this minimum viable composability, because it's all the way down to the bottom of this stack. We have this new level of composability that espresso is establishing at a higher plane between the layer twos. So we have these vertically spawning layer twos coming out of this base chain that is Ethereum. And then we have a higher plane that is producing composability between all of these layer twos. And that was like, what espresso has been building is building, is continuing to build. But what you are announcing, what you're talking about, your not pivot, but like additional scope for espresso is that when layer one proposers are also integrated into espresso, when they are also becoming a shared sequencer opportunistically, whenever some layer one block proposer, Ethereum validator is proposing a block, and that proposer is also part of the shared sequencer that is espresso, then we also get to include the Ethereum as one of these marketplaces that produces coordinate coincidences of wants into the shared system. And it's not every single block, because not every single proposer is going to be a part of Espresso, but when they are, they can join the marketplace that espresso is establishing higher up across the layer twos. I think that's kind of like my version of explaining this whole thing. |
C | Yeah, that was said exceedingly well. Yes, that was spot on. And just to come back to this, even though it might look like a subtle architectural change, I think the impact is massive. Because now rollups that are running on espresso are based rollups, it can be a choice that rollups make. So if a roll up, for example, is running on the espresso sequence sequencer and decides that it wants to build its state off of an old Ethereum state, so if it wants to build its state off of an Ethereum state that's already been finalized by Casper, for example, the impact of that would be that bridge transactions would have to be delayed, so they would be queued and would be delayed and would not materialize inside the roll up for, I don't know, say 15 minutes. If they're waiting all the way for Casper finality, and they wouldn't benefit from the fact that we're involving the l one proposer in the espresso protocol. On the other hand, if a roll up decides to, and then the benefit to that roll up would be that it enjoys the fast finality that comes from the espresso preconfirmation layer. This BFT finality gadget called hotshot, that is being run by restaked nodes, they would get very fast finality from that. There would be no risk of transactions being reversed if Ethereum reorgs because they're building off of a finalized Ethereum block. On the other hand, what they would be missing out on is the opportunity of an l one proposer to build simultaneously the next Ethereum block and their block, and enable atomic interactions between them based rollups would choose to use espresso and in a mode where they are building off the latest ethereum state, and then espresso is giving them finality. Conditioned, unpublished Ethereum blocks a slightly different type of finality, but it's still very strong, and it achieves a very nice balance between the additional composability you get with the l one and the finality that users want. And you can make up for the difference with insurance. So pre confirmations are basically an insurance that proposers sell to users to remove their risk. |
B | So, just to translate what Ben said, in my own words, there's two flavors of finality from the perspective of an l two. You can have finality which is unconditional, and you can have finality which is conditional on the l one reorgang or not reorgang, I should say. So if you want to have synchronous composability with DL one, there actually is a cost, which is that. What if you make a preconfirmation that depends on the state of the l one, and then the l one itself kind of rugs you? The l one reorgs. Now, one of the things that we're working on is this idea of single slot finality, whereby DL one can't reorg very, very deeply. And more specifically, what single slot finite gives us is, at least the latest design, gives us a real depth of three slots. So even though we have a block which is finalized at every single slot, there is a little bit of latency, a little bit of delay. And this delay is three slots. So that's 36 seconds, as much as could be reorganized in some of the worst cases possible. Now, one of the things that Ben keeps on alluding to, and I really want to dig into that, is this idea of fast l two finality. Right now on l one, we have these twelve second slots, and it takes several slots to reach finality. So it might take, let's say, half a minute to reach finale. But what if you are doing pure l two transactions? So l two transactions that don't touch the l one, can we have basically shorter block times and faster finality for those pure l two transactions? |
C | Yeah, we can. So the way that that works, and this is unique to espresso's version of base sequencing. Let's look at what vanilla based sequencing would look like the transactions are. We're only adding l two transactions through l published l one blocks. Okay, so let's say we have a block. Block twelve on ethereum, okay, and this includes some l two transactions, some l one transactions. After block twelve is published. We know the state of ethereum l one. We know the state of these roll ups. Now, the next block, block 13, is under construction. There's some proposer for the l one that's producing it, and it's also adding transactions to the l two. Well, these transactions that are being put into this block 13, even after they're published, let's say, are not final conditioned on block twelve. They're final only conditioned on block 13. So if block 13 were to be reorged or were missed or just lost in some other block, were producing a slot building off block twelve, all the transactions included in block 13 would go away. They would be gone. This is something that in today's architecture with centralized sequencers, doesn't happen because the smart contract gives sequencing rights to a centralized sequencer who is able to say these are transactions are going to be in this roll up no matter which way ethereum reorganizes, right. Espresso is able to do the same thing. So the way that transactions are pure l two transactions are added is through the espresso BFT finality gadget. Even an l one proposer who is adding on transactions is going to get them finalized first through this BFT finality gadget, which we can think of as a flavor of pre conformations, too. It's something that's happening between all the l one validators that have opted into this protocol and are participating in it. And you can, even before this block 13 is published, continue to build on the l two states after the published block twelve that will be final no matter what is published next on Ethereum. As long as block twelve doesn't reorg, those transactions will be included on any fork of Ethereum. Because they're built off the l two state, they're not affected by what happens after block twelve. You know, they happen logically before whatever comes after block twelve and their final condition on it. That's one of the unique contributions of Espresso's design of base sequencing that I think is very, very valuable, especially considering that you're going to have, I mean, the natural pattern is that you're going to have updates on the l one, and then you're going to have a flurry of activity on the l two, and then you have another update on the l one. Now, importantly, we also enable this synchronous interaction so whoever's constructing block 13 can also, once we have real time proving which is coming. I think I truly believe that real time proving is coming. 100 millisecond latency proving is coming just. |
B | As a piece of context here. Ben is a hardcore cryptographer, trained as a professional cryptographer, and has made a lot of contributions, specifically on snarks and folding schemes and recursive proofs. |
C | Yeah, my side hobby is writing papers on snarks. |
A | Same. I love that in my spare time. |
C | But once we have real time proving, then you'll be able to have deposits from the l one into the l two and withdrawals into the l one within the span of one ethereum block. That's only what base rollups can do. We can still do that, but we can also have, between published Ethereum blocks, many l two transactions finalized, conditioned on this already published thing. Yeah. |
A | So that's the idea with base proving. That one example, that means you could take a flash loan on the layer one, do some activity on the layer two, and then pay it back on the layer one inside the same block. Is that what it unlocks? |
C | Once we get real time proofing, we'll be able to do that. |
A | Okay, that seems like a very high. It's like, it seems like a gold standard of. |
C | Yeah, and this is going to, I mean, this is going to happen. There's many projects out there that are working on ASics for proving. Also with the innovations in recursive proofing, you really only need low latency on the construction of this final proof that summarizes all the individual proofs that happen. So it's getting a little bit into the weeds. But if you're the only proposed and you're constructing an l one block and also the l two blocks at the same time, you will make some l one state transitions. So you'll have some l one transactions to the EVM that you'll maybe do a deposit, you'll pass a message along with a proof. But this proof is, it's not small, it's like very large. And it was really easy to produce, and it doesn't need to be small because it's just in your head. As you're constructing this, you passed a that message and this proof to the l two. You do some l two transactions, you pass another message to the l one that does a withdrawal. It also has this large, easy to construct proof that's still just in your head. And at the end, you take all these proofs that were just in your head and you summarize them into one small proof. And it's that final summary, that final compression of these large in your head, easy to produce proofs that needs to be done in like 100 milliseconds. And we're going to the hardware to do that soon. |
A | Okay. Understanding at a high level how these systems work, somebody here in the actual running of the hardware, somebody here is extremely well networked with extremely strong computational resources. Right. That's just my intuition. Is that check out? |
C | Yeah. But that's also something that it can be outsourced. So if I'm, if I'm the proposer, I don't necessarily need to be extremely sophisticated if I can outsource some of this job to someone else. But I think, and this goes back to how we were saying, how we were talking earlier about how someone becomes a proposer. It's best done through some kind of economic mechanism that assigns the proposal rights to those who are able to create the most value. You want to do it in a way that they don't have a monopoly, that it's easy, it's competitive. There can be multiple of them. Maybe we should talk about censorship next. But I think that the entities that are going to be producing these blocks will be from a more sophisticated subset, but it will be done in a way that doesn't result in, in censorship or centralization. |
A | In trust, right? Yeah, yeah. |
B | And this is basically all the research that we've done with proposer builder separation. The proposer can run on a raspberry PI with a home Internet connection and still tap into the most sophisticated sequencer and builder markets in the world. And sure, there's extreme sophistication on the other side of PBS, but that is segmented away and isolated so as to not corrupt the decentralization of the proposal set. And one of the very important things is that these builders and sequencers can't censor through mechanisms like inclusion lists and encrypted mempools. And David, you mentioned the flash loans, and this is something I want to highlight as being the litmus test of essentially perfect composability. If you can do a flash loan across rollups, then you've solved the it, and we can do that. So we can have, for example, a flash loan of a million eth that originates at l one. You deposit that into roll up a, you buy some tokens, you send all of those tokens to roll up b, so you withdraw from a deposit into b, and then you sell those tokens for, let's say, 1,000,001 eth, and then you pay back the billion eth at l one, and you make a one Ethan arbitrage profit. All of this as a flash loan. |
A | And this is not a research problem. This is an engineering problem. Is that where we are in this development roadmap? |
B | Correct. This is purely an engineering problem. Just a few weeks ago, we've had this announcement from axial, which is a manufacturer of snark proving ASics that they have their first ASIC physical. We can touch it. And that actually does photos on the Internet that you can go and find. And not only that, but there's two separate projects that are also looking to build a snark probing ASIC for 2024. That's fabric and seisig. Once we have that, it's a massive unlock. And one of the things to highlight here is that the snark proving ASIC doesn't have to be trusted for safety. The whole point is that the stock proofs that are generated are totally trustless. They're mathematical proofs. The worst thing that could happen is that for some reason, all the asics from all the manufacturers just suddenly blow up and combust. And then now you have to fall back to things like GPU's and CPU's, which have higher latency. But if we're in a position where we have diversity of manufacturers, just like we have diversity of clients, then there's no reason why they will all suddenly blow up. Now, another thing that I want to touch on is basically this question of finality and ux. I had this opinion that it's actually okay for the l one to have long block times and for finality to take a long time. So you can think of Dr. One as being this ultra strong and decentralized settlement layer. And that means that finality takes a little bit longer than other chains can provide. But the reason why I think this is acceptable is because you can have preconfirmations that are on the order of 100 milliseconds. So you have the best of both worlds, the ultra strong security plus the 100 millisecond latency from a UX standpoint point. But what Ben is saying is that you can have even more than that. You can have this low latency finality specifically for transactions that are pure l two, meaning that they don't touch the l one. If they start touching the l one, then, yes, you start suffering maybe half a minute of latency for achieving finality. But if you're a pure l two, then you have 1 second, two second types of latency for finality. |
C | And in all cases, you still get the fastest preconfirmations from the proposers, which can be 100 milliseconds, but backed likely by less economic value. |
B | William, one final topic that I think is worth discussing is this idea of censorship resistance. We mentioned that there is potentially this hypercentralized builder and preconfirmer market, but I think in our discussions you've alluded to a potential clever solution, some sort of mitigation to that problem. Do you want to walk us through that? |
C | Yes. It turns out to be the same solution that enables this MeV redistribution. We were talking earlier about how we have this auction that assigns roll up block proposal slots to shared proposers, and I say shared proposers with an s plural, because there's one special outcome where there's one individual proposer who bids the highest on all these different roll ups. But we could also have multiple proposers who jointly partition the roll up space into a collection of different bundles. And that's very important when it comes to talking about censorship concerns. So one of the concerns that we hear people talking about is, well, I'm worried that if there's only one proposer that is going to be proposing for all rollups, then they will just ignore my roll up because it brings them too much risk for some reason. Maybe it's some jurisdictional reason, and that's concerning. Right. We'll be left behind. Well, if you are finding the most efficient economic allocation of roll ups to multiple proposers, then that will not happen. Because the superbuilders, let's say, who are only interested in these five roll ups, and they're just simply not interested in touching any transactions on this other roll up, will not bid on this other roll up. They'll bid on the bundle of these five rollups and somebody else will bid on this other roll up. So as long as there's always a market for this other roll up, and in fact, it's a special case, it can be that roll up itself. If you were going to run your own centralized sequencer or your own decentralized sequencer solution, then there already exists a market for running your roll up. And that's not taken away when you join this shared sequencer mechanism. And this is only possible with a solution that doesn't force one proposer on all the roll offs, but rather finds the most efficient economic allocation. So there is this additional censorship benefit. I think that a lot of people are missing that, so I'm glad that we got to talk about that today. |
B | Right. So there's two different kinds of censorship resistance that we care about. The first one is around making sure that transactions get included on chain. And for that we have this really neat solution called inclusion lists. So you can always get your transactions on chain, even if it's some sort of censored roll up. |
C | Yes. |
B | Now, the other concern that we have around sensor resistance is sensory resistance of the pre conformations. This 100 millisecond ux. We want to make sure that everyone has access to that. And what you're saying is that, that if we have a super sophisticated pre confer that happens for, let's say, regulatory reasons, to not be able to provide preconfirmations for one specific roll up. Well, that's fine. What we can do is we can have a second pre confer that is going to step up and say, I will provide preconfirmations for this censored roll up. Sure. |
C | Somebody will bid on it. |
B | Somebody will bid on it. And so you still have this 100 millisecond ux. And the reason why it's compatible to have two simultaneous pre confers is that they're actually acting on disjoint pieces of state. So whatever pre conformations they individually give will never conflict with each other. |
C | Correct. Right. |
A | It kind of sounds like maybe there's a similar pattern here to the meme that I brought up earlier about ethereum, the layer one as the foundation of composability. And it's also the foundation of censorship resistance. And with additional mechanisms that high watermark, that tide level increases up the stack for censorship resistance properties, we just kind of need to build it and integrate it. It sounds like a similar pattern I'm sussing out here. |
C | Yeah, no, yeah, I think so. |
A | Guys, this has been an immensely educational episode, I think just with the overall direction and also some of just like the more down the rabbit hole details, things, lots of positive things to get out of this. Is there anything that we lose? Blockchain systems are systems of trade offs. What are we trading off to get some of this stuff? Maybe, Justin, you can talk about some high level patterns and then we can go to Ben for some more nuanced details. |
B | Sure. Yeah. I think it is important to talk about potential downsides of espresso's approach to counterbalance some of the upsides. Two that come to mind are, number one, we're introducing an honesty assumption in order for this off chain fast l two finality to happen. We're assuming that half of the espresso validators are online and honest. I guess one of the questions is, what happens if they go offline? What happens if they're dishonest? Then I guess the second potential downside is the fact that we're dealing with restaking here, and that might come with risks that are associated with restaking. So throwing the ball to you, Ben, to address the downsides. |
C | Yes. I think that when we talk about downsides, I think that downsides are sort of a relative term, right? So we should be looking at, well, downsides compared to something else. So I think I look at it more as trade offs. What are the trade offs in the design space. Right. And certain trade offs may matter differently to different people, so it's just good to be informed about the trade offs. Espresso sits between, I guess, I would say pure, I guess, original idea of base sequencing and the centralized sequencing that roll ups to today, in the sense that when it comes to, I guess, liveness, it does involve this additional BFT protocol that has authority over updates to the roll up. Smart. Contract. Contract. And so even the l one proposer who's constructing the next l one block can't just stick in transactions into the roll of contract on its own. And it can do it through inclusion lists, but it can't just update the roll of contract on its own. It has to get not the approval of a centralized sequencer, but the approval of the BFT gadget that is being run through espresso. And so the concern that, or the trade off that Justin was pointing out is what happens if the BFT gadget is not live. Now you have this other thing that can stall progress. That doesn't mean that everything loses liveness entirely, because what you can do is if the BFT protocol is not live for a certain amount of time, then you can just allow the l one proposer in a future slot to take over and inject transactions. And this is similar to the design of force include transactions or escape hatches for roll ups, because the same concerns arise with a centralized sequencer that may go down, but that ends up being a trade off, because while you don't lose liveness entirely, there is still a chance that in some slots, even if the l one proposer of ethereum is ready to go, it will not be able to make progress on the l two because the BFT protocol down, that's a trade off. The benefit, of course, is what we already discussed, that now you can get this high throughput and fast finality for pure l two transactions, the same benefits that you get from centralized sequencers today. That was the first concern that you brought up, Justin. |
B | Yes, and let me just summarize this one before we move on. What you're saying is that if the BFT consensus protocol loses liveness, let's say more than one third of the espresso validators go offline. Well, for some period of time the roll ups can't advance because they need this certificate of finality from espresso in order to move on. But what we can do is basically have a timeout. So if we've detected that the BFT has been down for let's say five minutes, well, we just turn off all the goodies that espresso provides. So there's no more MEV redistribution, there's no more fast lt finality. And we kind of fall back to this plain vanilla based sequencing which is maximally robust and maximally live, but doesn't have these additional good. |
C | Well you don't turn off all the goodies necessarily, because the BFT gadget is separate from the MEV redistribution mechanism which is assigning the proposer. So you still have that MEV redistribution mechanism because this is an auction that's being run to determine who gets proposed and then that might be bought up by the l one proposer. It's the BFT gadget that's being used for this fast finality property that we get from centralized sequencers. And if that goes down for some period of time, you can turn it off and you won't get fast finality, but at least you can make progress. |
B | Okay, interesting. That's a detail I didn't realize. So you're saying that the redistribution aspect only relies on the l one. It's some sort of l one auction that's happening. It does not rely on the Bft. And so even if the BFT is broken, you still have the redistribution. |
C | Right. |
B | Okay, perfect. So I guess the second topic is risks around restaking. |
C | Risks around restaking, right. Maybe elaborate first and then I can comment because maybe explain your concern around restaking and then I can add color. |
B | Right. So there's two, I guess, classes of restaking risk that people are concerned about. One is around removing the level playing field that we currently have with staking. So for example, if restaking requires very high hardware requirements or lots of capital or things like that, it increases the barrier to entry and provides an uneven amount of APR for different entities. A separate class of restaking risks is around massive catastrophes and mass slashing. Let's say that millions of eth suddenly get slashed. Does that mean that the l one has to come in, the social layer has, sorry, the l zero has to come in, the social layer has to come in and do a bailout, which would be extremely messy and expensive. |
C | Right? Yeah. No, I've heard this concern as well, and I think that could be a valid concern. This idea that we need to be careful not to overlook social consensus when it comes to this additional role that restakers are playing. I think that it's important to consider the extent to which we are relying on what these restakers are providing. When we talk about pre confirmations, for example, from a proposer, a proposer has restaked some youth collateral and is making some kind of promise to users, and they may violate the promise. And there are risks here. There aren't inherent risks here. And it's maybe an improvement on the status quo of not having a preconfirmation, but we can consider the worst that can go wrong and whether we really would need some kind of social consensus to come in correct for things or not. I'm curious, Justin, how you would sort of look at this. |
B | I mean, recently I've been more and more optimistic about restaking because I feel that there's been a decoupling, a potential decoupling between staking and what's known as restaking. In some sense, restaking is a little bit of a misnomer, because you could put forward any asset as collateral. It doesn't have to be either. And you don't necessarily have to be staking. You don't have to be a validator. You could just put raw eth or whatever you want. And so really, the class of restaking applications that, from a research standpoint I'm focusing on today is those where you specifically have to be a validator in order to participate in a given AV's. And I think there's going to be the vast minority of applications. What I mean is that the vast majority of the time, you can not be a validator and still participate in an AV's. But espresso is one of these exceptions where really you do need to be a validator. You specifically need to be, for example, an l one proposer or an l one attester in order to come in. But now that I'm saying all of this, maybe this is not the case. We don't need the attesters to come in. We could just have random people who are not validators come in with collateral above and beyond or separately from the attesters. |
C | Well, actually, let me just, I guess, clarify one point on the design of these attesters in espresso. So to be in a tester, to participate in this BFT finality catch that espresso provides, you can stake one of two assets. You can stake espresso tokens, you can stake eth. You don't have to be an l one proposer in order to do this. We give the l one proposers this right of first refusal over proposal rights. But for a testers there isn't anything that says oh, you have to be validating for Ethereum. It's just that you can use your stake for Ethereum, you can restake it to participate as intestine. You can also restake espresso tokens. What we do is we will determine the ratio of weight between your eth and your espresso. That will determine how much weight we assign to yth versus espresso. And you can think of this is called a dual staking model by Eigen layer. And I think what it's very useful for is bootstrapping where initially the capital requirements to have 32 ether ready stake for Ethereum and then restake it for espresso is quite high. So you might not get a lot of participation until there's a lot of activity on the system. But if you have the option of coming in staking a different economic asset, then it can help with getting that initial participation. And then the economic activity on top of the system will increase and eventually at some point more e three stakers will join and then predominantly the security is coming from ETH and not from something else. |
B | Okay, understood. So what you are saying is that actually the collateral doesn't have to come from stakers, it doesn't even have to be ETH. And so there isn't this tight knit relationship that you're building with the attesta specifically? |
C | Yeah, I mean I think I view ETH restaking as a subsidy for ETH validators to participate in contributing a security to the system. So if you're already staking for Ethereum then you can reuse that capital to also contribute security to this pre conformation layer. But you don't have to be one of those nodes, you can just stake a different asset to participate. |
B | Pain, right? So just to provide a little bit of context as to why im asking this question is because there is a potential upgrade to Ethereum where we do whats called state capping or stake targeting, where we adjust the amount of issuance to go down to zero or even down negative, for example ten towards negative infinity as you start getting close to a cap. And in that context if the issuance is actually negative, meaning that you need to pay for the privilege of being a validator, then it actually doesn't make sense to be restaking as a validator. What makes more sense is that you withdraw your eth and then you stick it into an AV's directly so that you don't have to pay this negative yield. In the case of espresso specifically, you would do this because you're not forced to be an l one attester in order to be an espresso attester. |
A | Guys, this has been extremely all encompassing. I think this is bringing in a lot of different innovations that are all happening in parallel around the Ethereum sphere. The progress of layer twos, both in their technical prowess, their technical capabilities and their sheer number, along with Eigen layer restaking along with just composability innovations that's been going on. Justin Ben, this has been fantastic. Maybe one last thing to explore, which is further explorations. What's left in what didn't we touch in this conversation that might be left for future conversations? What is left to explore? What are some other unknown unknowns that are out there? I guess it's impossible to ask about unknown unknowns, but what are the known unknowns? Justin, if we were to do another episode in maybe two, three months, six months, what topics would you like to see more clarity on as we progress forward into the future? |