[Music] Let's get into it. We won't be holding everyone up too much. We've got a good talk today. Turns out we build developer tools, but we also want to talk about how we build them. We are kind of a product that was born two to three years ago, and everything we do is basically based in AI. Small models to do specific tasks, large models to do conversational interactions, vision processing models for all that stuff.
We're going to show today how we've built a world-class product that ships across macOS, Linux, and Windows to tens of thousands of daily and monthly active users. That's what a next-gen AI product looks like. So, we're going to get into it. Real quick, quick introductions. My name is Tsavo, Tsavo Knot. I am a technical co-founder here at Pieces and also the CEO, which is not super fun as of late. We just closed our Series A, so the past three months were not much code and a lot of Excel and PowerPoint. I'm glad to be here doing another tech talk. We're really excited about what we're building. Sam is here with us as well. Sam?
Yep, I'm the Chief AI Officer. I'm based out of Barcelona. So, we really span the globe, right? And with that comes a lot of challenges and informs what we do with our workflow. Awesome. Well, let's get into it. Okay, so the main thing we're going to get into first is what we're building and why. Then we're going to talk about how we've layered in machine learning in a very thoughtful way into every experience and aspect of our product. These days, as people hear about generative AI, vision processing, and all these capabilities, it's super important to think about how they compose into a product experience that makes sense. That's what we're going to show you. We also are building developer tools,
which means every single person here can use our product. We'll take the feedback from it, and we want to get your thoughts. We'll have some time at the end for questions. It's going to be a bit of a technical talk. Sam's going to get into some of the challenges and nuances of bringing AI into production, be it small machine learning models or large language models at scale. We were one of the first companies in the world to
ship Mistral-5 and Llama-2 across macOS, Linux, and Windows in a hardware-accelerated regard. Sam's an absolute beast when it comes to that. We'll show you some really cool stuff that's coming up with phase three and phase four of where the product is going. Anything to add there, Sam? No, that's about it. Great. Let's start out with what we're building and why. Maybe some of you were in the main
stage earlier and got a little teaser. Long story short, myself as a developer, I've been writing software since I was 16. It's almost a decade, 11 years now. I feel a lot of pain in the form of chaos, especially as I navigate large projects. Be it in the browser where I'm researching and problem-solving, or in a ton of projects where I'm looking at this Rust project, this Dart project, this C++ project. I'm jumping around a bunch. Our team went from six people to 12 people to 26 people very quickly, jumping around on cross-functional collaboration channels. For me, it was already painful a year ago, and it feels like it's getting more painful. Raise your hand here if
you've used GitHub Copilot or some type of autocomplete tool to write code. Well, guess what? Every single person now can go from the couple of languages they used to write, three or four maybe five if you're an absolute beast, to six, seven, eight, nine, ten. Your ability to move faster, write more code in half the time in different languages is going up.
But what happens when you increase your ability to write a couple of languages in a couple of different environments to now five or six? Well, that's five times the amount of documentation, five times the amount of PR requests you're able to do, or the cross-functional teams you can work on. We're seeing this advent where the volume and velocity of materials that you're now interacting with, the dots to be connected, are fundamentally hinting at how the developer role is changing in this era of generative AI. For us, we want to build a system that helps you maintain the surface area as a developer in those workflow pillars: the browser, the IDE, and the collaborative environment. So, this chaos, how can we distill it down and make it a little bit more helpful, a little bit more streamlined for you throughout your work in progress journey? I'll kind of power through some of these other slides. Maybe it's hard to see, but how many people, this is what your desktop looks like at the end of a workday: lots of browser tabs, lots of IDE tabs, maybe Gchat open or Teams or Slack. That's what chaos feels like. Then you close it all down because you want to go home and drink some tea. Then you wake up the next day and you're like, "Oh,
it's all gone. Where do I get back into flow?" That's what we're talking about, developer chaos. In reality, it starts to feel like there's so much to capture, so much to document, so much to share, and there's not enough time. How do we do that for you in a proactive way? How can we make these experiences where it's like, "I found this config, it's super important. I changed this implementation but didn't want to lose the old one," or "Here's a really great bit of documentation from a conversation I had, just fire and forget and save that somewhere," and have that somewhere be contextualized and fully understand where it came from, what it was related to, who you were talking to about that material, and then maybe give it back to you at the right time given the right context? What we're talking about here is the work-in-progress journey. Most of the tabs you interact with a day, most of the projects you have, most of the messages inside Gchat or Slack, all of that is lost. You maybe capture 10% in your mind, and that's why as a developer, our job is to half remember something and look up the rest because we simply cannot capture all these important context clues and these dots. We're kind of like in this
era now where that stuff doesn't have to be lost. We have machine learning and capabilities that can capture this stuff for us. That's what we're going to get into. Just to put it simply, we're talking about an AI cohesion layer, a layer that sits between the browser, the IDE, and the collaborative space and says, "This is related to that, and that is related to XYZ." We're going to capture that and save that for you, and then give it back to you given that same context at the right time. Phase one was really simple. It was a place for humans to save. This is our first principle, fundamental truth. People need to save things. As a developer,
you don't realize you should have saved something until you realize you didn't save it, and that's really painful if your job is to half remember something to look up the rest. We want to make a really great place for people to save things, and that was phase one, roughly two years ago. This is kind of like an AI-powered micro repository. I'll do a quick live demo here, where when you save something, all that stuff you didn't want to do but should have done to organize it—tags, titles, related links, capturing where it came from, who it's related to—all of that we want to do with AI. That's what we started out doing two years ago. I'll just do a quick example here.
Sam, maybe if you want to hold that for me, and Sam's about to get into the tech on this as well. By the way, I'm not even lying, look at my tabs that I have open right now, just like 50 tabs, docs, GitHub, the whole nine yards, got my email, everything like that. But I wanted to feel natural, right? I'm not up here with just a couple of tabs, like I'm in your workflow every single day, same deal as a technical co-founder. Let me go ahead and show you this. Maybe you're on, you know, open source, right? You're looking at an image picker implementation. So, I'll kind of show you what this looked like before. You're here, you're checking it out, maybe I'll use this, maybe I won't, I don't really know, but I'm going through the doc and I'm like, "Okay, I want to capture this."
So, this is what it looked like before. Maybe you go to your notes, I don't know, maybe use text files or something way more primitive, you paste it in there, right? It's not capturing the link it came from, you're not titling it, tagging it, it's going to be really hard to find this later, whether it's Google Docs, Notion, you name it. A couple of years ago, we said, "Can we build a product that has AI capabilities to do a little bit more magic for you?" We're going to grab that same code snippet, and by the way, I have very limited WiFi, so I'm actually going to do this in on-device mode. Here we go,
we're going to do this completely on-device, and I'll go ahead and... so what I just did there was I switched the entire processing mode to on-device to give you an idea of what this edge ML looks like. That same code snippet that's on my clipboard, I'm just going to go ahead and paste that into Pieces, and you're going to notice a couple of things. Maybe it's a little dark, so I'll switch to light mode. Classic, if Macs in the crowd, we got some buttons. Either way, what you see there is I just pasted that bit of code, and it right away said, "Hey, that wasn't just text, that was actually code." This is using on-device machine learning, and it also said it wasn't just any code, it was Dart code. We took it a step further. So that title I
was talking about, it's an image picker camera delegate class for Dart. If you flip it over, you're going to see there's a whole bunch more enrichment as well. We gave this thing a description, we're giving it tags and suggested searches, and we're also going out to the internet and aggregating related documentation links. When you save something, we're pulling
the outside world in to make sure that stuff is indexed and contextualized. So I can say, "Hey, this image picker is highly related to this bit of doc right here on the animated class." That's the power of machine learning. I just took a paste from what it looked like three or four years ago into my notes app, and I pasted it into Pieces. Instantly, with no internet connection, it enriched it. It said, "Here's all the details, the metadata that you're going to use for organization, for search, for filtering, and sharing later on." It also took it a step
further and said, "Here are the things that it is associated with—the people, the links, the docs, etc." We're going to get into that a little bit. That's just super simple paste. There's a whole lot more to Pieces that we'll talk a bit more about shortly, but let's keep it rolling. Sam, can you tell us a little bit about how that happened, what was going on there? Take us away.
One of the big challenges we faced was this happened incrementally. We started off by classifying code versus text, then language classifying, then tagging, then adding a description, a title, all the rest of it. Once we had all these models, effectively what we had were seven different architectures, each reliant on their individual training pipeline, each needing separate testing, optimization, spread instruction, and all the rest of it. One of the challenges we faced there was once you have all these models, how do you deploy those across these different operating systems, especially when you're training in Python but not deploying them in Python if they're on-device? We were really lucky. We sort of hit this problem just as the whole ONNX framework started stabilizing and being really useful. Eventually, we plumped for that. Of course, this had a lot of challenges. You're training seven different models,
each requiring different data, which is a huge overload on cleaning and preparation. It doesn't scale well for local deployment. We were sort of creaking at five models by the time we got to seven. Things were taking up a lot of RAM, and it was quite intense. It's also super slow and expensive to retrain and train these models. Each one's working on a separate pipeline. We're using CNNs, LSTMs, there's not much pre-training happening there.
One of the solutions around this came with GPT-3.5. Suddenly, there's one model that can do all of these tasks. Great solution, done. You can wrap GPT, wrap a cloud model, get all this information, fantastic. The only problem is the LLMs are too slow. What you
saw there was seven models executing like that. You can't do that with cloud calls, even if just over the network cost. It's not going to happen. What we saw when we integrated LLMs was this whole experience became super slow. Also, these cloud models, they're fantastic general learners, but they're really not there for coding, for your normal coding tasks. I was reading a paper a couple of weeks ago using GPT for
Python coding tasks. You had an issue, you gave it the bits of code it needed to change to solve the issue, and GPT-4 managed 1.4% of these tasks. So, two challenges: slow and not really there to work with coding. I would add one more: security, privacy, and performance. We have DoD users and banking users and things like that. They can't send their code over the wire and back. There was a huge thing around air-gapped systems that were able to do this on-device.
Exactly. We were looking for ways to take this sort of experience you have with the large language models, which is very nice, generating tags, generating descriptions, getting really nice stuff back, and put it locally, serve it on-device. One of the techniques that really helped was this low-rank adaption paper that came out from Microsoft, which solved a lot of our problems. Lots of you maybe, who's heard of LoRA before? There we go. I'll run
through what it does and why that really works for us when it comes to serving these models. With low-rank adaption, you take a general learner, a large cloud model, use that cloud model to generate data for your specific task. It doesn't have to be a lot of data. You then take a small model, something that executes very quickly, and you tune a very specific part of the weights in that model. We were using transformers, a T5, an excellent foundation model. Specifically, we were using a T5 that trained extensively on code, so it was good at generating code, very bad at generating descriptions. We found
if we took this general learner, used it to generate maybe 1,000 examples, we were then able to take this T5 trained on code and distill that behavior into that model. The way you do that is you focus on a particular part of the architecture, the attention layer. An attention layer, there are lots of analogies to what it does, but it's basically just mean pooling. What you do is you parameterize that and learn these weight parameters, which are a transformation on the attention layer. At inference time, you run inference as normal. When you get to the end of the attention layer,
you sum these LoRA weights with the attention layer, pass it forward, and what you get is a highly performant task-specific learner from a more generalized one. This had several valuable features for us. The first was no additional latency. There are a number of ways you can take the knowledge from a large learner and distill it to a small one. This one was particularly nice because we don't have to pay any extra cost for inference. That was number one. Second, training is much faster. You can get nice results with maybe 1,000 training examples, very nice results with 5,000-10,000, you're solid.
What that meant is we went from training seven different models with different data pipelines on hundreds of thousands of examples, sometimes millions, to reducing that to 5 or 10K, which is great for speed of iteration. It also means our engineers can train these models on their Macs. We're talking maybe six or seven minutes to train all of these locally, which means we can move a lot faster, deploy a lot more, and retrain and update these models almost on the fly. The other really interesting thing is this works really well with other methods. You can use LoRA. If you're not getting quite what you want, you can use prefix fine-tuning models. Most importantly, we went from serving seven separate models, which impacted the size of our app, the memory usage, and all the rest of it, to serving one foundational model, a small T5, coming in at 190 megabytes, and then just adding these LoRA modules on the way. This
meant we could serve seven different specialized tasks on the back of one foundation model with these modules. It meant we could move a lot faster. It was a huge turning point for us. Phase two. Phase one was okay; I can save small things to Pieces. I can paste some code in there, a screenshot in there. It's going to give me everything I need. When Sam's talking about those LoRA modules, those embedding modules, we're talking about modules for titles, descriptions, annotations, related links, tags, etc. We have this entire pipeline
that's been optimized. To sum it up, the LoRA breakthrough was pretty profound for us. Then we moved on to another challenge: how do we generate code completely air-gapped or in the cloud using your context of your operating system or a current code base or things like that? How many people out there are hearing about gen AI or AI features from your product teams or management? Who's getting pushed to do that? You're going to face challenges like which embedding provider to use, which orchestration layer to integrate, which vector database to use. These were all challenges we evaluated early on. We ended up at our approach, where we want our users to bring their own large language models. If you work at Microsoft, you can use Azure or OpenAI services. If you work at Google, you can use Gemini. We'll do a quick demo of this. From the initial idea that humans need a place to save, we moved to phase two, which was how do I conversationally generate or interact with all the materials I have saved in Pieces, and then use the things saved in Pieces to ground the model and give me the relevant context? When I go in and ask it a question about a database, I don't have to say what database, what language, or anything about that. It just understands my workstream context. It even understands questions like who
to talk to on my team about this because Pieces is grounded in that horizontal workflow context. I'll do a quick demo of this and then we'll keep it trucking along. Let's take a look at what it means to have a conversational co-pilot on device across macOS, Linux, and Windows. We're going to be doing this in blended processing only. I am offline right now, probably a bit far from the building. One thing I wanted to mention super quick before I get into this is Pieces understands when you're saving things. What I was talking about is the additional context: when you're interacting with things, your workflow activity, what you were searching for, what you were sharing with whom, etc. That's the when,
which is an important temporal embedding space for the large language model. Over time, you begin to aggregate all your materials. All of these materials are used to ground both large language models and smaller models on-device. These are the things I have
already, but let me go over to co-pilot chat. Something interesting here is that you can basically think of this as a ChatGPT for developers, grounded in your workflow context. I can come in here and it gives me suggested prompts based on the materials I have. I'll do a little bit of a setup. I'll use an on-device model, as I mentioned. You can obviously use your OpenAI models, Google models, whatever you like. I'll use an on-device model. I'm going to use the F2 model
from Microsoft. I'll activate that one. Let's ask a question. This is how you know it's a live demo. "How to disable a naming convention in ESLint?" Who's had to do that before, by the way? Any pain? Okay, well, you're lucky. Let's see what the co-pilot brings up here. I'm running on a MacBook Air here, an M2. We were really lucky to intersect the hardware space starting to support these on-device models. That's looking pretty good. It looks like we got a naming style with the rules.
Ignore naming style check. Great. When in doubt, if you're having linting problems, just ignore it. Besides that, that's a generative co-pilot grounded, air-gapped, private, secure, and performant. We think that this world of proliferated smaller models working in harmony with larger models is going to be the future, integrated across your entire workflow. Everything you saw here in the desktop app is available in VS Code, JupyterHub, and Obsidian, and all of our other plugins. Definitely worth a try. Let's get back to the tech. Just wanted to show you how this stuff actually looks, all the crazy stuff that Sam says. How do we actually productionize that?
Sam, you want to take us away? For sure. What you saw there, a couple of things are fundamental. First, you need to design a system that's isomorphic across cloud models and local models, which is super challenging. You need something you can plug into any model because the entire model space is moving so fast. If you don't
do that, you're going to be locked in and left behind. Second, you need to take full advantage of the system. You've got to use any native tooling, any hardware acceleration. Finally, you need a system where you can pipe in context from anywhere. A lot of AI products out there, you're stuck within one ecosystem or application. We wanted to break that. We're a tool between tools. We want to enable you to add context to your AI interactions from anywhere. These were our design principles for the co-pilot.
How did we do this? It was an emerging ecosystem when we got into this. We looked at ONNX Runtime, which was great but didn't have the tooling to run a vast selection of models. We settled for TVM because it abstracted away a lot of the hardware optimizations. Specifically, we used MLC. Has anyone used MLC or TVM? This is your go-to if you want to deploy a local model. MLC was great because it's open source. A lot of the work when you're compiling a
machine learning model is designing specific optimizations for certain operations across a wide range of hardware. That's almost impossible with a four-person ML team. Leveraging open source was fantastic. This is all in C++. We had to get it into Dart. Excellent tooling in Dart to do that. We used FFI to handle the integration into the app.
You've got this thing. Large language models, if you want them to perform well at a specific task, you have to put a lot of effort into your prompt. That's just a fact. You need a way to slam context into that prompt as well. If you're prompt engineering for a specific cloud model, that's not going to do well across specific local models or even other cloud providers. We built a system that allowed us to dynamically prompt these
models based not only on what the model was but where the information was coming from. We wanted to integrate multimodal data. A lot of your information is in text, but I watched a lot of videos when learning to code. My team insists on sending me screenshots of code, which is super annoying, but there you go. We wanted to take away the pain and allow you to add video and images to your conversations with these models. We did that using fast
OCR. Initially, we used a fine-tuned version of Tesseract. Does that ring any bells? Super nice open-source OCR system. We pushed that as far as it could go for us. Now we use native OCR, which at the time seemed like a great move away from Tesseract, a clunky system developed over two decades. Using native OCR had its own challenges.
Real quick on the Tesseract, who here uses Linux? Okay, we'll have to figure that out. What we're talking about next is the evolution of our product. Phase one was humans can save to Pieces, and Pieces will enrich that, index that, make it searchable, shareable, and all the like. Phase two is you move into the generate, iterate, curate loop, using conversational co-pilots to understand, explain, and write code grounded in the things you have saved. Now we are moving into phase three and phase four. We've just closed our Series A, but let's show what all this is about. You guys are some of the first in the world to check out this crazy stuff we're about to release before it's even on TechCrunch or whatever.
We want to build systems that are truly helpful, systems that you do not have to put in all the energy to use. You don't have to memorize what this button does and so on. We want systems that are fast, performant, and private, that feel like you're working with a teammate. We know you're going to need this, you're about to open this project, talk to this person, or open this documentation link. Can we, based on your current workflow context, give you the right things at the right time? The other thing is I don't want to build a 100-person integrations team. There are so many tools that developers use every single day. Building and managing those hardline
integration pipelines is a really difficult thing to do. Standardizing those data models, piping them into the system, and so on. We've spent a lot of time in the past year and a half on some in-house vision processing that runs completely on device, is extremely complementary with the triangulation and the large language models that process the outputs, but it allows us to see your workstream as you do. Understand who you're talking to, the links you're visiting, what you're doing in your IDE, what's actually changing. This is the same approach that Tesla took with their self-driving stuff. Those cars run on-device, see the world as you do, and aren't designed to hardline program every rule for every country for every driving scenario. We wanted to scale very quickly across all the tools, both in an air-gapped environment
or connected to the cloud. From there, can we truly capture and triangulate what's important? That's hilarious because this is straight from our Series A pitch deck. It says for designers and knowledge workers; we're talking about developers today. That's hilarious. Sam, do you want to add anything here? I would say something super important when we're deep in your workflow is we need to stop you from surfacing any secrets or sensitive information. A big component of this is filtering and secrets removal and giving you the ability to run it completely air-gapped on-device.
That's right. Just like the human brain, you take in so much more than you actually process. The job is to filter out 90% of the information and capture and triangulate and resurface the relevant 10%. That's what Sam's talking about. There's a lot of machine learning and high-performance models involved in the filtration process, but it's not just machine learning. I don't know about you, but I'm fed up with interacting with AIs through a text chat. I think there's so much more room for innovation around how we interact with these models beyond just assuming that's the best way. Part of this is really exploring that. Yeah, I would say the joke I make to investors is we went from a couple of keywords in a Google search bar and then opening a link and scrolling to now full-blown narratives in ChatGPT, trying to prompt and reprompt and dig it out. It feels like it's become a bit more verbose
and laborious to use these conversational co-pilots. We think the way they're going to be integrated into your workflow is going to be different. They're going to be more heads-up display or autocomplete-esque systems. That's what we're going to show you. Sam, do you want to hit on anything here? I think we've done it. That's good. Great. Let's do a quick demo here. The first one, which is going to be pretty awesome, hopefully we can see some of this. Basically, I'm in my IDE and Sam,
you could probably talk through some of this as well. This is your demo. For sure. Let me remember. This is my workflow. I'm just moving between windows. This is how you knew I was offline during the demo. There we go. I'm just doing a bit of simulation here. This is a super early proof of concept from a while back. Basically, we're moving between screens. To do this previously, we'd need four integrations: one for the IDE, one for chat, one for Not ion, and one for GitHub.
Let's see if I can switch to the hotspot real quick. There we go. T3 for the win. By the way, that's why you need on-device machine learning. If we can't load a video, how do you think we can do all the processing in real-time for you? It's not going to keep up with your workflow. That would be an absolute nightmare. All right, here we go. What happened here is we've skipped around the workflow, capturing the information from the active windows. We're using your activity in your workflow as an indicator of what's
important. We're not capturing everything, just what you're looking at. We're adding it in as context to part of the chat. This enables me to ask, "Could you tell me a little bit about Dom's PR?" and it's all there. This highlights one of our key principles:
AI is interesting, and these large language models are very convincing, but without the actual context, the outputs are useless. When you give them the context, you can get these magical experiences where you can talk about your team and all the rest of it. We decided to take that and push it further. I'll just add real quick. The way this will
actually manifest is in reality, you can go to your co-pilot and say, "Give me workflow context from the last hour or the last six hours." Instead of adding a folder or adding a document, you're just saying, "I want to use the last three hours of my workstream to ground the model." Then you can ask it, "What was Dom's PR? What was going on? What are the specific changes?" In this video, you can see this real-time output, real-time aggregation, or real-time triangulation as Sam is just moving through the browser, in Gchat, looking at research papers, looking at documentation. Over on the left, it outputs the links, who you're talking to, the relevant documents, filtering a bunch of stuff. That will also be used to ground the model. You can say for the last hour, "What was I doing?" We think those conversational co-pilots, as I mentioned, are a bit laborious. We
want to give you a heads-up display given your current context. Wherever you are, if you're in the browser looking at a webpack config, can we give you relevant tags that you can click on and expand? Can we give you snippets you've already used related to webpack? Can we tell you with high confidence you're about to go look at this npm package or check out this internal doc? Can we give you a co-pilot conversation you've had about this webpack config previously? Can we tell you to go talk to Sam or someone on your team that's worked on webpack? Also, can we point you in the direction of an introduction webpack project? I didn't have to do anything here. Pieces are running in real-time on your operating system, giving you what you need when you need it based on your current context in the browser, in IDE, in collaborative environments. I don't have to type out a sentence. I don't have to go search for anything. I switch my context, and the feed changes. These are the experiences we think are going to be profound not only for developers but for a lot of different knowledge workers at large. Having the right thing at the right time. You look at Instagram, YouTube, TikTok,
Facebook, every feed out there in the world. It's really good at giving you interesting content. Why can't we apply that to your workstream to help you move faster each day? That's a little about the workstream feed that'll launch later this quarter. We've got some crazy enterprise capabilities that'll launch later this year. I think we are at a wrap here. We'd love to have you come talk to us. We've got a booth with no words, but it's got a really cool graphic. You
can't miss it. It's black and white. Sam and I will be around to answer questions. Thank you. Did you say it was watching you scroll through a browser window? Yes, that's right. In real-time, on-device, no internet. So it would see the URL that you might hover, seeing what you're reading also? That's right. In a really smart and performant way, by the way. A lot of processing, so we do a lot of active cropping. I wanted to know if you could talk more about how you decided to subdivide the tasks from a larger model to smaller models. Running these models locally, did these smaller models provide more leeway for the tasks? You want to talk about this? Well, it was one smaller model. That's the key thing. We went from seven to one. That's a huge
saving in terms of memory and all the rest of it. In terms of what tasks we decided to focus on, that was really informed by designing a snippet saving app. All of this metadata we generate is very much keyed into you being able to find the resources you've already used. Things that are useful to collaborate. That's why we hit up links,
related people, descriptions, all the rest of it. Those tasks almost designed themselves. There are constraints. If you want to generate a description, there's a limit to how much memory you're using at that time. If you can generate that description in less than 100 milliseconds. We look at the problem and say, "What's the probability we can generate a two-paragraph description in less than 100 milliseconds?" We work backwards to solve for those tokens per second outputs. We work backwards to solve for GPU, CPU, memory footprint, whether it's cold started, you turn it off and on to free up that memory. There are all types
of challenges involved. I think generating a description is the largest, but generating tags, titles, those are all sub-problems of descriptions. We executed that pretty well. Other questions? When are we going to be able to use AI for no keyboard, no mouse interactions to do our tasks? We're waiting on Neuralink or Apple Vision Pro, not really sure which one's going to hit first. If I was thinking at the speed of OpenAI's output, those tokens per second, I'd probably be pretty bothered. We need to get that way up. We need to make it efficient. It's quite expensive right now to process all the data as we're just throwing it up at these large language models. It's kind
of like the human brain is a quantum computer walking around and operating on 13.5 watts. We need to hit that level first. I can't give you the answer, but I know we're well on our way. I love my keyboard. I don't think I'd ever want to be in a situation where I just talk to my computer. We're getting there, at least heads-up displays. You showed how you literally watch what's being done. Any way of going
the other way around in the future, instead of just being in a tiny window, you have a draw on top of the apps to show what buttons to click and what the definitions are? I think it'll be really tricky, but if we fail at building developer tools, at least we could open source the system so robots could order at McDonald's. We know that button's important to click, we're going to do that one. We're looking at the URL, the tab name, what's scrolling, and so on. Vision processing is going to be important. Our eyes are pretty important for us. These systems need to be able to do that. If we could build a great foundation on that to overlay recommendations of actions to take, that would be cool. But we'll get there. That's maybe Series F or something.
One more question. That's exactly right. Our models strictly understand technical language. If you asked it a question about poetry, it would not give you a good output. That's how we're
able to detect if you're working on technical language or natural language. Our models are biased towards code. We've trained them on a lot of code. When you distill a model down, you have to be very specific in what type of environment it's going to perform well in. When you take it
outside of a technical language environment into a heavy natural language environment, the performance is capturable but not helpful. If you ask the co-pilot about the bank statement, it might do all right, but we don't care about that. We filter all that out. Just don't tell your employer you're using your work laptop for 25% personal stuff. We
get around this with a lot of clustering. We came to co-pilots after we'd done a lot of work in embedding clustering and search. We leverage that work. At the end of the day, you can run this all on your device. If you're using that device half and half for personal and work, feel free to use our co-pilots half and half for personal and work.
One more. I'm trying to think of this in the context of my everyday workflow. Let's say I've been using Pieces for three months. It knows exactly what I've been doing. It knows my Kubernetes clusters and all that. Do you have something in your roadmap where it interacts with Kubernetes, where I say, "Hey Pieces, remember that API that I built that has this CICD workflow in GitHub and is deployed to this cluster. Do that again." Is that something?
We're really interested in human-in-the-loop systems. A lot of work is being done around these autonomous agents. We see a lot of interplay between co-pilots. When we get into the operational execution side of things, we'll slow roll that a bit before we say, "We know all the docs, all the people, go out and do it for you." We're quite far away from that and we'll let GitHub take the lead there. For us, the stuff you interact with as a developer, that's at least 30-40 tabs worth, a couple of conversations, and definitely a Jira or GitHub issue ticket. All those dots we
want to connect for you. When it comes to executing, hopefully, you're still the one doing that. Otherwise, if it's fully autonomous, we might have to change the way we write code. We'll deal with that later.
Cool, I think that's it. Awesome, thank you guys. [Music]
2024-06-21