Courtesy of SamsaraSanjit Biswas is the CEO and co-founder of Samsara, which helps businesses track and monitor assets like vehicle fleets and equipment. It’s using large language models to simplify the experience for customers and improve productivity. Biswas also co-founded the enterprise Wi-Fi firm Meraki, acquired by Cisco for $1.2 billion in 2012. Reed Albergotti: You’ve always been good at figuring out the bottlenecks in new technology trends, like with Meraki and wireless connectivity. It seems like we’re at a similar moment now with AI where it’s very cool, but there’s a lot of infrastructure that needs to be built to make it really useful. What are the bottlenecks now with AI? Sanjit Biswas: It does remind me a lot about Wi-Fi 20 years ago, when John [Bicket] and I first started working on it. You use it for the first time and you’re like, “This is amazing.” The first time you opened a laptop and you could just sit at your couch and surf the internet, that was a really big deal and it became obvious that everyone is going to want this. But how do you make it happen? At the time, if you wanted to build a big Wi-Fi network, you needed a Ph.D. in computer science. We’re seeing the same thing with sensors and AI. We have all gone through our ChatGPT moment. We know this is going to be world changing. But then, if I want to improve my risk mitigation at my construction job site or for my trucking fleet or something like that, there’s a lot of “how do we make this happen?” The data needs to be really clean. It needs to be trained on data that knows the answer, you need on-the-road data. And then you need to provide the driver with some real time feedback: “Please put down your mobile phone,” or provide them with some kind of coaching and scoring. That’s another bottleneck we’ve tried to break. We have the data. We have the insight. How do you take the action? You put all of this hardware in the world. Do you see opportunities for edge compute? We actually do a lot with edge compute and AI today. We have millions of these cameras that are deployed by fleets. They run AI inference at the edge in real time. Instead of having to wait hours for footage to go to the cloud and get analyzed, it tells you within a second, “Please put down your mobile phone,” or “You’re tailgating,” and it gives you real time coaching and feedback that’s all done with inference sitting at the edge. We use Qualcomm chips to do that. They’re very powerful compared to computers 10 years ago or 20 years ago. These are like little super computers. Are you able to run some of these newer transformer architectures? We started with the kind of original convolutional neural network models that were state-of-the-art a couple of years ago. We’ve moved to a transformer-based architecture that helps us do a ton of things. We can detect all different kinds of risks. We can detect very complex cases. And then it also lets us do more and more sophisticated detection. A recent one we rolled out was drowsiness detection. It turns out drowsiness can’t be detected with an image-based model. What you really want is a historic record, and a transformer model that is trained on what happened five, ten, 30 seconds before accidents. There’s amazing insights around how people move as they’re getting tired. We trained a whole model based on that and rolled it out to those cameras. Read on for the rest of the conversation, including Biswas on a world where you can track anything from space. → |
|