Guest: Ron Green, Founder at KUNGFU.AI
Host: Chris Featherstone, Sr. Director of AI/Data Product/Program Management at Salesforce
Published on: September 17, 2024
In this episode of the Earley AI Podcast Ron Green, Founder of KUNGFU.ai joins host Chris Featherstone. Ron's expertise spans over 25 years in artificial intelligence and machine learning. Starting his journey in computer science during the early days of AI, he has witnessed and contributed to the evolution of the field from modest neural networks to today's complex, groundbreaking systems. With a master's degree in AI from Sussex, he has firsthand experience in areas ranging from protein folding predictions to advanced graph models.
Key Takeaways:
Most AI driving business ROI today is domain-specific—trained to solve one problem very well—not general-purpose systems like ChatGPT that have broad but shallow capabilities.
Insightful Quote:
"Accuracy in AI models builds trust, and that's crucial for their successful deployment. But it's vital to remember that all models have biases; the key lies in identifying and addressing the unfair ones early on." - Ron Green
"When text-to-video systems generate a pirate ship battle in a coffee cup, they're not just creating pixels—they're simulating fluid dynamics, ray tracing, and buoyancy. None of that was programmed. It emerged from training on petabytes of video data." - Ron Green
"The biggest misconception I see is people equating ChatGPT with all of AI. Most business ROI comes from domain-specific models trained to solve one narrow problem extremely well, not broad general-purpose systems." - Ron Green
Tune in to hear insights from 25 years of AI evolution—from the challenges of the AI winter to today's transformer revolution—and why the future holds even more dramatic changes as models begin standing on each other's shoulders.
Links:
LinkedIn: https://www.linkedin.com/in/rongreen/
Website: https://www.kungfu.ai
Podcast: https://www.kungfu.ai/resources/hidden-layers
Ways to Tune In:
Earley AI Podcast: https://www.earley.com/earley-ai-podcast-home
Apple Podcast: https://podcasts.apple.com/podcast/id1586654770
Spotify: https://open.spotify.com/show/5nkcZvVYjHHj6wtBABqLbE?si=73cd5d5fc89f4781
iHeart Radio: https://www.iheart.com/podcast/269-earley-ai-podcast-87108370/
Stitcher: https://www.stitcher.com/show/earley-ai-podcast
Amazon Music: https://music.amazon.com/podcasts/18524b67-09cf-433f-82db-07b6213ad3ba/earley-ai-podcast
Buzzsprout: https://earleyai.buzzsprout.com/
Podcast Transcript: 25 Years of AI Evolution, Domain-Specific Models, and Ethical Challenges
Transcript introduction
This transcript captures a conversation between Chris Featherstone and Ron Green about the evolution of AI from the 1990s to today, exploring misconceptions about generative AI, the importance of domain-specific models, ethical challenges with bias, emergent capabilities in text-to-video systems, and why accuracy and explainability build trust in AI deployments.
Transcript
Chris Featherstone: Welcome to the Early AI Podcast. I'm Chris Featherstone and unfortunately Seth can't join me today, but it'll be be really fun to dive into, you know, our guest today in terms of talking about AI misconceptions and the emergence of AI capabilities, the role of knowledge and architecture and AI systems. Let me get to our guests because that's probably the most important thing that all of you listeners want to hear about anyway, instead of me rambling on. So our guest today is a digital guru with a background in computer science and a master's degree in artificial intelligence from Sussex. He's the founder of Kung Fu AI and brings over 25 years of experience working in artificial intelligence and machine learning technologies across industries like biotech, security and telecommunications. So without further ado, Ron Green, welcome to the show, my friend.
Ron Green: Hey, thanks for having me, Chris. Listen, it's been as you
Chris Featherstone: know, a extremely crazy ride across AI and ML in general, especially within the last, we'll say 18 months. And it drives me crazy. And yet it's super exciting at the same time that people believe that this is now AI. And I would love to know from your experience, because being in it 25 years, what did you work on back then? Like, what were the types of models and what were you trying to solve and things like that. I'd love to get, you know, not to go back to your history really. Like, it may bring up facial tics, but I'd love to know some of the things that you were working on back then. Yeah, yeah, absolutely. So
Ron Green: back in the 90s when we were building neural network based systems, they were just so much smaller than they were today. So the types of problems you could solve would be viewed as beyond toy problems now. So we were doing things like using neural networks to try to adjust the output of like DNA sequencing hardware where you would, you would fluoresce the nucleotides and you would run them through the DNA sequencer. But as, as those nucleotides sort of elongated through that process, the waves would get all mixed up and it was literally difficult to tell whether it was, you know, ACGT or was it was that C before the G, because the waves would kind of just wash over each other. So we were using neural networks to solve problems like that. We were using them for rudimentary vision problems. So within simulated environments, like imagine autonomous robots, little sort of car like figures giving them just enough of a visual signal that they could move left or move right. And these sort of dynamic, evolving systems interact with each other wayfinding things like that. But the, the biggest problem about AI back then was it just was so hit or miss. We're definitely still in the realm of sort of, it's more art than science. We don't have a super deep understanding what's going on. But back in the 90s, it was not even clear if we were on the right path. I've told this story many times, which is the folks working in symbolic AI, you know, one of the main, you know, sort of alternatives to connectionist and oil based AI they would joke us with all the time. They would say, when your AI systems don't work, you don't know why, and when they do work, you don't know why. And so there was this sense that it was pretty far afield of science and more just, just throwing stuff against the wall. And today we're, we're solving problems that would have been a dream, you know, back 25 years ago. So you, you said
Chris Featherstone: something that was interesting in terms of that even today it's still more art and science. Double click on that for a minute because I want to understand and then we'll get to like misconceptions and stuff. But I'd love to double click what you mean by that. Yeah, absolutely. So it's,
Ron Green: it's still more art than science in many ways, not the least of which is, you know, we don't understand at a deep level how our own thought processes work, how our own consciousness work, what it even means to know something at the philosophical level. And we're seeing that, we're seeing that right now with, you know, these neural networks, especially these large, large, large, highly parameterized systems with billions or trillions of parameters. So you can introspect them to an extent and you can like, let's, let's use a concrete example. We built a system that can accurately predict breast cancer risk using just mammograms five years out. It's, it's at the FDA right now for approval. What's amazing about this is it's not detecting cancer. There is no cancer in these mammograms. But it is picking up on biological signals of cancer, sort of whether it's, whether it's biological visual signals of cancer, correlation or early stage signs of cancer, we don't know. We flat out don't know. And if you were to show radiologists these images, they would give these women clean bill of health. And so even when we've done like ablation studies, even when we've done things like grad cam, where you can go in and you can look and say, show me exactly what the model is looking at and where is it focusing. And look at that attention across thousands and thousands of different mammograms. It's just unclear at a deep level what it's doing. And this flows all the way to language models where you can ask a language model a question and it's not clear exactly how or where that information is vetted, where that knowledge that's embedded in it because it's spread across, you know, so many of these sort of higher dimensions. So with those, with those, those vision services and stuff,
Chris Featherstone: and looking at the, I don't even know if you call it precursors to cancer, right? Because they're, I mean you're really just the pattern recognition of these. How do you get the clinicians on board with. When you're showing them something, you're like, hey, listen, by the way, look at this image. And they're like, yeah, that looks completely. Because I do radiology all day long, this looks completely clear. And you're like, you actually put this in a million of these images. You know that this is a five years out. How do you get them to like agree? Yeah, that, that's a really,
Ron Green: that's a really good point because you're, you're essentially asking these professionals who have years of experience to trust some sort of black box system, right? What we, what I always advise is as much as possible, try to avoid the scenario where you have some black box AI making a determination and then you blindly accept it. Whether it's, whether you're talking about loan approvals, cancer prediction, or even something as, maybe as prosaic as like summarizing content. Now it's always good to have some types of checks and balances and human in the loop. And we're at Kung Fu AI, we're really big fans of sort of AI as an augmentator to humans, not as a replacer of humans, but within the specific breast cancer example we're talking about there, it's literally just statistical. We, you know, to get an AI model like this to the FDA for approval, it is an astoundingly expensive, time consuming process because you essentially have to convince the FDA that you have no bias within the model. And the amount of bias, it would just blow your mind. For example, one of the things you'll run into is as patients become older and the chance of cancer increases, or maybe they even were screened and there were some warning signs they will go have follow on scans at better machinery. Well, it's very easy for these AO models just to note that the type of mammogram correlates with cancer. So we had to do all kinds of work to make sure there was none of that sort of bias within the model. And then it's simply a matter of overwhelming statistical evidence. And so we, on this particular example, we had terabytes of mammograms longitudinally spread out across over a decade, from three different continents, across multiple manufacturers. So the point is that at that point you have, you have such broad statistical evidence of efficacy that even, you know, even clinical professionals can't help but look at the evidence and be convinced. I mean, I can see this across
Chris Featherstone: definitely across healthcare because the outcomes is driving like life and death, financial services, same thing, right? In terms of trust, what other misconceptions do you run into? Just in terms of like areas, just besides trust, like, and, and accuracy. But what other misconceptions run into? Probably the biggest conception,
Ron Green: misconception that I see these days is since Chat GPT came out almost two years ago, the average person, if you stopped on the street, they would equate that with AI full stop. Right? It's just, you know, chat models, large language models. That is a, and that is amazing. And in some ways it's understandable because ChatGPT, in my opinion, is a really impressive accomplishment. The fact that scaling up language models to the trillion parameter level is allowing us to build systems with these broad capabilities is amazing. But most of the AI that is being used in the real world right now to drive business ROI is what I call domain specific. So it's a type of AI that was trained to solve one problem very specifically and be very, very good at that. But that's all it can do where you can talk to ChatGPT and you ask it to write a poem or summarize or classify. And it has sort of broad capabilities. Most of the ROI out there that businesses can reach for now using AI is something like, like we talked about automating certain processes, maybe loan decisioning or document processing or sales prediction or product recommendations, whatever it may be. And AIs are incredibly good and powerful at those specific narrow tasks. And so the idea that generative AI is synonymous with all AI is probably one of the biggest misconceptions I see out there. Oh
Chris Featherstone: yeah, there's so many times we have to help people unlearn and then relearn to open up the pathways of, oh, there's more enlightenment that comes from this. I'd love to get your Take I don't know if you saw in the interwebs this week about they're trying to map it to IQ and they've been doing these IQ tests. Right. With these large language models. And to me, IQ is I think, a little bit contextual, of course, as well as semantic. Yep. What are we talking about IQ level just in terms of the vernacular reading and comprehension or recall or what?
Ron Green: Yeah, that's exactly right. We're actually working right now on a problem that's kind of related to that, which is trying to ascertain and basically provide a metric for how much incremental knowledge any one single piece of training data may provide to a language model. And it's really, really interesting because it goes back to what we were talking about at the beginning of the conversation, which is, you know, what does it even mean to know something and what does it even mean to trust these black box models where the information and the knowledge is so spread out across all these sort of like high dimensional embedding spaces. And so I think it's very closely tied to some of our own lack of understanding about what it even means for us to know or understand anything because we still don't have a strong understanding or really any capability to introspect our own consciousness yet. Yeah, it's
Chris Featherstone: crazy. I know. In the pre call you guys, there's some discussion around just the rapid advancement and merging capabilities, especially in text to video systems. Yes. Which is borderline creepy in a lot of ways. Right. Just in terms of what we see out there, in terms of, hey, listen, I'm going to create this and now all of a sudden I've got this rendition of just some text in a really, really succinct, crisp way. But I'd love to get your take on what you think, you know, and their ability to, you know, simulate maybe complex phenomena without a lot of, let's say, instruction. Right. Teaching. Yeah, give me your take on that. I so this is an area
Ron Green: that I am deeply fascinated with because if you like to maybe kind of set the table for this. When text to video first started being pioneered, which was just a few years ago when that long, the videos were really blurry and they would really quickly sort of lose context. And so if you're familiar with this, I'm sure everybody who's listening to this is hands would come and go and new arms would come and go and you know, maybe a person would be. One moment they're on the beach and then the next moment they're in the office. So Continuity was just really problem at that point. And now the new. If you look at like Runway ML and some of these, these other text to video creators out there, it is astounding how consistent the context is and how realistic. But even somebody like me who's living and breathing in this every day kind of missed something that I think in hindsight was staggering. If you look at those videos, they're like one of my favorite examples is the OpenAI Sora example of two miniature pirate ships having a battle inside a coffee cup. And it's photorealistic. And if you look at it, just immediately you're impressed, but it takes a moment to realize it is simulating fluid dynamics. It is simulating ray tracing as they move around and the light is bouncing off. It is simulating buoyancy and smooth movements of these ships. And none of that was programmed. None of that was programmed. All of that was sort of emergent capabilities from just being trained on, you know, petabytes of video data. And so that means that not only are these sort of text to video systems capable of generating accurate lifelike renderings of whatever you type in, but they have embedded in them entire physics engines that are compressed in some really, really deep latent space. And that just kind of fell out of the training emergently. And to me, that is fascinating. And nobody understands what's going on at a deep way there yet. See, that's, that's the scary part, right?
Chris Featherstone: I mean, I love the fact it's, it's so refreshing to see a, you know, a tried and true practitioner like yourself get excited about what's next and all these different aspects and kind of blows your mind, right, with the foundation stuff. And at the same time, there's some scary notions out there, right? And you talked a little bit about bias, right, with how important part of that bias is, as you can imagine, walking into the, the, the different governing bodies and stuff that are like. Did you take all these edge cases into consideration too? Right? Like, right. Oh, listen, you're talking about mammograms, but you're male, right? Do you have even those types of things? How many, you know, what about ethical considerations? Right, because part of this is, you know, in the beginning you were talking about there was no explainability, and yet we. Explainability is a big, you know, audacious thing now that everybody pivots on it. Yet you still can't explain a lot of these kind of things, like you're saying, without divulging, maybe IP without divulging, process all that Kind of stuff. So let's go down the ethical route, rabbit hole a little bit. And yeah, I mean, explainability is different from ethics, but, but there's still a lot of it because one begets the other, that kind of thing. But that's totally right. What do you think about that? What do you think about the risks are with some of this? It's a really, really important
Ron Green: area. And I'm always candid about this. I've been doing this so long that I remember when ethics would come up, let's say in the mid-90s. And my reaction was ethics, this stuff doesn't even work. Like, who cares about that? Right. And I use this analogy, it's really analogous to the Internet. When the Internet was first built, coming out of the, you know, the late 60s or early 70s, they weren't focused on security, they were just focusing on reliable network connections. And then once the Internet took off and they, they'd really sort of ironed out most of the, that networking stack, it was time to take security seriously. And it kind of had to be layered on top. It kind of had to be like wedged into the, the networking stack because it really wasn't designed for that. From the beginning, AI is really seeing a lot of that as well. We went from systems that were a toy to systems that could kind of work but were absolutely not production ready to just almost overnight it feels like these systems that are superhuman capable can even be autonomous. And now we are having to rush to kind of get our act together on sort of the bias ethical front. And a really great example of that is if you think about like loan decisioning, we have real first hand experience with this. We were building a system to do loan decisioning. And one of the beautiful things about something like that is you could say that this system was unbiased. It doesn't see race, it doesn't care about any of those classic ethical bias issues that plague humanity. But the data set it was trained on actually had all of that historically embedded in it. Yeah, okay. And so when we train that system, anticipated this. When we trained that system, we then did a bunch of studies to figure out what were the racial or ethnic biases based into that loan decisioning model. And then we were able to remove it. We were basically able to suppress those. But too many, too many of the, of the customers that we work with day to day aren't really aware of that issue. And if you train AI systems on historical data and you are just blind to the bias baked within it, that AI model will replicate all of that bias. And you, you will not be in a position to say that it's making independent, objective decisions because you've basically taught it to be bias.
Chris Featherstone: So then where do you. Because it seems to me like the inverse of that is profiling. Profiling is what we look at for classification and for recommendation engines and stuff and which alone type of approval decision making process is based off of a profile and a recommendation of variance. Right. Because I look at my. What are all the factors that go into. Shoot, I don't even know because like credit, for instance, credit scores. Like it seems like it's talking about a black box. The same, like the same algorithm that's deciding that is the one that's deciding airplane ticket prices and Vegas hotel prices, like running that. You know what I mean? Right. So, so where does the profiling then come in? Because you just, you know, we talk about you remove all the bias, but then you remove all of the ability to make a sound decision too. It's a really good point. The way I think about
Ron Green: it is every model is going to be bias in the same way that, in the same way that you, there's no way you could grab all of humanity and get them a degree, 100% on anything. Everybody, every human being has some type of bias and much of that bias that we all have, and I guarantee you I'm in this category, I'm not even aware of. I have some just built in bias that is sort of below my consciousness. So it's not really a question ever of saying, hey, we're going to build models that have no bias. It's just getting comfortable with what type of bias it's going to be. And, and from my mind, what you want to do is you want to focus on the bias that is unfair. So if you're making, let's say loan decisionings, you probably really do have to look at somebody's income as a part of that decision process. And you can make an argument that that's a bias, but you could also make the argument it's an appropriate bias for this particular scenario, but looking at their age, maybe less justification. Looking at the race, absolutely no justification. And so it's really kind of a spectrum and there's no one size fits all. You just the most important thing about bias is to make sure that the people building the AI systems and whoever is, whatever the domain is, that they collectively are working and surfacing these issues early into the process and addressing them. Because if you wait till the end, it is Very, very difficult to kind of retroactively remove the bias in a systematic way. I think to me that is one of the better
Chris Featherstone: definitions of explainability. Knowing, you know, actively what you're doing as opposed to, well, here's where we're getting data from and here's the path to the algorithm and stuff. No, no, no, no. Here's the explicit things that we are looking for in a pattern, the explicit things we're excluding and then you get into how to explain it. Yeah, and another thing too, going back a little bit I guess, to like
Ron Green: common misconceptions, is we run into this all the time where we'll work with companies and they think that if there is some type of AI model that will be deployed, that it's kind of all or nothing. It's like all black box, you can't control it. You are essentially letting go of the decision making around whatever that model is touching. And we actually advocate really strongly against that which you can put in all kinds of oversight. Like you mentioned, around explainability. You can have thresholding, you can have policies that override. So for example, with loan decisioning, you might have models that are making recommendations. They could be scoring along many dimensions, you know, anticipating fraud, anticipating, you know, likelihood of repayment. All the normal dimensions that you might think about around loan decisioning. But you can have separate thresholds that you tune and that you change over time around what confidence level on those different dimensions you live with. And completely independent of that, you can put in policies. They can be just old school traditional software policies that say, hey, if these things, if these conditions are not met, I don't care what the AI model says, we are going to maybe reject the decision or we're going to put this through some type of human review. So that's another really common misconception is the lack of explainability therefore forces these models to be entirely autonomous. Do you find too that
Chris Featherstone: there's also a misconception on what data is needed? We work with data scientists all the time and yet, and I feel like, by the way, just my opinion is that we put either a lot of pressure and, or a lot of gratis on them, that they know all and all data sources and everything else, which to me there's, they're like another model. They're only as good as the data they are provided. Right. To look at. How do you, you know, when you talk to your customers and stuff, what does the data picture and perspective look like? That is a really important area. I'm,
Ron Green: I'm Actually really, really glad you brought that up. And there's like sort of three things that come to my mind is one, we've been doing this now, we've been building AI production systems for almost seven years. This has only happened a handful of times, but we've had some clients where we are early stages of the product development and it becomes clear that the data they thought they had was actually not there. And it was because either it was deleting it or they were compressing it in some way to save money and basically destroyed all the signal. So I always tell clients, look, the, you know, it's like the old proverb, the best time to plant a tree was 20 years ago. The second best time is today. Same thing with your data. It's, it's a, it's a, it's a very, very valuable, you know, source of intellectual property. So make sure you're not throwing it away or compressing it too much. The other thing we'll run into is, is, is, you know, companies underestimating how much data they need. Depending upon what type of model you're building, it can vary. So if you're building, let's say a language model and you're able to work in a space that is, let's say, somewhat generic, like you're just trying to, let's say you have a language model that you want to fine tune on your own proprietary documents, but you're within some vertical and there's a little customization. But for the most part the, for the most part the language and the semantics are, you know, common English or whatever your language may be. That's one thing you can use off the shelf open source models and, and really move pretty quickly. If you're working in sort of specialized domains, legal, medical, the bar is higher on, on the data acquisition. And then the last one, and we do see this every now and then, is some people just say, hey, I have this data lake and there are terabytes of data there. Can we just go release an AI model, just like let it go, look at it all and come back and be smart. And it doesn't work like that at all yet obviously because most AI systems today are supervised learning based, you have to have labels, meaning you have to have what the input to the model would be and what the correct output of the model should be. And very often sort of the worst case scenario is that labeling is an incredibly, you know, intensive human process. Maybe it's 30 minutes, maybe sometimes. We've literally had instances where they say it could Be a week of discovery that has to go into something to just come up with one label. For one example of training data. In those instances really you, you really can, you know, almost conflating sort of like data mining with AI. And so those again are some of the issues we run into Data on the data side, I'll finish by saying the good news is most companies fall into that middle category where you don't have to start from scratch. You can leverage open source. And the lift on getting additional labeled data is not that high. So whether it was implicit
Chris Featherstone: or something, I love what you said around that. The data that you're deleting your signals, right. And that data are, that's exactly what they are, is your signals coming in to know exactly what to do. What are your thoughts too then? Because you know, we also, I'll run into quite a few customers that we'll, we'll talk about them believing that generative is everything and they forget that there's classic recognition models that they need or pattern recognition models that they absolutely have to have. Find it and then explain it kind of a thing as opposed to. No, the generative model is going to do it. All right, where do you see like a good graph environments coming into play for these? Because I feel like there's a lot of customers who miss the whole boat on the ability to use a graph model, especially with data lake and things. What are your thoughts there?
Ron Green: Yeah, we're seeing more and more utilization of graph, especially things like sort of graph based rag meaning, you know, graph based retrieval, augmentation generation. It's really, really especially handy in situations where you have relationships that need to be a part of the modeling process. And so I always think about it like this. You know, there's the famous no free lunch theorem that there is no one algorithm that will always be optimal for solving any problem. And if you think about computer vision and the use of like convolutional neural networks and just, you know, from sort of a non technical perspective, the way to describe that is convolutional neural networks allow the model to kind of not really care about where something is in an image. It's sort of translation invariant. It's more than that, but that's one of the main ideas. Well, convolutions sort of bring that inductive bias to the model. It forces it right out of the gate to not care if the cat is in the top left or the bottom left. It's still a cat with graphs. Graph neural networks. The beautiful part about that is you're bringing this inductive bias to the model where you're saying relationships really, really matter, and then you're forcing the model to learn through sort of that lens. And I think that's incredibly important. The interesting thing, maybe one, you know, kind of one more note on that is that's also the main reason we're seeing such amazing results right now with the other AIs. We've been talking around large language models and text to video and things like that. It's because they're based on the transformer architecture, which also allows that sort of contextual weighted attention to be applied in the model. And so one of the big takeaway, I guess the point I'm trying to make is anytime you can align inductive bias, meaning forcing your model to kind of think or operate in some certain way that aligns with the problem, you're going to get, you're going to get better results and you're actually going to streamline the path to explainability. Yeah, I think of this as
Chris Featherstone: it's in any AI models, like even when I'm talking to my family or folks that aren't in tech, right. The goal is always accuracy, right? Accuracy builds trust and then you can actually understand how to use it in, in various ways. Part of that for me especially graph was always, like you said, super interesting around the metadata about all these different things. I worked early on Sparkle Base with one of my good friends, Barry Zane, and it was, it was, it seems like to me, in fact, I saw something on, on the, the web that was basically is this is 2024, the year of the graph. Right. And I think part of it is what you're teasing out is the fact that when we talk about generative and or pattern recognition models, the goal is always accuracy. Of course, part of that accuracy comes from, hey, if I want to take a, just a base foundation model or model off the, off the cart, I've got to do some really neat and interesting things around the guardrails of what I can and can't do with it, meaning ask it questions or what I can't ask it and things like that. And that comes in the form of that accuracy. The second part of that is I don't know all the edge questions that I could or couldn't ask it to come back with, maybe some false positives or false negatives. So let me put a graph environment in there, create all those associations, and at least what I'm doing is I'm increasing exponentially the ability to Gather way more accurate result sets because I'm going to take those, put them into an embeddings environment, vectorize it. Now I've got a really great representation as you know, in addition to all this not structured data to actually deliver my results. I feel like it's like almost essential especially talking about customer experience related items and fraud based items and you know, anyway, all those kind of things. Yeah, I couldn't agree more, 100%. So it's, yeah, it's, I'm always, especially with somebody gets so excited about some of these types of topics which I appreciate all the energy you bring too because it's, it's palpable just in terms of like, hey, there's something new out there. Well, we're gonna embrace it, figure it out and then this is how we do things. So. Yep. So let me, let me switch gears a little bit because it's always interesting to, to know the, the person behind all of the great work that you do. So tell me a little about yourself personally, like what are you hobbies, interests, all those things. And please don't say that I spent all my, my extra time, you know, training AI and ML models. Yeah, yeah, absolutely
Ron Green: I do. You know, it is funny to stay up to date. It is kind of shocking how much time you do actually have to spend a reading these days just to stay, stay on the, the cutting edge. But you know, probably, you know, some of my favorite pastimes are playing the piano. I'm pretty passionate about that I've been playing for, for decades. I would say I spend a good chunk of my dedicated to that big reader. Love to read, love to travel and you know, when it comes to AI, I just kind of feel like, you know, for sure for somebody who, who, who lived through the AI winter that, that, that AI winter from the 90s, early 2000s and to be have the entire field not only resuscitate and now going through this sort of of absolute renaissance, this revolution. It is, it's actually a joy when I spend my weekends reading white papers or prototyping with new models or trying out new packages. I mean I just, I spend most of my, most of my week with a smile on my face because I just am so excited about where we're at right now and where we're going in AI. I think the vast majority of people out there fail to recognize just how dramatic the changes are going to be in the near future in the most positive possible ways because we're going to, we're going to start moving up sort of that exponential curve. And the best example of that is just the fact that there's more and more evidence this year. I feel like every major model release I see, if you look, if you read the white paper, they say it was trained on a considerable amount of data that was output by the previous model. Right. So we're, we're, you know, the models are starting to stand on the shoulders of each other and it's just astounding. So I'm, you know, that is actually, you know, one of my favorite hobbies is to stay up to date and catch up on AI on the weekend.
Chris Featherstone: Do you think, do you think that you ever seen the movie Multiplicity? Right. If you multiply something of the multiplet, you know, multiplies and starts to diminish. Do you think we're going to reach those types of scenarios as well where the output of a model, as fictitious as it may be, may actually really poison the well of the next model?
Ron Green: Right, right, right. Yeah, that's a really good question. I have a couple of thoughts on that. One is, I think that if you naively take the approach that, let me back up and say this, the classic example would be, hey, we've trained this particular large Lang model, let's say it's chat GPT4 on all of the publicly accessible information on the Internet. And so there's no more training data. What do we do? Oh well, we'll have it generate data and then now you've got this sort of feedback loop where it's training on its own output and you just kind of, you end up with this sort of like, you know, averaging out effect. Right. And you end up with actually a loss in abilities. I think that sort of naive approach is definitely not the way to go. But what we're seeing are more sort of teacher student approaches where you can have multiple separate models making determinations and then grading the output of models. And so there's this, there's this recent move towards, I think it was Microsoft's paper Phi 3 on Phi 3 where they, where they were sort of Sebastian was joking about like textbooks for all you need. They found that if you, that if you concentrate on really, really high quality training data, you can build small models that outperform larger models because you're just basically concentrating that signal. And I think that that's, we're seeing that all over the place. In fact, last week Alpha Proteo, which is a sort of a protein fold prediction model from DeepMind, they said it was trained on 100 million predicted proteins from alphafold. So that means that this problem that was unsolved as of like five years ago and then which is that is how would an amino acid sequence fold into a protein. We now have models that are so good at solving that problem that you can generate 100 million training examples and then use that to train a downstream model model that has a new capability beyond just protein fold prediction, but it can actually have the ability to predict how proteins will bind with other proteins. And so I'm, I, I, that's a long way of saying I am deeply not worried about that problem right now. I think there's a lot of room at the top to grow still. That's awesome. So do you have a favorite
Chris Featherstone: proprietary LLM? Oh, good
Ron Green: question. I'm, I'm, I'm pretty excited about llama 3.170 b. I can literally run it on my laptop on like literally an Apple laptop and then Phi 3 which I actually mentioned a second ago because it's such a small model. Anything I can run locally on my laptop I'm a big fan of.
Chris Featherstone: I'm impressed with the speed and actually the long miles too. You know I work with a number of, of a lot smarter folks than, than I am and, and we're, we're doing use cases where we need instant response on a single stream it of course and things like that and it's amazing. It's coming but it really
Ron Green: is. Well listen Ron, thanks so much for, for the time. Is there
Chris Featherstone: anything that, that I should have asked you that I didn't? No, no.
Ron Green: This was a really fun, wide ranging conversation I would say feel free to reach out to me anybody listening if you have any questions. Ronungfu AI I'm happy to connect and I have podcasts called Hidden Layers if you want to dig in on different AI technical topics, check it out. Awesome. Yeah, we'll make sure that
Chris Featherstone: way to get a hold of you and to see where you're at on the web is in the show notes and stuff. But listen, thank you so much for spending the time and taking time out of your busy schedule to do this and wealth of information and really really appreciate it.
Ron Green: I really appreciate it too Chris. Thanks for having me. Thanks brother. Take care. Take care. Visit early.com to find links to the full
Chris Featherstone: podcast on all audio platforms to listen on the go. Thank you for watching. Contact us today@infoearley.com.