
Sidecar Sync
Welcome to Sidecar Sync: Your Weekly Dose of Innovation for Associations. Hosted by Amith Nagarajan and Mallory Mejias, this podcast is your definitive source for the latest news, insights, and trends in the association world with a special emphasis on Artificial Intelligence (AI) and its pivotal role in shaping the future. Each week, we delve into the most pressing topics, spotlighting the transformative role of emerging technologies and their profound impact on associations. With a commitment to cutting through the noise, Sidecar Sync offers listeners clear, informed discussions, expert perspectives, and a deep dive into the challenges and opportunities facing associations today. Whether you're an association professional, tech enthusiast, or just keen on staying updated, Sidecar Sync ensures you're always ahead of the curve. Join us for enlightening conversations and a fresh take on the ever-evolving world of associations.
Sidecar Sync
AI Innovations from Gemini and DeepSeek to Break Your Brain | 76
In this episode of Sidecar Sync, Amith and Mallory dive deep into two major AI model releases—Google’s Gemini 2.5 Pro and DeepSeek’s V3—and explore how they’re reshaping the landscape of artificial intelligence. They discuss the technical breakthroughs, from million-token context windows to open-source innovations, and what it all means for association leaders. This is a jam-packed episode with strategic insights for the forward-thinking association exec.
🔎 Check out Sidecar's AI Learning Hub and get your Association AI Professional (AAiP) certification:
https://learn.sidecar.ai
📕 Download ‘Ascend 2nd Edition: Unlocking the Power of AI for Associations’ for FREE
https://sidecar.ai/ai
📅 Find out more digitalNow 2025 and register now:
https://digitalnow.sidecar.ai/
🚀 Find out more about the upcoming Blue Cypress Innovation Hub! https://bluecypress.io/innovation-hub-chicago
💡 Join Sidcar’s AI Mastermind Group:
https://sidecar.ai/association-ai-mastermind
🛠 AI Tools and Resources Mentioned in This Episode:
GPT-4o ➡ https://openai.com
Gemini 2.5 Pro ➡ https://deepmind.google/technologies/gemini
DeepSeek-V3 ➡ https://huggingface.co/deepseek-ai
Midjourney ➡ https://www.midjourney.com
HeyGen ➡ https://www.heygen.com
Amith’s Comic Strip ➡ https://shorturl.at/2Khn5
Chapters:
00:00 - Introduction
02:02 - Catching Up: Innovation Hub in Washington DC
07:58 - Gemini 2.5 Pro
15:58 - Google’s Position and Counter-Positioning Strategy
20:57 - Visual Breakthroughs in GPT-4o’s Image Generator
24:00 - Small vs. Large Models
28:36 - DeepSeek-V3 Release: What’s New and Different
33:29 - Why Open Source Models Matter for Associations
38:04 - The Impact of Lower AI Costs on Strategy
41:23 - Dream Bigger: Breaking Constraints and Thinking Abundantly
44:48 - Final Thoughts
🚀 Follow Sidecar on LinkedIn
https://linkedin.com/sidecar-global
👍 Please Like & Subscribe!
https://twitter.com/sidecarglobal
https://www.youtube.com/@SidecarSync
https://sidecarglobal.com
More about Your Hosts:
Amith Nagarajan is the Chairman of Blue Cypress 🔗 https://BlueCypress.io, a family of purpose-driven companies and proud practitioners of Conscious Capitalism. The Blue Cypress companies focus on helping associations, non-profits, and other purpose-driven organizations achieve long-term success. Amith is also an active early-stage investor in B2B SaaS companies. He’s had the good fortune of nearly three decades of success as an entrepreneur and enjoys helping others in their journey.
📣 Follow Amith on LinkedIn:
https://linkedin.com/amithnagarajan
Mallory Mejias is the Manager at Sidecar, and she's passionate about creating opportunities for association professionals to learn, grow, and better serve their members using artificial intelligence. She enjoys blending creativity and innovation to produce fresh, meaningful content for the association space.
📣 Follow Mallory on Linkedin:
https://linkedin.com/mallorymejias
You shouldn't think of AI video generation or AI whatever as a scarce resource, that you can only use it in a tiny number of areas. You should be thinking that by the end of this year, you'll probably be able to do it across everything. Welcome to Sidecar Sync, your weekly dose of innovation. If you're looking for the latest news, insights and developments in the association world, especially those driven by artificial intelligence, you're in the right place. We cut through the noise to bring you the most relevant updates, with a keen focus on how AI and other emerging technologies are shaping the future. No fluff, just facts and informed discussions. I'm Amit Nagarajan, chairman of Blue Cypress, and I'm your host. Greetings and welcome to the Sidecar Sync, your home for content at the intersection of associations and AI. My name is Amit Nagarajan.
Speaker 2:And my name is Mallory Mejiaz.
Speaker 1:And we are your hosts. Today, we are going to cover some crazy and awesome and exciting things that are happening at the forefront of artificial intelligence, and we're going to tell you how they might apply to your world as an association leader. Before we do that, though, let's take a moment to hear a quick word from our sponsor.
Speaker 2:If you're listening to this podcast right now, you're already thinking differently about AI than many of your peers, don't you wish there was a way to showcase your commitment to innovation and learning? The Association AI Professional, or AAIP, certification is exactly that. The AAIP certification is awarded to those who have achieved outstanding theoretical and practical AI knowledge. As it pertains to associations, earning your AAIP certification proves that you're at the forefront of AI in your organization and in the greater association space, giving you a competitive edge in an increasingly AI-driven job market. Join the growing group of professionals who've earned their AAIP certification and secure your professional future by heading to learnsidecarai. Amit, how are you doing today?
Speaker 1:I'm doing great. I just got back from DC. I think I may have picked up a little bit of a cold or something on the way back, but here I am. I'm back in New Orleans and doing well. How about yourself?
Speaker 2:I'm doing pretty well myself. Nothing like a little dose of sidecar sink to make you feel better, right?
Speaker 1:That's right. It's like an adrenaline shot. It's awesome.
Speaker 2:Exactly. I have been itching to find out how the Innovation Hub went. I was not able to attend and I haven't really talked to you yet, amit, about it, so this is the first time I'm hearing. But how was it? How did it go? How was your session?
Speaker 1:It was amazing. So this is the third annual DC Innovation Hub. We have the Chicago one coming up in two weeks in Chicago April 8th, and that's also going to be an amazing event. It's at the American College of Surgeons office downtown and the DC Innovation Hub was at the American Geophysical Union just north of DuPont Circle in DC. Beautiful location, amazing conference, wonderful hosts. We thank our friends at AGU for that and we had great turnout. We had about 70 people show up. This is our small community event we do in DC and Chicago each year. Our main flagship event, as many of you know who are listening, is Digital Now, which will be in Chicago this fall.
Speaker 1:But back to the Innovation Hub in DC. It was really a cool moment in time, I think, in a way Mallory, in that we really are seeing people build. They're past the contemplation stage. Many of the people that were in the room were there to listen and learn, but they were also there to share. That's the idea behind the innovation hubs. We started those as informal community gatherings to certainly share some content and things that various folks across our family of companies are doing, but more than anything to kind of take a feel of the community and say, hey, like what's going on, and so people are talking about deploying AI in a lot of interesting ways. So it was super fun. I learned a ton, met some great people I hadn't had a chance to meet in person before. All around, Really well worth it, and I hope those of you that are listening that are even somewhat close to Chicago consider joining us on April 8th. We still have a little bit of room left.
Speaker 2:Were there any key takeaways or any like challenges, any patterns that you noticed emerge of all the associations there said, oh, we're really struggling with X, anything like that.
Speaker 1:You know, I think, more than anything, people seem ready. So even last year at this time, the way I felt was that people were still nervous, they were contemplating, they were learning, but they were still kind of like, hmm, I don't know if we should do this. I'm not sure if AI is ready to support the critical work of our association at scale. A lot of people a year ago were already, you know, doing personal experiments and actually, in some cases, significant work with tools like Claude and ChatGPT. But to deploy it as an organization, right to go out into the world and say this is our association's AI for knowledge or this is our association's tool for search or personalization, few were doing that. And this time around, what I felt was that people were saying, yeah, we are doing this. It wasn't a question of if, and I think that's exciting because you know AI is not going to wait for any of us, whether we're, you know, do-gooders in the not-for-profit sector or somebody else. Like AI is not hanging out and just saying, hey, we'll wait for you however long you want. So I was excited to see people taking action, because that's really the learning loop, right.
Speaker 1:We talk at Sidecar all the time about the AI learning journey. It's not an event. It's a continuous and forever process, and that's true in any domain. It's definitely true in AI, and when you go and do the thing yourself, that's when you really learn and you build organizational reps. That leads to organizational strength, obviously leads to culture change. It's cool, so I was really pumped up about that. We heard about people talking about a lot of different kinds of AI as well, so some people doing interesting analytics stuff, some people are doing personalization at scale, some people are doing knowledge assistant work. So it was a lot of fun.
Speaker 2:That's awesome. I feel like you and I, and through the podcast, through Sidecar, getting to witness this journey from its inception all the way to now, all the way into the future. I'm thinking of the AI Mastermind group that Sidecar runs for Association C-suite leaders and thinking how, in its early stages, the AI Mastermind group we had to lead most of the sessions because there just weren't any use cases out there. It was more abstract. We were talking about concepts, getting your data ready, ready strategy, and now in our AI mastermind iteration, we're having more and more participants present because they're actually like rolling out these projects themselves. So it's been really neat to see in 75 episodes how far we've come.
Speaker 1:Yeah, I think we had to with the mastermind. We had to get the party started. But now it's raising.
Speaker 1:So it's pretty cool. And that mastermind group is awesome. So a good friend of ours, mary Byers she and I are the leads of the mastermind. We host this virtual monthly 90 minute session with a small group of engaged leaders from a variety of different associations, and if you're interested in going deep on AI once a month, consider joining that. We have information on our website there as well.
Speaker 1:But, yeah, you're right, I mean this journey that we're on. We are witnesses to it, as you pointed out, Mallory, and we hope to be, you know, in various ways obviously, educators and sources of inspiration, but really a conduit through which the community can share their experiences with AI. And so, to build on the point I made with respect to Innovation Hub, what we'd love to do is hear more from our listeners, from the folks who watch us on YouTube. Give us feedback on things you're doing. In some cases, we may be able to bring you on the pod as a guest or feature something that you're doing in an article on the Sidecar blog, in our newsletter, etc. We'd love to hear from you. That's the most powerful thing you know. We can talk about our perspective as the leaders of Sidecar and Blue Cypress. At the end of the day, what matters is what you guys are doing.
Speaker 2:In today's episode, we are talking about AI models. Is that a surprise? We're talking about the release of Gemini 2.5 Pro, and then we'll be talking about DeepSeek's latest upgraded model. So, starting off with Gemini 2.5 Pro, it's Google's latest and most advanced AI model introduced this week to the public. Gemini 2.5 Pro is part of a greater trend of thinking models or reasoning models, which are advanced AI systems designed to mimic certain aspects of human thought processes, particularly in problem solving and logical reasoning. These models use complex algorithms and techniques to analyze information, draw conclusions and make decisions based on that analysis decisions based on that analysis. We've chatted at length about Claude 3.7, OpenAI's O1, and now Gemini 2.5 Pro is joining the ranks. The model uses techniques like reinforcement learning and chain of thought prompting to simulate reasoning. This process involves the model thinking through its responses by verifying facts and logically deducing answers before providing them With a large context window of 1 million tokens. Gemini 2.5 Pro can process extensive amounts of data, and not just text. It can also process audio images, videos and large datasets like entire code repositories. The model's agentic coding abilities allow it to create complex applications like fully functional video games from a single prompt.
Speaker 2:I did a little brief test with 2.5 Pro in the Google AI studio before we recorded this pod, and so far I like it. It breaks down exactly what it's thinking, which a lot of the models that I mentioned do as well. You all know that we produce blogs from our Sidecar Sync transcripts. If you've listened to the podcast before, we've talked about this several times. Typically, my go-to model for that would be Claude, but I decided to run a little experiment with 2.5 Pro and I was quite impressed, mostly because I like to start that process by asking identify three topics from this podcast that we could write a blog about, instead of just asking it to generate an entire blog at once. Identify three topics from this podcast that we could write a blog about, instead of just asking it to generate, you know, an entire blog at once, and it gave me some really compelling topic ideas, and then it also provided support from the podcast.
Speaker 2:So at minute 10, Amit mentioned this. This is why I included it in this topic, so I thought that was interesting. That's something that Claude does not typically do when I use it. Amit, what are your initial thoughts with 2.5 Pro? Amit Bhandari.
Speaker 1:I'm pretty impressed with it. I haven't personally sent it a single prompt, so I'll disclose that I intend to over the next couple of days. But I have seen several videos of demos of Gemini 2.5 Pro you, an AI. The context window really refers to the amount of short-term memory it has. So when you send a prompt in and kind of the history of your conversation with that model not all conversations, by the way, but that specific conversation you're on in chat, gpt, in cloud or in the Google Studio the aggregation of all of the back and forth you have with the model that all has to fit into what's called a context window and there's a variety of techniques for dealing with really long conversations. But the idea is that the bigger the context window in theory, the more powerful the model could be because it has more of this short-term memory. So when we talk about a million tokens a token being approximately equal to a word, for our purposes that means that's a lot of words. That's something on the order of magnitude of 15 to 20 business books. It's a lot of content, whereas the other models that are out there that are similar in intelligence, like Cloud 3.7 and GPT-4.0, those tend to be limited to 128,000 tokens, so about an eighth of the total capability of these models, sorry, of this particular model, gemini 2.5 Pro. Now, google has been a leader in long context models for some time, since the first Gemini release. Actually, they had very large context windows. Some of the ways to think about this for associations is that if we have really complex tasks we want to take on, where we want to feed in many different pieces of content let's say, from our journals or transcripts from conferences, and we want to be able to look across a lot of content at the same time, gemini is a tool that stands on its own at the moment because these other tools are limited essentially to this fairly small context window on the one hand. But remember original ChatGPT, if you recall, had 4K of context, so 4,000 tokens. It was very, very limited. So in any event, the point I would make is that by itself is cool and really more than anything. That's not new, that's just a feature of Gemini that seems to be a key differentiator for big, complex pieces of content, but really the intelligence of the model is pretty amazing.
Speaker 1:So one of the people I follow on YouTube, this guy named Matt Berman. If you like slightly more technical content. He's a great YouTuber to follow. I watch quite a few of his videos and he breaks down fairly complex topics in a really nice way, in my opinion. Anyway. So he had this video showcasing using Gemini 2.5 Pro for coding, and I tend to look at those examples fairly quickly when a new model comes out, because coding is both it has to represent, both the ability to do like fairly complex reasoning but also to understand pretty complex prompts, especially with these days.
Speaker 1:You know, people are putting in requests to coding tools like this to do very complex things Like the two examples in his videos. One was a Rubik's Cube simulator, which was essentially a three-dimensional Rubik's Cube simulator of any number of dimensions, so it could be three by three, six by six, a hundred by a hundred, and the AI was asked to build the codes that you could visually represent this. You could spin it around, you could see it from any angle, you could zoom in and out, you could pan tilt, et cetera, and it did that. And then, on top of that, the code it was requested of the AI to make it so that the AI itself could solve the Rubik's Cube, so it could randomize the cube state and then it could solve it and you can watch it visually solving the Rubik's Cube, which is pretty cool. What's impressive about this is that in this video it was a single shot right, so it was just a prompt, and then immediately had a working piece of code in a browser in a single HTML page that did this. That's a non-trivial bit of software development to write that right. Even a really good developer would take quite a bit of time to build something of that order of magnitude. The other example was also similarly three dimensional visual kind of thing in the browser to build a Lego simulator, to be able to snap Lego bricks together of any sizes and shapes and colors, and what he demonstrated there was pretty cool. So I found that particular example really compelling, both because it was clearly showing the ability to do fairly complex reasoning with that level of coding.
Speaker 1:This is not a trivial coding exercise. Like you know, some simpler games that people have asked like build a snake game in Python, which is I wouldn't say that's trivial, but it's fairly simple. Comparatively speaking, this is an order of magnitude more complex. So that was impressive. I think your example was great and when we, when we're talking about this, just continuous evolution of these models. The thing we always have to point out is there's so many options, right. So this is now Google getting into the game in a way that I think really puts them more on the radar for a lot of people. We talk about OpenAI, we talk about Claude, we talk about the open source models, but you know, google's just been kind of behind, at least in the public perception and in terms of usage. Certainly they've been a distant, maybe not even third place is what I was going to say maybe fourth or fifth. So I think this is going to put them back on the map for a lot of people to really consider and think deeply about.
Speaker 2:And that was one of my follow-up questions too is, at least from my perspective, it sounds like from yours too. Google seems to be late to the party oftentimes when it comes to AI, but when they arrive to the party, right, they're well-dressed, people want to hang out with them, like this model is really impressive, but it seems late. So is that kind of Google's MO you would say, like do they take more time to bake things because they want to put out something really quality?
Speaker 1:I definitely think that is an element of it, that I think, in fairness to them, I think that's a part of what they need to do because of their scale and because of their brand that they want to put things out that are fairly well thought out.
Speaker 1:The flip side of it is, I think they were just behind and you know they were not behind in terms of fundamental AI research. They've been leaders in that in many ways for years and years, and, for those that aren't familiar, google actually invented the transformer architecture, which is the type of neural network that has powered all of the language models you've been hearing about, including the original chat GPT, and it still powers the vast majority of language models. Back in 2017, they invented that architecture, and so they know a thing or two about AI. These are some really smart folks with a lot of resources, but, you know, one of the things to maybe consider also is organizationally. These guys are the incumbents in search, and so when they saw LLM start to scale quickly, they're thinking about their own business model, and so it all of a sudden made it, you know, a strategically critical thing to be in this game.
Speaker 1:So I don't know if they're thinking about it more from the perspective of how do we use it within Google search and other Google products, and that's their number one priority versus producing models for the rest of the world to use, as compared to OpenAI and Anthropic, whose only purpose is to produce models other people use. So I don't know if it might be that perhaps you know it's also in any organization, no matter how big and how well resourced and how smart the people are, you have to, you have to make priority choices. I'm entirely speculating about this, but I sense that there's elements of that going on.
Speaker 2:It seems like Google might need to displace itself which we've talked about before on the pod, this idea of counter positioning to displace its own search function or at least greatly change it. And it's kind of already done that with AI overviews, which have gotten better in my personal opinion. But that'll be interesting to watch it play out.
Speaker 1:Definitely. Yeah, I think that you know counter-positioning oneself. It's better to be Netflix than Blockbuster, than Blockbuster. But it's a better idea to potentially say, hey, how can we be? You know, how can we essentially be the company that creates the new business model that has superior customer experience ourselves. So I definitely see Google heading in that direction.
Speaker 1:You know, my bottom line is that there's so much happening in the area of models that you know, even those of us who spend all of our time thinking about this and talking about it and playing with these models and building software on top of these models, we can't keep up. And so just this week, you know, aside from the two models that we're talking about, with the new DeepSeek V3 and also Gemini 2.5 Pro, we also had a release from OpenAI with the GPT-4.0 new form of image generation, and that is it's a really stunning capability. So it's a capability. If you haven't experimented with it since it came out, I really recommend that you get on ChatGPT and ask it to create an image, or take an existing image, drop it in and ask it to modify it. It's pretty stunning and that's gotten quite a bit of attention, but it's hard to keep up with all this stuff. The advice that I always give people about that issue of how to keep up with this is that look for the trend line right, look for the pattern and look for how you would like to use these models, not just the specific model. What are the problems you're trying to solve today? If it's a use case, you already have working totally fine for the last 12 months with GPT-4, maybe you're not so excited by like this new model. That's an even better blog, when the existing blog was pretty darn good. Of course, you know those of you listening to this podcast probably are always looking to take the next step and improve and evolve what you're doing.
Speaker 1:But what I also think is interesting and important is think about the things you couldn't do with AI. So my example where I did have hands on experience in the last 24 hours, mallory is I used the new Chachi PT 4.0 image generator to create a comic strip and I threw it on LinkedIn. It's not perfect, it's not great, but it just talks about this conundrum that associations are in with respect to AMS replacement. Right, our favorite topic of the pod for a lot of people other than AI, of course, is AMS replacement, and you know we're not trying to, you know, pick on it, but it's a tough, tough thing to do and it's really long and really expensive and the value creation oftentimes is marginal at best. And so you know why do that when you could do a lot with those resources experimenting with AI. So that's what I just basically said.
Speaker 1:What I just said to open AI's model yesterday and I got a four panel comic strip. Then I said, hey, give me four more panels that kind of conclude, like what happens two years later if all you do is focus in your AMS and don't do AI, so go check out my LinkedIn if you want to see the comic strip. But in the past I've had this idea to create comic strips or infographics or whatever, and none of the image generators could do it. So had I not experimented with this, last night I would still be thinking in that mindset that wouldn't it be cool if I had this creative outlet to create comic strips or infographics with just an AI model? But in my brain I would say, oh, but they do a horrible job with text.
Speaker 1:Oh, but they can't do different styles, like a comic strip style is very different than the kind of you know. We all know these AI images have started to look like for the last couple of years. They're very, very similar. But now my brain is all fired up about, like, all these new capabilities we have, right, so experimentation is good, but the trend line is it's not just we're excited about OpenAI's image generator. All the other ones are going to be like that within half a minute, right, you're going to see the same thing for mid journey. Probably within days You're going to see the same thing from all the other multimodal models. We know Anthropic is no slouch with their cloud product, so it's exciting. So the trend line is image generation really can be used for things that even, like a day ago or two days ago couldn't be used. And that's what I keep looking for is what are the next use cases? That are the next unlocks?
Speaker 2:Yeah, that's a great point. Your comic strip was good. I highly encourage you all to go to Amis LinkedIn. We can include that in the show notes as well for you to check out. I have never asked Midjourney, which is my preferred image generator, to do a comic strip, but I'm pretty sure it wouldn't do a good job with it, based on how frequently I use it. So I'm going to use GPT-4's image generator the next time I publish a blog, which might be today, and see what I can come up with.
Speaker 1:The publish a blog, which might be today, and see what I can come up with. The text side of it is really compelling because the image generation, even in the past, when we've had really stunning images generated by AI, when they have attempted to incorporate text, whether it's a sign in the background.
Speaker 1:It's always been, you know, kind of garbled up, and even when you prompted the AI to say do not include text, it would still oftentimes include text. So this really does represent a pretty significant leap in image generation and that's useful in so many respects. You know, you think of it as well again, when we're going from something that was once scarce to something that is abundant. You know, a comic strip would take a lot of work to put together. Right, you need talented illustrators, you need the idea for the comic strip, you need all this stuff, and I certainly wouldn't be attempting to do that. I really can't do much more than a stick figure to save my life, you know.
Speaker 1:But I have ideas and so I'd love to be able to express those ideas in different ways that are both interesting and, hopefully, effective in communicating. So I find this really, really exciting. So then you take the comic strip and you say, hey, take an image to video generator and now animate that comic strip and add audio to it, and you know, so, on and on and on. This goes right.
Speaker 2:Wow, talking about trend line, I want to zoom out a little bit. It seems like we've been talking a lot lately about these thinking models that I mentioned, or reasoning models, which seem to be, in general, pretty large models in terms of parameter size, but in the prior few months we spent a lot of time talking about small models. So I'm kind of wondering how you see small models fitting in to these greater thinking and reasoning model conversations and do you see like that chain of thought, prompting and reasoning as something that will be needed in small models, or should we just leave that to the big ones?
Speaker 1:Well, I think that the simulation of the chain of thought, reasoning, is something that some of the small models are starting to do in some ways, where they're not really doing reasoning in the same sense as the bigger models. Now, gemini, 2.5, pro, all use this internal thinking process where, rather than trying to give you an answer as quickly as possible, they decide that the problem is complex enough, that they're going to break down the problem into steps and they're going to solve each of the steps one at a time, typically sequentially, although not always, and then bring back the results from each of the steps, compile a result and then evaluate the result and then possibly iterate again and again until the model determines it's gotten to a good result. And this process is this reasoning process, as it's called is compute, intensive, it takes more time, it takes more compute and it produces remarkably improved results compared to just quickest possible answers, which is you know, we've compared it in the past to the idea of system one and system two, thinking of the intuitive, reactive thinking versus the reasoning style thinking that our biological brains are doing so well, I think. Coming back to your question about small models, will they incorporate actual reasoning processes? I'd be surprised if they didn't at some level.
Speaker 1:But at the same time, part of what we are seeing is people distilling into smaller models things that reflect the intelligence of larger models, and we've seen that from big LLMs to small LLMs for some time now, for a couple of years, where the small models keep getting smarter and the model architectures are getting better. In some ways they're getting smarter, but it's also about how the data set that the small model is being trained on reflects back on what the big model knows. So, for example, lama 3.3, the $70 billion parameter kind of mid-sized model that was released in December is smarter than the LAMA $405 billion parameter model that was released last April. And the way they did that is there were some advancements in the tech, but what Meta did is they used LAMA 405b to generate a bunch of sample data that they then trained the new model on. So this distillation process is a way of really taking the essence of the intelligence of the larger model and packing it into a much smaller frame, and I think you're going to see that with reasoning. Specifically, you know, to foreshadow our next topic with respect to DeepSeek v3, that's actually a lot of what they did taking R1's reasoning and dropping some of that intelligence into a non-reasoning model in v3. But you're going to see more and more of this.
Speaker 1:What I keep coming back to and this is particularly important for our friends in the world of associations and nonprofits is that you know this is a somewhat technical topic it's important to understand as a business leader, because we're at this forefront of this emerging technology.
Speaker 1:We want to know what's possible, how we can use it to better serve our constituents, and on and on. But eventually and I think that eventually might be in the next 12 to 24 months you're not going to be talking too much about is the model a reasoning model or is it a straight inference model? I think most of these models will have a reasoning ability and how much of that reasoning ability they use or don't use will be dependent on what you ask it to do, and that's essentially what Cloud 3.7 does, right? So Cloud 3.7 knows that it needs to go into thinking mode, just like you or I would if I'm given a really complicated math problem or something else that takes, you know, time to reason through. I'm going to do that. I'm not just going to guess at the answer, which is essentially what the, you know, the fast inference mode is what language models have been doing up until these reasoning models.
Speaker 2:Yeah, I think you're right. Looking back and again, this AI journey that we've been on since the beginning, I remember you and I would talk about multimodal models and models that can understand text and audio and images, and I feel like that's just a commonplace thing now. So maybe you're right, 12 to 24 months we won't even be having this conversation, but that's a good segue for topic two of today, which is DeepSeek's V3 that Amit mentioned. We saw this model upgrade this week as well from a Chinese AI startup called DeepSeek, which we recently covered in a prior episode, for its R1 model released. To give you a quick recap of that, its R1 model just cost $6 million to train over 89 times cheaper than OpenAI's rumored $500 million budget for its O1 model and the release of R1 led to a $1 trillion drop in the US stock market. So lots of waves were made with DeepSeek and the plot continues to thicken. So DeepSeek has now recently upgraded its v3 large language model to DeepSeek v3-0324.
Speaker 2:Again, we love these AI model names. This new version is seen as another competitive attack on major AI players like OpenAI and Anthropic. Deepseek's v3 offers enhanced reasoning and coding capabilities compared to its predecessor. The model has 685 billion parameters, up from 671 billion in the previous version, and it's important to note that these models are open source, so the model is available under the MIT license on platforms like Hugging Face. Deepseq V3 challenges the dominance of US AI companies by offering competitive performance at lower costs, and the model's release underscores the growing presence of Chinese AI startups in the global AI scene, perhaps shifting the balance of power in tech development. I feel like Amit with DeepSeek the things to note are lower cost, competitive capabilities and the fact that it's open source. Do you agree with that?
Speaker 1:I think. So, you know, I think credit is definitely due to these folks. They are brilliant. The papers they put out there detail, you know, with a lot of granularity how they built these models. So they've open sourced not only the code and the weights of the models so anyone can run it anywhere, but they also have published a ton of research explaining how they've advanced their technology. So this is not a copycat or a clone. They have fundamentally advanced the science of AI and they're contributing back to the global community. So I view this as a very positive thing and I'm hoping to see this from a lot of other parts of the world jumping in because, especially as you see, the resource constraints seem to decrease over time as people get really creative when they have limited resources, and that leads to fewer resources being needed or being perceived as being needed Right, and that, in turn, leads to more innovation, which leads to more and more growth, and that's that's exciting.
Speaker 1:So to me, I think that that's one really important thing to recognize. One of the things that these guys have done a really good job of is they've advanced the way mixture of expert models, or MOE models, work. It is a 600 plus billion parameter model, but you can still run it on fairly. You know fairly reasonable hardware pretty quickly because it only has about I think it's like 32 billion parameters are active on a per token basis. So what that means essentially is it's really behaves like a 32 billion parameter model and that's and that's not a particularly huge model. It's half the size of Lama 3.370 that I just mentioned a few moments ago. So that's important because they've been able to really optimize with this mixture of experts model and do some other things around efficiency. They have a highly performant and really intelligent model.
Speaker 1:All the other players that are out there are really paying close attention to everything DeepSeek is doing, incorporating their ideas. So probably today already we've seen OpenAI and Anthropic and Google incorporate DeepSeek's ideas, and on and on right. And so the proprietary vendors don't really share what they're doing, but I guarantee they're taking advantage of open source and the open source community keeps on compounding on itself. So I would point out, the open source bit probably is the most important part of it, because these things are moving so quickly.
Speaker 2:And when you say mixture of experts, architecture, that just means parts of the model, the only parts that activate within the model, or the parts that are needed for that prompt.
Speaker 1:Exactly so. You know, if I say, for example, I would like DeepSeek to write code, maybe there's a portion of that model 600 plus billion parameters that are focused on coding, and only those. You know 30 billion parameters get activated. If I ask it to, you know, write a poem, it might be a different section of that. You know that synthetic brain essentially that gets activated. So it's a very large brain, so to speak, but it only uses portions of it at a time and it's not running the whole thing.
Speaker 1:And you know, most of these AI architectures are MOE or mixture of experts models. What these guys have been doing a good job with is making them more efficient, making them more performance. Part of what's happening is is like who makes the decision about which section of the model should activate for a given token, and that's something that, of course, is really important, because if you don't route that to the right part of the net, you don't actually get great results. So I would give these guys a lot of props, as I've been doing, in terms of their scientific advancements. You know my point of view in terms of open AI, and probably Anthropic as well, but more so open AI is that. You know my point of view in terms of open AI, and probably anthropic as well. But more so open AI is that you know you have these companies that have had I wouldn't say unlimited, but they've had substantial resources and their perception has been that they need those substantial resources to produce these. You know world-class gains and here you have a challenger with far fewer resources, not just money, but also they used equipment that's considerably less powerful. Due to export restrictions, they don't have access to the latest chips, at least that's what's been reported. So you know, that tells you a lot.
Speaker 1:Again, constraints can be incredibly powerful. If you give people a very narrow time frame to do a job, a lot of times you get a better result than if you say you know how much time you need or if you give them longer timeframes. I'm a big fan of setting narrow deadlines for small chunks of work. I don't like saying, hey, what's your, what's your priority for the next 12 months. I'd rather know what you're going to do in the next week. Not that I don't think about the next 12 months, but like it's more about how do you, how do you put a constraint? That's near term, and the smaller the constraints usually the more creative people get.
Speaker 2:Necessity breeds invention, which we've said on the pod before. Amit, I didn't realize this and maybe you had heard it, but there was a bill introduced in the House last month potentially banning federal employees from using a Chinese AI platform, DeepSeek, on their government issued devices. So I imagine there might be some organizations that would be dissuaded from using DeepSeek on their government-issued devices. So I imagine there might be some organizations that would be dissuaded from using DeepSeek models under the threat of a potential ban. But on the flip side, the models are open source. So can you explain whether that is a valid concern or not, or what to keep in mind there?
Speaker 1:It's totally a valid concern if you're using their website. So if you're going to deepseekai, I think is the website, if you go to that website, that model that you're going to deepseekai, I think is the website that, if you go to that website, that model that you're interacting with is hosted in China, which is, you know. It's not an inherently bad thing to have a model hosted in another country, but the point would be that if you're a government employee and you are asking a model, something related to you know, whatever it is you're working on, maybe that isn't the best thing to send overseas, right? Maybe that's something we should be running within the United States and preferably in, you know, a government cloud of some sort that's secure. So I think that's one piece of it, but that should be separated from the idea of the model itself.
Speaker 1:It is an open source, open weights model. It can be reproduced and you know you can run it anywhere. You can run it in your own data center, you can run it in a public cloud infrastructure like Azure, aws, gcp, oracle, et cetera. You can run it. You mentioned Hugging Face earlier. They provide a variety of services as well. There's a lot of places you can run these models and so to ban the model, I think, is really misinformed, because the model itself is just a piece of software. That's totally something you can inspect. One of the really nice things about AI models is that if you say it's open source, by the way, versus open weights, they sound like similar things.
Speaker 1:The model itself. The actual number of lines of code behind these models is remarkably small. It's in the single digit thousands typically, or even smaller, and so the models themselves don't have a lot of code. It's all about the weights. So the weights are, of course, like harder to understand, right? They're just a bunch of numbers.
Speaker 1:But you if, like someone was worried about like a backdoor existing or like phone home you know phones home and sends your data back. That's not a thing Like you can verify that the software is not doing that. You can also contain these models in ways when you inference them in your own hardware so that there's no possibility of them doing that. So I think that the mindset should be that we care deeply about where we inference these models. That should be done, thought about from a security perspective, that should be thought from a vendor trust and vendor reliability perspective, and we should also deeply care about which models we use.
Speaker 1:But you know, whereas I probably would not sign up to do any workloads for any of our products in China right now, for a variety of reasons, one of which is just whether or not that will continue to be available, but also it's a question of sensitivity, obviously, of the data. But I'd be totally happy to run these models as long as they're inferenced in places where I believe there's, you know, a better degree of transparency, visibility, control, et cetera. So you know, we've talked about Grok, we've talked about AWS and Azure Foundry. You can run most of these models in most of those places.
Speaker 2:Okay, so it wouldn't make sense for the government to ban source code or like the actual weights.
Speaker 1:I don't believe that's what's being discussed. I think it's access to the website. So that's why I think it's important to separate using the model on your own equipment or on equipment that's run by someone you trust versus connecting to DeepSeek's website and using it as a consumer. I believe that is what the bill is intending to ban.
Speaker 2:Mm I want to zoom out again to the trend line because, as you mentioned with the first topic, that's what's important to keep in mind. This competition as a whole seems to be driving AI costs down, down, down, really quickly no-transcript.
Speaker 1:Having the knowledge that the costs are going down at such a rapid rate should open your mind to the possibilities of applications that you couldn't afford previously. So if you were to say, hey, we want to go through every piece of content your organization has ever published and we want to auto-generate a taxonomy from it, or we want to do other things with it, right In the past maybe that would have been prohibitive from a cost perspective. Maybe the quality wasn't great enough either. But let's just say that you think the quality is awesome now, but the cost might've been a factor where you're like oh, that would cost us $3 million to go through all that content and now it doesn't. Right Now maybe it costs you $3,000 or something like that. So these shifts in cost should open your mind up to at-scale applications, doing things at scale. Or what about all the unstructured data that you have in your email or elsewhere? What can you do with that kind of stuff? So the way I look at it is the applications. When cost is going down and performance and speed are going up. It opens up new possibilities. So that's important conceptually, because a lot of what we're talking about in this pod and in the rest of our work at Sidecar is plotting your course, really your strategic direction, understanding the tools, understanding the technology, understanding how they all fit together, but thinking about where you should go with this.
Speaker 1:So clearly, we want to optimize current business processes. That's the obvious part, right? We want to make what we do happen faster, better, cheaper, right? Traditionally, you'd pick maybe one or two of those, but you don't get all three. Now you can get all three. But the bigger question is what should you be doing? As opposed to how do you do what you currently do better? And the what should you be doing is informed by constraints, and the constraints are the capabilities of these AIs, but it's also the cost. So knowing that the cost keeps relentlessly going down should open your mind to the possibilities that six months, 12 months, 24 months from now, video generation will be close to free. For example, right now, if you want to do a lot of hours of video generation with HeyGen, you have to sign up for plans that will cost you in the five or even six figures. That's a constraint. We're, by the way, extensively using HeyGen in a variety of our content production, for our AI learning hub and for other things, and we're making the investment because we think it's worth it, but we also know that it's not really a long-term recurring investment because there is competition in every one of these dimensions.
Speaker 1:So that's part. One is that I think you can think bigger and think a little bit longer term, something that you know you wouldn't say oh well, I was going to invest a million dollars in a new AMS and next year it's going to be $100,000 and the year after it's $10,000. That's not a thing, right? That's not the way people think and that's not the way these systems tend to work. But with AI, that's literally what's happening with the fundamental models, because of competition, because of the amount of capital being thrown at it. So that's true for what I'd call kind of the raw materials of AI. It's the fundamental building blocks, really the models at that layer. It's also happening with inference competition, because so many people are going after being your cloud provider for AI, whether it's the hyperscaler major players like AWS and Azure, or if it's people who have a new approach to hardware, like our friends at the Grok which is GROQ, to always repeat that for clarity or a number of other folks that are out there doing cool things with inference. That's a hyper competitive market. You're going to keep seeing costs come down there.
Speaker 1:But to come back to the second part of your question about like well, how come we aren't seeing costs for the application layer which consumes these? Ais is for one, it's market economics. There's fewer competitors than there are at the model level. There's also more complexity in terms of integration with systems. There's kind of the nuance of what a lot of times is referred to as the last mile of the solution actually solving the problem. So you can take the model that has Gemini 2.5 Pro caliber thinking, but then to wire it up into your ecosystem and make it work just right for you. Still today there are specialized pieces of software that have to be integrated. There's a lot of labor that's required. There's domain knowledge in terms of what the best processes are.
Speaker 1:So the actual end solutions, I think you are going to see them come down in cost overall, but it's slower. It's kind of like if we said, hey, let's just imagine that the price of steel was close to zero all of a sudden, would that mean that the price of cars immediately goes close to zero as well? You know, when you're further down in the supply chain, you know that can happen. Right. Cost of materials can go up and down. Sometimes it has a ripple effect and sometimes it takes a lot longer to work its way through the system, but ultimately this is a really good thing.
Speaker 1:The component costs of the ultimate solution obviously are a major factor in that solution, but more than anything, it's this abundance mindset that I'm trying to really hammer into people's minds that you shouldn't think of AI video generation or AI whatever as a scarce resource, that you can only use it in a tiny number of areas. You should be thinking that by the end of this year you'll probably be able to do it across everything. So every single blog post that you have, let's have a great video on that post, every single time, completely AI generated. Today that might cost you tens of thousands of dollars. By the end of the year it'll probably be considerably less.
Speaker 2:I love that. I call it my version of that. So you said dream bigger. I call it often breaking your brain, because it's constantly, every day, all day, having to say, like what else could I do with this model, breaking these barriers of what you thought was possible, especially when we talked about member services. Right, having a member service agent that doesn't just answer basic questions, that actually has access to your whole knowledge repository and can answer domain specific questions, that's something that you have to really open your mind to. So I agree, I think that's essential and actually, now that I'm thinking about it, that it could be an interesting take, maybe for for Digital Now 2025. I don't know if we have a theme hammered out, but kind of the idea of dream bigger, breaking your brain.
Speaker 1:I like that. I like breaking your brain. I like that. I like breaking your brain. That sounds fun.
Speaker 2:Uh-huh, I don't know. So maybe you all are the first to hear our theme for this year. But with that, everybody, thank you for tuning in to today's episode. I know Amit mentioned Grok with a Q. We do have a special interview coming up. Not sure when we're going to post it, but in the next few weeks or so. So be on the lookout for that and we will see you all next week.
Speaker 1:Thanks for tuning into Sidecar Sync this week. Looking to dive deeper? Download your free copy of our new book Ascend Unlocking the Power of AI for Associations at ascendbookorg. It's packed with insights to power your association's journey with AI. And remember, sidecar is here with more resources, from webinars to boot camps, to help you stay ahead in the association world. We'll catch you in the next episode. Until then, keep learning, keep growing and keep disrupting.