Sidecar Sync

Intel Bets Big on Chips Again as Meta Drops Llama 4 | 77

Amith Nagarajan and Mallory Mejias Episode 77

Send us a text

In this thought-provoking episode of Sidecar Sync, Mallory and Amith dig into the fascinating world of semiconductors and how a historic joint venture between Intel and TSMC is reshaping the global tech landscape. They explore the underlying tensions between vertically integrated business models and specialization — a conversation that holds key lessons for association leaders navigating change in the age of AI. From reflections on the Innovation Hub Chicago event to an insightful breakdown of Llama 4’s powerful capabilities, this episode is a timely reminder that adaptability is everything — in both tech and associations.

🔎 Check out Sidecar's AI Learning Hub and get your Association AI Professional (AAiP) certification:
https://learn.sidecar.ai

📕 Download ‘Ascend 2nd Edition: Unlocking the Power of AI for Associations’ for FREE
https://sidecar.ai/ai

📅 Find out more digitalNow 2025 and register now:
https://digitalnow.sidecar.ai/ 

💡 Join Sidcar’s AI Mastermind Group:
https://sidecar.ai/association-ai-mastermind

🛠 AI Tools and Resources Mentioned in This Episode:
Llama 4 ➡ https://ai.meta.com/llama
GPT-4 ➡ https://openai.com
Claude 3.7 ➡ https://www.anthropic.com
Gemini 2.5 Pro ➡ https://deepmind.google/technologies/gemini
DeepSeek R1 ➡ https://deepseek.com
Aqcuired Podcast ➡ https://www.acquired.fm/episodes/tsmc
The Information (Article) ➡ https://shorturl.at/xymgG

Chapters:

00:00 - Introduction to Sidecar Sync and This Week’s Topic
08:36 - Semiconductor Basics Explained: Fab vs. Fabless
10:20 - Intel and TSMC’s Joint Venture
16:08 - Lessons for Associations from Intel’s Journey
20:53 - Specialization vs. Integration in Association Work
26:41 - How AI Changes the Capacity to Do More
31:39 - LLaMA 4 Overview: Scout, Maverick, and Behemoth
34:29 - Multimodal Models and Mixture of Experts
40:47 - Why LLaMA 4’s Context Window Matters
45:03 - Open Source, Privacy, and Inference Provider Choices
49:48 - Final Thoughts

🚀 Follow Sidecar on LinkedIn
https://linkedin.com/sidecar-global

👍 Please Like & Subscribe!
https://twitter.com/sidecarglobal
https://www.youtube.com/@SidecarSync
https://sidecarglobal.com

More about Your Hosts:

Amith Nagarajan is the Chairman of Blue Cypress 🔗 https://BlueCypress.io, a family of purpose-driven companies and proud practitioners of Conscious Capitalism. The Blue Cypress companies focus on helping associations, non-profits, and other purpose-driven organizations achieve long-term success. Amith is also an active early-stage investor in B2B SaaS companies. He’s had the good fortune of nearly three decades of success as an entrepreneur and enjoys helping others in their journey.

📣 Follow Amith on LinkedIn:
https://linkedin.com/amithnagarajan

Mallory Mejias is the Manager at Sidecar, and she's passionate about creating opportunities for association professionals to learn, grow, and better serve their members using artificial intelligence. She enjoys blending creativity and innovation to produce fresh, meaningful content for the association space.

📣 Follow Mallory on Linkedin:
https://linkedin.com/mallorymejias

Speaker 1:

Forget about what's happening to you and your association. But in your sector, what will people in your sector look for from the association? That's the open question. Figure that out and then go build that.

Speaker 1:

Welcome to Sidecar Sync, your weekly dose of innovation. If you're looking for the latest news, insights and developments in the association world, especially those driven by artificial intelligence, you're in the right place. We cut through the noise to bring you the most relevant updates, with a keen focus on how AI and other emerging technologies are shaping the future. No fluff, just facts and informed discussions. I'm Amit Nagarajan, chairman of Blue Cypress, and I'm your host. Greetings everybody and welcome to the Sidecar Sync, your home for content at the intersection of all things associations and semiconductors. Actually, I meant to say artificial intelligence, but today we are going to be talking a little bit about semiconductors and it's going to be a lot of fun and we'll tie it back to AI. We'll tie it back to all the things you're used to hearing on the Sidecar Sync, so we'll jump right into that momentarily. My name is Amit Nagarajan.

Speaker 2:

And my name is.

Speaker 1:

Mallory Mejiaz and we're your hosts. And before we jump in, here's a quick word from our sponsor.

Speaker 2:

If you're listening to this podcast right now, you're already thinking differently about AI than many of your peers, don't you wish there was a way to showcase your commitment to innovation and learning? The Association AI Professional, or AAIP, certification is exactly that. The AAIP certification is awarded to those who have achieved outstanding theoretical and practical AI knowledge. As it pertains to associations, earning your AAIP certification proves that you're at the forefront of AI in your organization and in the greater association space, giving you a competitive edge in an increasingly AI-driven job market. Join the growing group of professionals who've earned their AAIP certification and secure your professional future by heading to learnsidecarai. Heading to learnsidecarai. Amit the intersection of associations and semiconductors. I never thought we'd be talking about this on the podcast, but I mean it when I say I'm very excited to have this conversation.

Speaker 1:

I am too. I've been looking forward to this. I kind of geek out about this stuff and it's fun to understand a little bit more about what goes on under the hood, so to speak. You think about what actually powers AI and what powers really most things in the world. It's chips, and how are they made? Who makes them? These are questions a lot of people really don't know the answer to, and it's a complicated answer.

Speaker 1:

So I think that right now is a perfect time for us to spend a little bit of time unpacking that, because, you know, we are in this really interesting moment in time globally. We're at this interesting moment in time in terms of, like globalization or deglobalization is what I mean by that and also with respect to what has been really the leader of the pack for many, many decades in the United States Intel no longer being what they once were as an industry titan and really having fallen from that and what happens there. So I think that's going to be a really great conversation to have, so I'm excited about it. But you know, semiconductors power AI. We talk a lot about exponential growth. Yesterday at the Innovation Hub in Chicago, I had the pleasure of opening the event and talking about exponentials and you know a big part of it is the underlying advancements that's happened in semiconductor design, but also the manufacturing process. So it's just a. It all ties together in this way.

Speaker 2:

I'm excited to unpack this, but first you mentioned the Innovation Hub in Chicago, which I haven't really talked to you about yet, so this will be the first time I'm hearing. You said you kicked off the event. You also said you talked to a good few people about the podcast this podcast. So how did it go Like? What were the vibes at the Innovation Hub Chicago?

Speaker 1:

It was awesome. I had a number of people tell me that they really enjoyed listening to the podcast, which I know makes both of our days when we hear that and people were saying, yeah, it was just a really it's a helpful way to stay in sync with the latest that's happening in the world and in the world and a contextually relevant way for associations, which is, of course, our goal is to be with you on this journey, and that was awesome. And then, as far as the attendance, we had record attendance both in DC two weeks prior and in Chicago. These started out as really like informal small community gatherings about three years ago and this is the third time we've done this event and the idea was, hey, let's get together in the springtime in both Chicago and DC just to have some small local events, get the community together and share ideas around innovation, and we thought that would also be a good counterweight to Digital Now being on the calendar every fall, and so it's an opportunity to engage our community face-to-face. And both in DC and Chicago we actually had quite a number of people fly in for the event, which was I wouldn't say it was surprising necessarily, it was just it was a delight to hear that people were making that investment in time and dollars and energy to join us because they were saying the value was so strong. So that was awesome.

Speaker 1:

What I would say is similar to how I felt about the DC event a couple of weeks ago. People are doing things. They're not just talking about things, which a year ago. Some people were doing things, of course, but most associations were kind of just like testing the water a little bit and maybe starting to learn a little bit, but not really doing things.

Speaker 1:

Now I see associations actually running active experiments and starting to deploy technology at scale and seeing impact, seeing members tell them that their quality of service is better, seeing their engagement statistics improve, seeing their financial results improve as a result of improved service or services that they've been able to deploy. That would not have been possible without AI. So to me that's really gratifying to see people do that in the community. Of course, I do realize that our community of folks that are coming together, that watch our content and come to our events are, you know, self-selecting in that are probably a little bit ahead of the general curve of the association markets adoption. Nonetheless, I think it is, you know, setting great examples for associations who you know may be a little bit less advanced in their at the moment in terms of their AI journey.

Speaker 2:

Yep, yeah, that's great to hear. My follow-up question was going to be did you have the same takeaway of associations are actually doing this innovative work? They're actually working with artificial intelligence? But it sounds like you did. Were there any kind of distinctions between the DC and Chicago event?

Speaker 1:

It was much colder.

Speaker 2:

Okay.

Speaker 1:

So Chicago is colder. It was early April like, oh, it's going to be great, it's going to be beautiful early spring weather in Chicago and actually I would say on the day of the event it was a beautiful morning. It was about 29 degrees but it was crisp and the sky was blue and it wasn't super windy, so it was beautiful. And we were very happy to have the event hosted by the American College of Surgeons at their beautiful office on the top floor of their building. They have this unbelievable conference space overlooking the city and you can see the lake, so it's a beautiful space. Everyone seemed to enjoy that.

Speaker 1:

Being in Chicago is always fun. I love the city. It's a great place to go. Tons of wonderful association things there. It's just a fun place to hang out. I was there a few days early, caught the Cubs home opener at Wrigley, which was super cool, and then my son was in town as well for checking out colleges. We're taking a look at some universities up there for him is up there for him, so it was a great trip all around.

Speaker 2:

Awesome, awesome. Well, everybody, we won't have another round of Innovation Hubs until the spring of next year, but be on the lookout for those dates and sign up if you want to attend. All right, our topics for today. Well, you know, we're talking semiconductors. Particularly we're talking the Intel TSMC joint venture, which is kind of shaking up the semiconductor industry, and then we will be talking about the release of LAMA4, which we're really excited about. So I've got to give you a little bit of deeper background, I think, for the Intel TSMC joint venture, and then Amit will kind of do the same in our conversation after. So, bear with me.

Speaker 2:

Before we dive into this industry news, I want to just make sure we're starting with kind of the basics and the foundations. What is a semiconductor? So a semiconductor is a material that has electrical conductivity between that of a conductor, like metal, and an insulator, like glass. These materials, usually silicon, form the foundation of modern electronics. When precisely engineered with various elements added to them, they become the microchips that power everything from your smartphone to supercomputers. So now that we've got an understanding of semiconductors, let's talk about the business models that have been at play historically.

Speaker 2:

So, traditionally, the chip industry has operated under two main approaches. First, there is the integrated device manufacturer, orm model. Companies like Intel both design and manufacture their chips in their own factories, which are called fabs. They control the entire process, from concept to finished product. These functions so Fabless. Companies like NVIDIA, amd, apple and Qualcomm focus exclusively on designing chips, while foundries like TSMC, which is Taiwan's semiconductor manufacturing company, and Samsung specialize in manufacturing those designs for others. About 20 years ago more or less, we saw a major industry shift as companies like NVIDIA chose to go fabless, focusing all their resources on design while outsourcing manufacturing to specialized foundries like TSMC. This approach reduced capital requirements dramatically, since building and maintaining modern chip fabrication plants cost tens of billions of dollars. It also allowed for faster innovation cycles and let each company focus on their core strengths.

Speaker 2:

Tsmc emerged as the dominant foundry, especially for cutting-edge manufacturing processes. Intel, however, maintained its traditional IDM model. While this offered certain advantages, it increasingly struggled to keep pace with the specialized expertise of TSMC's manufacturing capabilities expertise of TSMC's manufacturing capabilities. That context is essential for understanding what might be the most ironic twist in semiconductor history. Intel's financial crisis may be over with support from its biggest rival. So Intel and TSMC have reached a preliminary agreement to form a joint venture aimed at revitalizing Intel's struggling chip manufacturing business. Under this partnership, tsmc will take a 20% stake in Intel's US-based chip fabrication facilities, with Intel and other US investors controlling the remaining shares. This represents a seismic shift in the semiconductor landscape. Intel, which has faced declining revenues and technological setbacks, reported a staggering $13.4 billion operating deficit in its foundry division in 2024 alone. Despite efforts to transform under its IDM 2.0 strategy, intel has struggled to compete with TSMC's advanced manufacturing capabilities. So this joint venture has huge implications for the capacity of chip production and ensures the US continues to have meaningful manufacturing capacity for advanced chips. With geopolitical tensions rising, particularly around Taiwan, where TSMC is headquartered, establishing robust American manufacturing capabilities has become a national security priority.

Speaker 2:

So, amit, that's a lot of context. You sent me an email with this topic idea maybe a few days ago, and you kind of included a long blurb with it to help explain this to me, because I was not super familiar with this whole industry prior and even with that email, I was a bit confused. So I went on a deep dive. I listened to the Acquired podcast episode on TSMC, which is fabulous. It is two and a half hours and I know what you all might be thinking like I don't know if I'm going to listen to that. It's really interesting how they frame the whole story of the founder of TSMC. So, amit, I want to get your initial thoughts, but I kind of also want you to lay the landscape on, like what this industry has looked like in the recent history.

Speaker 1:

Well, I think you did a great job providing an overview of the competing models, where you think about what is the fundamental business model and what has worked historically. And if you go back in time several decades prior to the era that you're describing, you know, which really kind of started in the 90s, where NVIDIA, you know, at the time was a startup and they started off as a fabulous chip company and they never had their own plants and many other companies followed suit and so these companies, as fabulous chip designers, were kind of this first generation that were able to go out there and essentially leverage, outsourcing. And at the time there were arguments against that, saying well, if you're not fully integrated, vertically integrated where you have your manufacturing, you have a weakness, you don't have an advantage there. But it allowed you to do what you described, which is the speed at which you could innovate when you have those. Separations of concerns are really another way to put it is essentially specialties, right Areas of focus, and you know people have been doing various forms of specialization of labor since the beginning of time and that has something. That's something here that you're seeing play out in chips. So Intel, prior to that era when fabulous manufacturing became a thing, was, you know, really the dominant chip manufacturer for the microprocessors specifically, that powered the PC revolution, and well into the internet boom, and their IDM model of both designing chips and then manufacturing chips was what AMD, their only other competitor in that particular category. Right, they did the same thing up until actually late 2000s, where AMD spun off their entire manufacturing business, and we can talk more about that later.

Speaker 1:

But the essence of that shift is really interesting to think about, because you had a company that had tremendous advantages in scale. They had advantages in process, right, because Intel was known as being extraordinary in its manufacturing process, and then, over a period of time, they started to have a decline, even though they had, effectively a monopoly. Not exactly, there was kind of a designed duopoly, with AMD being kind of a manufacturer that had licensing access to their IPs. They could make, you know, basically, these things called x86 compatible chips, which is, you could get a PC with either AMD or Intel CPUs for a long time. But the point and that was really a government satisfaction thing, so that Intel wouldn't be broken up. But the essence of what I'm describing, though, is a shift in an industry which is worth considering, because part of what was going on that enabled TSMC to rise is a maturity of the industry, the acceleration of the underlying technology that made it possible to manufacture chips at scale.

Speaker 1:

Prior to that era, the chip manufacturing process was so incredibly specific to the exact chip you were building that it was very, very difficult to consider, hey, I'll be able to manufacture chips for someone else, because it was really really. Every single piece of equipment was super custom. It was very early days of that so that as that process became more and more and more sophisticated and the machinery became more adaptable and more capable of producing different kinds of semiconductors, it made it possible to say, hmm, what if we actually had a separate sector focused on the fab process, which is the manufacturing process, and Intel completely missed it, right? So Intel, as strong as they were, really relied upon an aging architecture and while that architecture still powers the vast majority of PCs, they completely missed out on GPUs, which is the math processors that power graphics, but also power AI, which is a similar mathematical operation, and they obviously had this fundamental issue in their business model of IDM versus going fabulous. It's relevant to the AI revolution that we're in the midst of right now, but also because it represents an opportunity to reflect back on a dominant player.

Speaker 1:

Not that long ago, people would have thought of Intel to be a surefire blue chip stock to preliminary agreement, but it's not that far off from that in some ways too, which we can. We'll talk more about that in a sec. But I think associations need to pay attention because in many respects, they're the dominant player in their space. They have a business model that's existed for decades, or in some cases over a century, and they also live in a world that is rapidly changing. And so do they want to be the Intel or do they want to potentially adapt and find a better model because the environment's changed?

Speaker 1:

That's the main reason. Actually, I wanted to talk about this. I think the technology is interesting. I think the sector is something we can all benefit from learning more about. Most people know very little about semiconductors, know very little about all this stuff, so that'd be an interesting kind of educational topic. But to me it's more about hey, wait a second. There's a company that even 10 years ago, few people would have predicted the decline of Intel to this extent. And now here we are.

Speaker 2:

Yeah, man, I think this stuff is so fascinating because it's easy to look back with hindsight and say, well, yeah, it makes sense that we would break this out right into chip design and chip manufacturing. It doesn't seem that novel from our perspective, but listening to that acquired podcast that I mentioned, you realize at that time, when TSMC was founded, they were creating a solution to a problem that didn't really exist, because all of the chip companies were designing and manufacturing their own chips. So they were going out to startups and saying, hey, now you can form your company because we can help manufacture these chips. So it wasn't a given. It seemed very controversial at that time to have this idea, so I just wanted to upplay that a little bit. So, amit, would you say Intel, by going into this joint venture, preliminary joint venture, is essentially admitting defeat on the IDM model. Like that is no more.

Speaker 1:

I'm not entirely sure. I think that part of the thinking is Intel has a giant amount of manufacturing capacity in the United States. It's not necessarily all at the cutting edge, but some of it's close and with that footprint the idea is that TSMC can come in and help Intel build better chips. Basically, so TSMC is part of this deal. I think it's going to own a chunk of the company or it's some relationship along those lines. It's a little bit murky, based on the latest news that's out there, but ultimately what they're contributing isn't so much capital as it is process expertise. So we talk about Hamilton-Helmer's seven powers and we talk about the seven powers that show these routes to enduring differential returns, or really a strategic moat that some people would call it. One of them is called process power, and TSMC is an extraordinary example of process power, probably best illustrated this way that if you took all of TSMC's equipment and even all the documentation they have about how to run the equipment and how to run their business, I don't think it would be possible for other people to go into their plants and actually execute the same playbook, similar to, like the Toyota manufacturing process. So TSMC has developed this culture, this technology stack. They have, you know, an incredible concentration of PhD level scientists who help them continually optimize what they do. So they're going to go help Intel in this agreement, make their manufacturing a lot better. Of PhD-level scientists who help them continually optimize what they do. So they're going to go help Intel in this agreement, make their manufacturing a lot better. So now, will that mean that ultimately, intel's fabs become TSMC fabs effectively, maybe? And then is Intel just going to focus on competing more effectively with ARM on the CPU front, or competing in the GPU race finally in some meaningful way, maybe?

Speaker 1:

I think there's still a lot of questions to be had, but I do think it's exciting from an American perspective to, whatever the ownership structure is, the revitalization of American semiconductor manufacturing capacity and being at the cutting edge is important because as a country, we're behind at this point, like the manufacturing side. Not on the design side, we're still world leading at the moment, but on the manufacturing side we're clearly behind. The closest in terms of fabs behind TSMC would probably be Samsung, which is in Korea. Global Foundries, which used to be part of AMD, has some advanced manufacturing capability, but not nearly at the level or scale of what TSMC offers. So I think it's a good thing for Intel, because the alternative is to essentially fire sailing their manufacturing capabilities, and I don't think that's good on a lot of levels.

Speaker 2:

You were talking about this idea of specializing in meat, this idea of expertise, and in the podcast I keep mentioning the Acquired podcast they had a phrase on there that I wrote down. It's something you've probably heard before, but the phrase was you can only do one thing well, and that really stood out to me. I started thinking about associations. In what way do you think associations can learn from this joint venture, from the idea of specializing, from the idea of expertise? Do you feel like there's one area that associations do incredibly well that they should kind of funnel all their resources into?

Speaker 1:

Or do you think the current I think let me come back to that just a second. I just want to add one thing to what you said earlier on the acquired pod. For those of you that haven't actually had an opportunity to take in one of their pods, I would highly recommend it. The TSMC episode is great. They have a series on NVIDIA, they have a recent episode on Microsoft and they also have episodes in a variety of other companies like Costco and Hermes and most recently on Rolex. So if you're interested in kind of like a very well told story about the history of a particular business that you might be a fan of or you are a consumer of, I would definitely recommend it.

Speaker 1:

I absorb their pods, you know, in 15 to 30 minute chunks, usually while I'm out and about or in a car. I don't typically listen to them, you know, end to end, because they tend to be very, very long, some of them actually the two and a half hours you mentioned. Malware and TSMC is one of their shorter podcasts, the one I just listened to on Rolex, I think it's like five hours or something. It took me like a month to get through it, partly because it's not a company I'm like, personally super interested in, whereas the Microsoft one, I think, was like four and I listened to it in like two, you know, in two chunks, because that's something I'm super interested in. So, at any event, for listeners, if you want to go deep on business history of a particular company or sector, I highly recommend Ben and David at the Acquired podcast. They do fantastic work, but one of the things they talk about that's similar to the way you described it is a lot of their episodes they talk about hey, what makes the company's product the best in the category? And one way they describe it is what makes the beer taste better, right? So if you're a beer producer, if you're a brewery, only do the things that actually make your beer taste better. So if it doesn't actually make your beer taste better like owning your own fields of barley, that probably doesn't make your beer taste better than sourcing those grains from the best producer that's regional to you.

Speaker 1:

Can you get an advantage by vertically integrating or by owning more of the process? In the case of Rolex, rolex actually manufactures their own steel because they have a specification and a proprietary approach to it that they believe is the best way in the world to do that particular type of steel. Is that necessary? Maybe, maybe not, but in their particular case maybe that works. But vertical integration, as successful as it can be in specific instances, if you study industries more broadly, tends to not be the norm. You tend to see specialization of labor, aka high distribution of supply chains. So if you look at, for example, auto manufacturing or any kind of complex product, you tend to see this whole ecosystem of companies that highly, highly specialize in very specific things. Like in auto manufacturing. You might have a company that just specializes in wiring harnesses, for example. That's all they do, and on and on and on. There's everything, every single thing you break down.

Speaker 1:

So the question I would ask of the association world is does the association need to be a fully integrated delivery mechanism to deliver value through engagement, right? If you think about why people are part of the association, they come for value to be delivered right, and that's true for probably any business. In the case of an association, what's the value? What's learning? Sometimes it's content. A lot of times it's connecting with other people. It might be professional advancement because the association helps them find new career opportunities. Those are some examples of value creation by the association.

Speaker 1:

But the association tends to have like this fully integrated stack of services where they run events and they provide member services and they run technology and do all this other stuff, and they tend to have a radical distribution of different specializations. We often say there's an association for everything, and so the question in my mind is does it make sense for all associations to run their own technology stack? Does it make sense for all associations to manage their own events? Does it make sense for all associations to produce all their own educational content? In many cases it might, and in some cases it might not. Would it make sense for associations to outsource some of those things? Would it make sense for associations maybe to join forces if they're in adjacent verticals with other associations, which, of course, is not a novel concept? That's happened for a long time.

Speaker 1:

But oftentimes you see people kind of repeating the same playbook and everyone's manufacturing their own steel, and so I would ask you do you need to manufacture your own steel? Do you need to have every single thing vertically integrated in your association? And a lot of times that would be like an extreme way to describe it. Associations obviously use a wide variety of vendors and all that, but my question is really just an open one. I don't have an answer for it necessarily.

Speaker 1:

It's just with what's happening with AI. Does your current model for creating value for the end consumer make sense, right? Are there other forms of value that you need to be able to create that you're not seeing because you're so focused on making this deal, because you're so focused on that lower level work of producing the event or creating the courseware or whatever it is? Does that make you somewhat blind to the direction of your sector, where your profession is heading, what people are actually in need of, and are you able to respond to market demand by creating the value that they want three years from now in the world of AI? That's really the question I have, and if you can anticipate those changes and try to be where the curve is heading right in terms of that future creation of value that offers opportunities for players who are able to think that way. I don't know that means that their whole business model needs to change, but it could mean that, and the era of powerful AI that we're in now we're just starting in means that there's more options.

Speaker 2:

Yeah, and actually what you just said at the end there, now that we're in this age, this era of AI, I wonder if it is possible to do more than one thing. Well, I mean, my gut instinct is like focus on your one thing and execute like heck essentially on that, but maybe in the advent of AI, maybe you can kind of juggle all these things. What do you think about that?

Speaker 1:

I mean, I think that it's a great question, right. I look at Blue Cypress, our family of companies and all the different things that we do across there. I think the only way that works, at least in our case, is that we have each business not isolated, but at least somewhat separated and defined, where Sidecar is its own thing and Sidecar has its mission to educate a million people in the association sector on AI by the end of the decade, and that's very clear, and everything Sidecar is doing is focused on delivering on that mission. Similarly, the folks at the Rasaio platform are focused on delivering personalization at scale for all associations, right, and they're not thinking that much about AI education. They know about it, but it's like there's a degree of separation that allows for separation of focus as well. It seems to be working for us pretty well, but we are ultimately, you know, one integrated company, On the other hand, because there's obviously common ownership and you know these are related businesses, as we talk about a lot, but they're separated enough. So that seems to work in our case. I know that in a lot of other cases, you know, you have companies that have multiple different products and divisions and services, and it's really hard to see that work.

Speaker 1:

I think associations do have a singular focus in the sense that they're there to create value in the lives of their members. That's ultimately what they're trying to do. The question is is that is there? Is there something disjointed under the hood? Right? Do you need multiple, different, fundamentally different skills to execute on the future of that vision, and what does that mean in terms of the best way to source those materials? If you think of it as a manufacturing supply chain, saying what's the best place to get my event production or my event design or my planning and I'm picking on that one because it's top of mind since we just ran an event but, like that particular critical process, does that make sense? Or does it make more sense to do that a different way, Right, I think? I think events are a critically important part of the association formula. That's a form of engagement that's likely to be durable for centuries to come, because we're all going to want to get together. That's just a thing that people want to do. So I think associations are well positioned to deliver there.

Speaker 1:

But I guess ultimately it's more of a question than anything else. That I have in my mind is that when you see something like this happen when companies that you grew up with that were the dominant. You know behemoth blue chip. You know companies that people look to and even say, hey, we're going to base our methodology for management, Like the whole OKR systems that people talk about, objective, key result. That's an Andy Grove thing from you know, decades ago, that he implemented when he was running Intel actually before he was CEO and it's still a fantastic methodology, by the way. We use it ourselves and I highly recommend it to anyone who's looking for a better way to frame their priorities and execute on them. It doesn't mean that OKRs are bad All of a sudden. It means that something happened right. This company that was at this unbelievable pinnacle lost its way. Perhaps, or maybe the model you know no longer made sense, so that to me is the big opportunity to ask that question openly.

Speaker 2:

So maybe, as an association leader, consider whether your organization might be the Intel and if there's something like a TSMC out there that's doing some element of your business in ways that you couldn't imagine that you could partner with them.

Speaker 1:

Exactly right. I think it's just being open-minded about it, because Intel identified like deep in its culture as a manufacturing company, as much as they were a semiconductor design firm and still are right. They deeply identified as a manufacturing company and a big, big part of their culture was, you know, the pride in that, and I think, associations all of us, right, we have that the rootings of where we came from. Um, that isn't to say that that's bad, necessarily. It might be the thing that you focus on and you say you know what. That's actually what's going to differentiate us, because we're so good at producing world-class events. We do indeed do it better than anyone else, and by doing that, our beer does taste better, right?

Speaker 1:

So that's what you have to figure out is what is it that people are wanting from you? And I think that's part of the key question I'm asking us to all ask is what will people want from you in two years time? Forget about 10. Two years from now, given what's happening with AI in your sector, forget about what's happening to you and your association, but in your sector, what will people in your sector look for from the association? That's the open question. Figure that out and then go build that and then determine whether or not building that means retooling the way you're structured. It might it might not. It might mean you might be perfectly positioned for it, but I suspect there's a good number of us in this world of associations that might find that does make sense to reconsider existing business models.

Speaker 2:

That's a great place to leave this and move on to topic two. We're going to be talking about the release of Lama4, which happened on April 5th of this year just recently. This release marks a significant step forward in Meta's open source LLM ecosystem, with models designed to push the boundaries of multimodal and multilingual capabilities. So I want to talk about the family of models in Lama 4, and I've got to give them a shout out. The names are much better than what we typically see with these family of model releases. So we've got Lama 4 Scout, a compact model with 17 billion active parameters and a total of 109 billion parameters. It supports an industry-leading 10 million token context window, which we'll talk about in a bit, making it ideal for tasks like multi-document summarization and reasoning over large data sets. We've also got Lama for Maververick also has 17 billion active parameters, but a total of 400 billion parameters. It excels in reasoning, coding tasks and long context processing. And then, aptly named, we've got llama for behemoth, which is still in development, a massive model with 288 billion active parameters and nearly 2 trillion total parameters. It is positioned as one of the most powerful AI models globally, but has not yet been released. Lama4 models are natively multimodal, capable of processing text, images and video inputs while generating outputs and the models use a mixture of experts framework, or MOE, which enhances efficiency by activating only the necessary experts for specific tasks instead of the entire model. This design enables high performance while reducing computational costs.

Speaker 2:

So when I was looking up the release of LAMA4, I watched a few YouTube videos. It seemed on my end like some of the reception to the release was mixed, so I want to cover a few of those items. Due to European Union data privacy regulations, lama 4 cannot be used or distributed within the European Union, which I think is interesting. Also, reports suggest that Lama 4's release was accelerated due to competition from Chinese AI lab DeepSeq's cost-efficient models, which outperformed earlier Lama models, saying that perhaps this release was a bit rushed or panicked. There's also some commentary on the open source side, so the restrictive open weights license has drawn criticism. For example, large-scale commercial users require special approval from Meta. These limitations contrast with more permissive licensing models offered by competitors like Deepsea, and some developers were disappointed by initial performance issues reported through APIs. Additionally, discrepancies between benchmark scores achieved by an experimental version of Lama 4 and the publicly released model sparked criticism of Meta's transparency. So that was at least what.

Speaker 1:

I was seeing Amit. I know that you have.

Speaker 1:

I think that you're excited about this release and you're very impressed by it. So what are your initial thoughts? A couple of things I'd point out to add to what you said. One is that they are natively multimodal. So Lama 4 was trained on a combination of text, image, video, audio etc. And so by fusing together the training content that the models were trained on, you can really see the model reason across modalities. So we talked about this recently in the context of GPT-4.0's new omni-modal image generation and how that image generation is in the context of the whole conversation you have with ChatGPT in that particular conversation. And that's because it's a single model that's doing both text output and image output. Now, at the moment, lama 4, the two released versions or the two released sizes the Scout and Maverick are text output only, but I believe that's a temporary thing because they are truly omnimodal models. My suspicion is that with the Behemoth release, which will probably be only inferenced either on Lama or Azure or other large scale clouds, it's such a large model that they'll have a true omnimodal model in terms of output and that'll be able to compete with the latest from OpenAI on the image front. But what's interesting about it is the level of understanding the model seems to gain from being trained on this. Fusion of text and image and video is significantly better in a lot of other ways that are not related to the question of did it produce images as an output. So that's one thing. Another thing is we've been talking about mixture of experts, or MOE architectures, for quite some time, and the concept is not new, but the technology continues to improve considerably. So DeepSeek actually has done some tremendous work in improving MOE models and they have open sourced all their stuff and all their research is out there, and Lama4 is also an MOE model family.

Speaker 1:

So I want to point out something that you mentioned earlier. For example, with Scout, you're talking about a model that has over 100 billion parameters in total, but has 17 billion active parameters, so a fairly small percentage of the total parameters. So what does that actually mean? It means that for each token that the model is looking at, only 17 billion parameters are active at that moment in time for that token. The next token, it might be different experts within the model. So what's happening there essentially is you have this high degree of specialization within the model that allows for really, really high skills, essentially in different categories, and the model is smart enough to be able to switch around dynamically within a given prompt to use multiple models. What this means is it's more efficient. So it's 109 billion parameter model, which is roughly it's about the same size as Lama 3.370, which is what we talked about in December when it came out. This is a tiny bit larger, but it's actually going to be way more efficient than that and smarter because of the MOE architecture and a number of other things. So that's super, super interesting.

Speaker 1:

There's a lot of noise being made about the 10 million token context window and, I think, appropriately so. There's a lot more testing that needs to be done on these models in the public sphere. Now that these models are out, the question will be like, how good is Lama Force Scout at doing things with that many tokens? So if you say there's this thing called the needle in the haystack problem, which is, you know, can you find something very specific in a very, very large context? And just as a reference point, you know a token being roughly equivalent to a very, very large context and just as a reference point, a token being roughly equivalent to a word, not exactly, but for our purposes that's a close enough approximation. It's about 10 million words, which is a lot, right, that's equivalent to very large code repositories. It's equivalent to, I think, somewhere in the order of magnitude of a couple hundred books. So it's a lot of content. And so what can you do with that? Well, in theory the model has enough breadth of insight to look at a very large corpus of content let's just say an association's entire knowledge base or a very large chunk of it and to be able to inference across that, in theory, what that would allow you to do is to have an even more performant knowledge assistant, even have a more capable database analyst and so forth. So larger context windows should, in theory, give you better intelligence at the aggregate level.

Speaker 1:

The reason I kind of hedged that statement a little bit is because with models like Gemini that have had long context windows for a while, performance has been good in certain cases but doesn't necessarily bump up the overall kind of understanding of the content that the model has had, because there's some limitations in the underlying architecture still that make that valuable. But it's not necessarily, like you know, the silver bullet we've been looking for that says, hey, you can have unlimited content. Also, if you did take 10 million tokens and drop them into Scout. The time it would take and the cost for inferencing would be unsustainable. So if you do use the full context window or anything close to that, you're talking about processing a very large amount of content. It's going to slow down, it's going to be way more expensive. So don't necessarily take that as being, like you know, the ultimate solution to LLM shortcomings. It's not that, but it's still exciting. Nonetheless, it's a valuable tool. So those are some of my initial thoughts. I think that you know what it is. It's exactly what we expected from Meta. Right About a year ago they released Lama 3.

Speaker 1:

Now it's time for Lama 4. Notable is that they do not have a reasoning model. So there is no like equivalent to DeepSeek R1, cloud 3.7, extendedended Thinking Mode or OpenAI's O1, slash, o3 models. And for those that haven't heard us talk about reasoning models, the quick synopsis of that is these are models that are smart enough to realize that they need to break down a problem into chunks, work through those problems that are those complex problems step by step, even check their work before producing an answer. But they do take longer to produce the answer than a model that's just doing a quick shot and answering your question as fast as it can. Lama4's release was accompanied by a little bit of a cryptic message from one of their senior execs saying that reasoning is on the way. That isn't to say that it's going to be part of Behemoth necessarily, but it is something that cannot be lost on the folks at Meta that that's a critical part of a frontier all I'm offering at this point. So I find it exciting. I think you're going to see a DeepSeek R2 very soon, maybe this month, and DeepSeek R2 is probably going to leap past Llama 4 in some ways. Right.

Speaker 1:

And then you know OpenAI and Google. Google released Gemini 2.5 Pro, which is no slouch. It's a very powerful model. It's actually free for consumers to use through their UI. It's not free at the API level, but it's free for consumers. And then, of course, there's the folks at Anthropic that make the cloud model. So this is just a crazy competitive landscape and this is just the latest.

Speaker 1:

You know, the latest bit to keep track of what it should tell associations is the same thing we've been saying for quite some time now is you have choice, you have optionality. You don't have to go down a singular path saying, oh well, I've heard a chat GPT. Therefore, I'm going to use OpenAI for everything. That very well may be the right solution for certain use cases, but there's so much choice. Many association leaders talk to me about privacy, they talk to me about data security and they say, hey, I don't want to take all of my data and drop it into ChatGPT. I agree with that. You should be very cautious about whoever you share your data with, whether they're an AI vendor or a traditional SaaS vendor or anything else for that matter. In the context of OpenAI or any other major lab, do you really want to, you know, have all of your data residing in their ecosystem? Now, their terms of use, their license with you assuming you're a paying customer does indicate that they will not cannot legally use your data for training future models. But that's just an agreement. Does that mean that that agreement will be abided by? And some people believe it, some people do not believe it. So that's up to you to determine.

Speaker 1:

The reason I raise all of that stuff about privacy and data security is that, if you are talking about open source, you have your choice of inference providers. So you can run Lama 4 on Grok with a Q G-R-O-Q. You can run Lama 4 on Azure. You can run Lama 4 on Azure. You can run it on AWS. You can run it on tons of different places. You could even set up your own infrastructure to run Lama 4. It's an open source model. You can run it anywhere, and that's true for all open source models that have open source and open weights. You can run them wherever you'd like.

Speaker 1:

Why is that important? Well, think about everything in terms of economics and incentives. Why would you be concerned about OpenAI, anthropic Google or even Meta itself, if you use their service, or DeepSeek, for that matter having your data, even if the agreement said they're not allowed to use your data for training? Why would you consider that to be a potential risk? Well, potentially, there's an incentive to use your data to make new models and use everyone's data to make new models. That's far larger than the downside risk of using that data, even violating agreements in some cases.

Speaker 1:

If you're the model developer and you are also the inference provider because there's no separation of concerns in theory there could be value in inappropriately using some of that data.

Speaker 1:

That is something that could be argued In comparison if you're an inference-only provider, someone like a Grok or some of the other providers that are out there, even some of the cloud providers that might have relationships with the model developers, but they themselves aren't the model developers. There's a separation of concerns that may cause you to have more comfort. So, for example, the folks at Grok who we've talked about they do not train models, they do not build models. They don't have a horse in the race. They work with Meta, they work with Mistral, they work with a bunch of other model providers. They even have one of OpenAI's actual open models running on their cloud as well. So they don't really have an economic incentive to, you know, misappropriate your data at all. So there's that piece of it as well that I encourage people to be thinking about. So open source means optionality, it means cost reduction, but it also means better security.

Speaker 2:

I was smiling there because, as you were talking about model providers versus inference providers, I was thinking of Intel and TSMC and, like you know, it's expertise and it's all kind of we're talking about the same thing in a circular way. Amit, you said the idea here, which we go back to all the time, is that associations have choice. I know we've mentioned on the pod before that several of the products that fall within the Blue Cypress family of companies use Lama models. I don't know if they still do so what is your kind of personal take or your business take on Lama 4? Is that the direction you want to continue going? Are you impressed with what you've seen or are you a bit kind of skeptical?

Speaker 1:

Well, we haven't tested it in any significant way yet. We've played with Lama 4 already, but we haven't plugged it into Skip or Betty or any other products in the family in any meaningful way. So that'll come shortly. We are completely model agnostic. We're also inference provider agnostic. That's the way we architect everything. We do again to allow our clients, who are these associations, to have optionality. So with any of our products, you can choose different models on whichever inference provider you like, and that's a really powerful thing to know that you have that available.

Speaker 1:

Now. Different models have different capabilities. So if you were to say, hey, I want to use Lama 3.0 instead of 3.3, for some reason, you know, some things in certain products aren't going to work as well or might not work at all right. So, for example, certain features within our data analyst AI, skip, do not work with models that are weaker than GPT-4.0 or with LAMA 3.3. It just wasn't possible to do the current level of capability we have in skip until those models got as smart as they've gotten. So we tend to look at it really in like terms of model class. We'd say, okay, lama 3.3 is roughly equivalent to GPT-4.0. Lama 4 is a step above that. You know, whatever comes like Cloud 3.7 is a step above that, probably as well. So we look at it more in terms of, like, the power level of a model, and so within that, you know, we're constantly swapping models in and out.

Speaker 1:

The way we've designed all of our software architectures is to be completely model agnostic and you can literally plug and play these things in a number of different ways, and sometimes our products use a kind of a medley of models where we're using, oh, we use Lama 3.3 for certain things, but then certain tasks are way more complicated, so we use a reasoning model like an O3 mini or a DeepSeq R1. And then, you know, we tend to run our inference, you know, either within Azure or with Grok, and so we are. You know, we have these separations built into the way we've designed stuff. That I think is really important to have in place, because there's so much change happening right now that you cannot predict which of these providers is necessarily going to be the best. I think the advantage for someone like Meta and models that are either Lama for itself or one of the many fine-tunes that will happen, you know, when an open source model is released very, very quickly, especially for the larger ones, there will be dozens, if not hundreds, of fine-tuned versions that have been, you know, adjusted in different ways to make them better at certain tasks, like coding or whatever the task might be.

Speaker 1:

So but using those mainstream models sometimes is beneficial because there's so much money and so much development going into them that that can be helpful. But largely these things are commodities. Largely it's more about like you know what power level, if you will, are these things at and you can plug and play them. So different models have different advantages, of course, but you know what power level, if you will, are these things at and you can plug and play them. So different models have different advantages, of course, but you know, if you're getting to the point where these things are all really really good, that optionality leads to radically lower costs, which is what we're seeing.

Speaker 2:

The last question I want to ask you, Amit, because I know you've got a lot of knowledge here. There was some criticism on its open weights license. Are these models truly open source still? Can you kind of explain what that means?

Speaker 1:

Yeah, well, I mean, you know Meta is essentially saying that people who compete with them can't use the model without permission. So they're essentially saying is hey, if you're, you know, if you have more than I think it's 750 million monthly active users was the specific language, if I recall correctly. You know, that's certainly not us. You know maybe one day, but that's not us today or anywhere close to that, and that's not any association I know of, and it's probably people like Snapchat, it's people like Amazon, it's people like Microsoft, so it's their competitors at that scale that are barred from having access. And it doesn't actually say they can't use it. It just says Meta has to get permission, which they probably would not.

Speaker 1:

But I also think that those companies probably wouldn't want to use Meta stuff. So I think it is less of a big deal than some people are making it. I don't think it's unreasonable for a company to have some degree of restrictions on their open source. I don't think all open source needs to be a do whatever you want, no matter what kind of license. I think that there's lots and there's lots of different flavors of open source licenses that are out there, some that are actually far more restricted than this and some that are far more permissive. So I think it's totally fine for the association use case.

Speaker 2:

Yeah, I don't think you all have 750 million members, but you might, especially if we reinvent that business model.

Speaker 1:

Everybody. Thank you for tuning in to today's episode. We will see you all next week Ascend Unlocking the Power of AI for Associations at ascendbookorg. It's packed with insights to power your association's journey with AI. And remember, sidecar is here with more resources, from webinars to boot camps, to help you stay ahead in the association world. We'll catch you in the next episode. Until then, keep learning, keep growing and keep disrupting.