We're sunsetting PodQuest on 2025-07-28. Thank you for your support!
Export Podcast Subscriptions
cover of episode Hugging Face and watsonx: Why Open Source Is the Future of AI in Business

Hugging Face and watsonx: Why Open Source Is the Future of AI in Business

2023/10/3
logo of podcast Smart Talks with IBM

Smart Talks with IBM

AI Deep Dive AI Chapters Transcript
People
J
Jeff Boudier
Topics
Jeff Boudier: Hugging Face是领先的开源AI平台,提供大量的预训练模型和工具,方便AI研究人员、数据科学家和开发者使用。开源是AI发展的关键动力,它促进了模型的快速迭代和改进。Hugging Face与IBM watsonx的合作,旨在将Hugging Face的开源资源与IBM的企业级平台相结合,为企业提供更便捷、合规的AI解决方案。构建一个能够完成所有任务的单一万能AI模型是不现实的,针对特定任务训练的多个模型更有效率和实用。企业应该构建和拥有自己的AI模型,而不是依赖于单一的第三方模型。 Malcolm Gladwell: 本期节目探讨了开源AI的重要性及其在企业中的应用,特别关注Hugging Face和IBM watsonx之间的合作。 Tim Harford: 与Jeff Boudier的对话深入探讨了开源AI模型的特性、Hugging Face平台的功能以及与IBM watsonx平台的合作如何帮助企业利用开源AI。讨论涵盖了Transformer模型、开源库、以及如何选择合适的AI模型来解决特定的业务问题。

Deep Dive

Chapters
Jeff Boudier introduces Hugging Face as a leading platform for AI builders, where researchers share their work and data scientists access pre-trained models.

Shownotes Transcript

Translations:
中文

Hello, hello. Welcome to Smart Talks with IBM, a podcast from Pushkin Industries, iHeartRadio, and IBM. I'm Malcolm Gladwell.

This season, we're continuing our conversation with new creators, visionaries who are creatively applying technology in business to drive change, but with a focus on the transformative power of artificial intelligence and what it means to leverage AI as a game-changing multiplier for your business.

Our guest today is Jeff Boudier, Head of Product and Growth at Hugging Face, the leading open source and open science artificial intelligence platform. An engineer by background, he has a self-professed obsession with the business of technology.

Recently, IBM and Hugging Face announced a collaboration bringing together Hugging Face's repositories of open source AI models with IBM's Watson X platform. It's a move that gives businesses even more access to AI while staying true to IBM's longstanding philosophy of supporting open source technology.

With open source, businesses can build better AI models that suit their specific needs, using their own proprietary data while browsing a ready catalog of pre-trained models. In today's episode, you'll hear why open source is so crucial to the advancement of AI, how IBM's Watson X interacts with open source AI, and Jeff's thoughts on why the singular, omnipotent AI model is a myth.

Jeff spoke with Tim Harford, host of the Pushkin podcast Cautionary Tales. A longtime columnist at the Financial Times, where he writes the undercover economist, Tim is also a BBC broadcaster with his show More or Less. Okay, let's get to the interview. Hi, I'm Jeff Boudier and I'm a product director at Hugging Face. So I'm immediately intrigued. Hugging Face. Is this a reference to the Alien movie or something else?

It is not. And it may be not obvious to a listener, but Hugging Face is the name of that cute emoji. You know, the one that's smiling with his two hands extended like that to give you a big hug. That's Hugging Face. So basically, we named the company after an emoji.

And it is, I saw your website and it is a very friendly emoji. So that's nice. So tell us a little bit about Hugging Face and about what you do there. Of course, Hugging Face is the leading open platform for AI builders.

And it's the place that all of the AI researchers use to share their work, their new AI models and collaborate around them. It's the place where the data scientists go and find those pre-trained models and access them and use them and work with them.

And increasingly, it's the place where developers are coming to turn all of these AI models and data sets into their own applications, their own features. So it's like the Facebook group or the Reddit or the Twitter for people who are interested in particularly generative language AI or all kinds of artificial intelligence?

All kinds of AI really. And of course, generative AI is this new wave that has caught the world by storm. But on HagenFace, you can find any kind of model. The new sort of Transformers models to do anything from translation or if you wanted to transcribe what I'm saying into text,

then you would use a transformer model. If you wanted to then take that text and make a summary, that would be another transformer model. If you wanted to create a nice little thumbnail for this podcast by typing a sentence, that would be another type of model.

So all of these models you can find, there's actually 300,000 that are free and publicly accessible. You can find them on our website at HagenFace.co and use them using our open source libraries.

And so this is fascinating. So there are 300,000 models. Now, when you say model, I'm thinking in my head, oh, it's kind of like a computer program. There are 300,000 computer programs. Is that roughly right or not really? It's a general idea. A model is a giant set of numbers.

that are working together to sift through some input that you're going to give it. So think of it as a big black box filled with numbers. And you give it as an input maybe some text, maybe a prompt. So you're giving an instruction to the model. Or maybe you give it an image as an input.

And then it will sift through that information, thanks to all of these numbers, which we call in the field parameters. And it will produce an output. So when I told you, hey, we can transcribe this conversation into text, the input would have been the conversation in an audio file. And then the output would have been the text of the transcription.

If you want to create a thumbnail for this podcast episode, then the input would be what we call the prompt, which is really a text description. Like a French man in San Francisco talking about machine learning. And the output would be completely original image.

So that's how I think about what an AI model is. And I think what we're starting to realize is that this is becoming the new way of building technology in the world. It has been for the field of dealing, understanding, generating text for quite some time. But now it's sort of moving across every field of technology. We have models for...

to create images, as I say, but also to generate new proteins, to make predictions on numerical data. So every kind of field of machine learning is now using this new type of models. But what's interesting is that if you're, say, a product manager at a tech company and you say, hey, I want to build a feature that does this,

A few years ago, the approach would have been to ask a software developer to write a thousand lines of code in order to build a prototype. And the new way of doing things today is to go look for an off-the-shelf pre-trained model

that does a pretty good job at solving exactly that problem, so you can create a prototype of that feature fast. So it's a new approach of building tech. I'm not a programmer, but I am aware that there was this idea of open source code, and now we have open source models. So what does it mean for something to be open source?

Open source AI actually means a lot of different specific things. It's the open source implementation of the model. So if you use the Hugging Face Transformers library,

to use a model, you're using an open source code library to use that model. Just to interrupt on the transformers, these are these ways of turning a picture of a dog into a text output that says, "Hey, this is a picture of a dog," or, "This is a French text and the transformer is helping you turn it into English text," or it's doing all of these things that you've been describing. The transformer is the engine at the heart of that.

Yes, exactly. And we call them transformers because they correspond to this new way of building machine learning models that was introduced by Google, actually, with a very important paper called Attention is All You Need. And that was published in 2017 by researchers out of Google DeepMind. Wow. That's just six years. That's so new.

It is very new and ever since the pace of innovation of like new model architectures has really, really accelerated. But it really started from this inflection point that came from this paper and its implementation in what is now called transformer models.

the transformer that has conquered every area of machine learning since. Okay, so sorry to interrupt. So you've got this library of transformer models and they're open source and that means what? Anyone can use them for free or that anybody can implement them for free? What does it mean?

So again, there's lots that go into it, but the most important thing is for the model itself to be available so that a data scientist or an engineer can download them and use them. And also, there are a lot of considerations about quality.

how you make them accessible. And a very important one is whether or not you give access to the training data, all the information that went into training that model and teaching it to do what it's trained to do. So I might have fed millions of words into a language transformer, or I might have fed

millions of photographs into a picture transformer. Yes, and now it's trillions. And the accessibility of that training data is very, very important. What's the relationship between the Hugging Face libraries and GitHub? Which, if I understand GitHub correctly, it's this, the repository of open source libraries.

Lots and lots of lines of code and routines and programs that are shared and updated and tracked and they're all available on GitHub, which sounds similar to what you're doing with Hugging Face for AI. So what is the interaction or the relationship there?

Yeah, I think you nailed it on the head. So Hugging Phase is to AI what GitHub is to code, right? It's this central platform where AI builders can go find and collaborate around AI artifacts, which are models and data sets. So it's quite different than software. But we play this central role in the community to share and collaborate and collaborate.

and access all of those artifacts for AI, like GitHub offers for code. And that community must be incredibly important. I mean, the open source is nothing if you don't have a community of people working on it. So how have you been able to foster and nurture that community?

Well, I think it goes to the origins of the transformer model and Hugging Face roll into that. So when the first sort of open model came out, it was called BERT, and it came out of Google. The only way you could access it was to use a tool called TensorFlow. But it happened that most of the AI community was using a different tool.

called PyTorch. And something that Hugging Face did is to make that new model, BERT, accessible to all PyTorch users. And they did it in open source. It was a project called BERT's Pre-trained PyTorch, or BERT PyTorch Pre-trained.

So this is like being able to play my Zelda game on an Xbox or a PlayStation, right? Or am I not really understanding what's going on? No, that's exactly what it is. And the thing is, everybody was using the Game Boy. And so it became very popular. And from there, the community sort of gathered to make all the other models that were then published by...

by AI researchers available through that library, which was quickly renamed from BERT, pre-trained PyTorch into Transformers to welcome all of these different new models. And today, that open source library, Transformers, is what all AI builders are using when they want to access those models, see how they work, and build upon them.

What's striking about this field is that it's changing so fast, it's improving so quickly. How do open source models keep up with that? How do they get iterated and improved? Actually, it's not so much that open source is keeping up with it. It's actually open source that is driving this pace of change. That's because with open source and open research,

data scientists, researchers can build upon each other's work. They can reproduce each other's work. They can access each other's work using our open source libraries, et cetera. So in a sense, it's not really that open source AI is a new idea. It's rather the opposite. There's been a blip of time in which closed source AI seems to be the dominant way, but it's really a blip.

In fact, none of the incredible advances that we marvel about today would be possible without open source. We're standing upon the shoulders of 50 years of research and open source software. So I think that that's really important. If it wasn't for that, we'd probably be 50 years away from having these amazing experiences like ChatGPT or stable diffusion, etc.,

So it's really open source that is fueling this pace of change, all these new models, all these new capabilities.

To give you an example, Meta released LAMA large language model just a few months ago. And ever since, there's been this Cambrian explosion of variations and improvements upon the original models. And today, there are over thousands of them that we host and track and evaluate. So yeah, open source is really the gas and the engine for that.

Jeff just made it clear that it is open source, not closed, that sets the pace for AI innovation. If that's true, then forward-thinking businesses shouldn't shy from leveraging open source AI to solve their own proprietary challenges. But how? Businesses can face serious obstacles when trying to adopt open source technologies, like complying with government regulation or making sure their customers' data stays protected.

In the next part of their conversation, Jeff and Tim discuss how IBM's collaboration with Hugging Face empowers businesses to tap into the open source AI community and how the WatsonX platform can enable them to customize those AI models to their needs. Jeff wants to ask about the partnership between Hugging Face and IBM. How did that come about? Well, it came through a conversation.

a conversation between our CEO, Clément Delangue, and Bill Higgins at IBM, who's really, really close to all the amazing research work and open source work that's happening at IBM.

And that conversation sort of sparked the evidence that we needed to do something together. We share a lot of values in terms of the importance of open source, which is fundamental to us with the importance of doing things in an ethics-first way to enable the community to incorporate ethical considerations in how they're building AI.

And we sort of have a different audience to start with, which is all the AI builders use Hugging Phase today to access all the models we talked about, to use them using our open source model.

and build with them. And IBM has this incredible history of working with enterprise companies and enabling them to make use of that technology in a way that's compliant with everything that an enterprise requires. And so being able to marry these two things together is an amazing opportunity. And now we can enable the largest corporations that have

sort of complex requirements in order to deploy machine learning systems and give them an easy experience to take advantage of all the latest and greatest that AI has to offer through our platform. Let's talk about this idea of a single model or a variety of models, because what I've been hearing you say, you've been saying, oh, there are lots of models. There are hundreds of thousands of models available.

available on Hugging Face, but you've also said there's a single thing, the transformer, and they're all transformers. So if they're all basically the same thing, why can't you just build one super clever model that can do everything? That's a really interesting idea and very much a new idea.

The reason we have over a million repositories, 300,000 free and accessible models on a Hugging Face platform is that models are typically trained to do one thing. And they're typically trained to do one thing with specific types of data. And what became new and evident in the research that came out over the last couple of years is that if you train a big enough model,

With enough data, then those models start to have sort of general capabilities. You can ask them to do different things. You can even train them to respond to instructions. So with the same model, you can say, hey, summarize this paragraph, translate this into English, start a conversation in French and then pivot to German. And so these are general sort of language capabilities.

And I think when ChatGPT came online and the world sort of discovered these new capabilities, there was at least for a short period, this sort of idea, this sort of myth that the end game of all this is maybe one or a handful of models that are so much better than anything else that exists, that they can do anything that we can ask them to do.

And that's the only model that we will need. And I, for one, I think it is a myth. I don't think it is practical for a variety of reasons. Say you're writing an email and you have like this great suggestion of text to sort of complete your sentence.

Well, that's AI. That's a large language model. That's a transformer model that does that. So there are a ton of existing use cases like this. And these use cases are powered by specific models that have been trained to do one thing well and to do it fast. If you wanted to apply these sort of all-knowing, powerful Oracle type of models,

You would not be able to serve millions of customers through a search engine. You would not be able to complete people's sentences because the amount of money that you would need, the number of computers that you would need to run such a service, it just exceeds what is available on the planet. So one

A reason for which it's not a practical scenario is that it's just very expensive to run those very, very large models. What I'm hearing is it's like, look, if you want to screw in a screw, you need a screwdriver. You don't want an entire tool shed full of tools if the task is to screw in a screwdriver. And sure, you could bring the tool shed.

There are all the tools, there's a screwdriver there, but it's not necessary. It's incredibly expensive. It's incredibly cumbersome. And that cost exists, even though maybe as the user who's just typing into a prompt box, the user may not see it, but it's still very real.

That's right. And then another one is performance. So taking the screwdriver example. And by the way, we're not quite there at this moment where we have this all-knowing, powerful Oracle that is still sort of a sci-fi scenario. But we have screwdrivers, but we also have the Leatherman, the multi-tool model.

a Swiss army knife. And that's sort of the moment that we are in today. But now, if I'm trying to open up my computer, turns out that it requires a specific kind of screw, like these tiny little Torx screws. And

Having a torque screwdriver will get me much further than trying to use my Leatherman where maybe I'll get the knife blade and it will mess up the screw and maybe eventually I'll get to what I need. But my point is that if you take a very specifically trained machiner,

model for a particular problem, it will work much better. It will give you better results than a very, very generalistic big model that can do a lot of things. And so for things like search engines or things like translation, for things that are very specific, companies are much better off using smaller, more efficient models that produce better results. That's really interesting. And presumably then,

Being able to know which model to use or being able to know who to ask which model to use becomes a very important capability. Yes, and that's what we're trying to make easy through our platform. So tell me about how this works with IBM's Watson X platform. How do you see Hugging Face's customers benefiting from that?

The end goal is to make it really easy for Watson X customers to make use of all the great models and libraries that we talked about. All the 300,000 models are today on Hugging Face.

And to do this, we need to really collaborate deeply with the IBM teams that build the Watson X platform so that our libraries, our open source, our models are well integrated into the platform.

If you are a single user, if you are a data science student and you want to use a model, we make it super easy, right? We have our open source library. You can download the model on your computer and run with it then. But in enterprises, there is a vast complexity of infrastructure and rules around what people can do and how the data can be accessed.

And all this complexity is sort of solved by the Watson X platform. This season of the Smart Talks podcast features what we're calling new creators. Do you see yourself as being a creative person?

I think it's a requirement for the job. I mean, we're in such a new and rapidly evolving industry that we have to be creative in order to invent the business models, the use cases, etc.

of tomorrow. My role within the company is really to create the business around all of the great work of our science and open source and product team. And by and large, the business model of AI within the whole ecosystem is still something that companies are trying to figure out.

So creativity is really important to really have the conversation with companies, understand what they're trying to do and then build the right kind of solution. So that's like where creativity comes into play. And one of the things that you've been talking about is just this growing number of models, this growing number of capabilities, this growing number of use cases, enormously exciting. But also I think,

completely bewildering for most people who are trying to navigate their way through this maze of possibilities that is growing faster than they can even learn about it. So how are you helping people navigate and make choices in that environment? And how does the partnership with IBM help with that? Well, as I said, our vision is that AI machine learning is becoming the default way of creating technology.

And that means every product app service that you're going to be using is going to be using AI to do whatever it is better, faster. And I guess there are two competing visions of the world coming from that. There is this vision of the Oracle all-powerful model that can do everything.

And our vision is different. Our vision is that every single company will be able to create their own models that they own, that they can use, that they control.

And that's the vision that we're trying to bring to life through our open source tools that make this work easy through our platform where you can find all those pre-trained models that are shared by the community. So we really want to empower companies to build their own stuff, not to outsource all the intelligence to a third party.

And the WatsonX platform from IBM gives those tools to enterprise companies so that you can use the open source models that Hugging Face offers. Then you can improve them with your own data without sharing that data to a third party.

And then you could do all of this work in compliance with whatever governance requirements you have for your company. Maybe your finance services company and you have a specific set of rules. Maybe your healthcare company and you have very strong privacy requirements for patients' data. Maybe your tech company and your...

You have your customers, your users' personal information. So you need to be able to do this work respecting all of that. Jeff Boudier, thank you very much. Thanks so much, Tim. It was fun. To create the AI models of the future, we're going to need open source. That means there's a place for business in the open source community to harness the game-changing potential of AI innovation.

Like Jeff said, businesses face unique challenges they need to solve at scale. Without proper support systems, tapping into open source AI at enterprise level is daunting. Finding the right size model for the job, fine-tuning its purpose, all while addressing governance requirements around data, privacy, and ethics.

So for businesses, IBM's collaboration with Hugging Face is a mark of progress because it signifies that business can tap into open source AI while preserving enterprise level integrity. Businesses should embrace the open source community and the AI future, much like Hugging Face and its emoji namesake suggests. I'm Malcolm Gladwell. This is a paid advertisement from IBM.

Smart Talks with IBM is produced by Matt Romano, David Jha, Nisha Venkat, and Royston Besserve with Jacob Goldstein. We're edited by Lydia Jean Cott. Our engineers are Jason Gambrell, Sarah Bruguere, and Ben Talladay. Theme song by Gramascope. Special thanks to Carly Migliore, Andy Kelly, Kathy Callahan, and the 8Bar and IBM teams, as well as the Pushkin Marketing team.

Smart Talks with IBM is a production of Pushkin Industries and Ruby Studio at iHeart Media. To find more Pushkin podcasts, listen on the iHeart Radio app, Apple Podcasts, or wherever you listen to podcasts.