Listen "Emad mostaque on logan bartlett show"
Episode Synopsis
https://www.listennotes.com/podcasts/the-logan-bartlett/ep-46-stability-ai-ceo-emad-8PQIYcR3r2i/[00:00:00.000 --> 00:00:02.580] (upbeat music)[00:00:02.580 --> 00:00:07.560] - Welcome to the 46th episode of Cartoon Avatars.[00:00:07.560 --> 00:00:09.840] I am your host Logan Bartlett.[00:00:09.840 --> 00:00:11.960] Welcome back for break.[00:00:11.960 --> 00:00:13.280] Thanks everyone for bearing with us[00:00:13.280 --> 00:00:16.160] as we took a pause over the last couple of weeks.[00:00:16.160 --> 00:00:18.200] We're excited for this episode.[00:00:18.200 --> 00:00:19.880] This, what you're gonna hear on this episode[00:00:19.880 --> 00:00:22.480] is a conversation that I had with the Mod in the Stock.[00:00:22.480 --> 00:00:27.400] And Mod is the founder and CEO of Stable Stability AI,[00:00:27.400 --> 00:00:29.840] which is the largest contributor to Stable Diffusion.[00:00:29.840 --> 00:00:32.280] Stable Diffusion is the fastest growing[00:00:32.280 --> 00:00:34.120] open source project of all time.[00:00:34.120 --> 00:00:39.080] It's one of the leading platforms in generative AI.[00:00:39.080 --> 00:00:41.720] And Ema and I had a really interesting conversation[00:00:41.720 --> 00:00:43.080] about a bunch of different things,[00:00:43.080 --> 00:00:47.720] but we dive into the state of artificial intelligence today,[00:00:47.720 --> 00:00:50.920] why this is possible, when it wasn't in the past,[00:00:50.920 --> 00:00:53.160] where this is going in the future,[00:00:53.160 --> 00:00:56.760] how he differentiates versus competitors like OpenAI.[00:00:56.760 --> 00:00:59.560] Really fun conversation and appreciate him[00:00:59.560 --> 00:01:01.120] for powering through.[00:01:01.120 --> 00:01:02.760] He was a little sick as we were doing this.[00:01:02.760 --> 00:01:07.760] So it was a fun conversation and I appreciate him doing it with me.[00:01:07.760 --> 00:01:09.840] And so before you hear that,[00:01:09.840 --> 00:01:11.560] we talked a little bit about this before break,[00:01:11.560 --> 00:01:13.240] but we are gonna make a more concerted effort[00:01:13.240 --> 00:01:16.440] to get people to like and subscribe[00:01:16.440 --> 00:01:20.240] and share and review the podcast itself.[00:01:20.240 --> 00:01:22.560] And so if you're whatever platform you're listening on,[00:01:22.560 --> 00:01:25.720] if it's YouTube, if it's Spotify, if it's Apple,[00:01:25.720 --> 00:01:29.040] whatever it is, if people could go ahead and like[00:01:29.040 --> 00:01:31.640] and subscribe and leave a review,[00:01:31.640 --> 00:01:33.880] share with a friend, all of that stuff.[00:01:33.880 --> 00:01:35.480] We're trying to figure out exactly what direction[00:01:35.480 --> 00:01:36.640] to take this in.[00:01:36.640 --> 00:01:40.080] And so that validation and feedback[00:01:40.080 --> 00:01:42.520] and also the growth that comes along with all that stuff[00:01:42.520 --> 00:01:44.880] is super appreciated.[00:01:44.880 --> 00:01:46.880] It's not something we had been comfortable[00:01:46.880 --> 00:01:48.240] asking for to date,[00:01:48.240 --> 00:01:51.080] but as we kind of figure out what direction we're gonna go,[00:01:51.080 --> 00:01:54.840] we'd love to see more shares, more reviews, more views,[00:01:54.840 --> 00:01:56.200] more likes, all that stuff.[00:01:56.200 --> 00:02:00.520] So really appreciate everyone's support in doing that.[00:02:00.520 --> 00:02:02.520] And so without further delay,[00:02:02.520 --> 00:02:04.280] what you're gonna hear now is the conversation with me[00:02:04.280 --> 00:02:06.880] and I'm on Mistock from Stability AI.[00:02:06.880 --> 00:02:09.920] All right, Iman Mistock.[00:02:09.920 --> 00:02:10.800] Did I say that right?[00:02:10.800 --> 00:02:12.120] - Yep. - Perfect.[00:02:12.120 --> 00:02:13.680] Thank you for doing this.[00:02:13.680 --> 00:02:16.560] Founder of Stability AI,[00:02:16.560 --> 00:02:20.480] one of the main contributors to stable diffusion.[00:02:20.480 --> 00:02:23.480] Thank you for coming on here today.[00:02:23.480 --> 00:02:24.440] - So pleasure, Logan.[00:02:24.440 --> 00:02:25.960] Most I have be here.[00:02:25.960 --> 00:02:26.800] - Yeah, totally.[00:02:26.800 --> 00:02:30.120] So maybe at a highest level, we can start off with[00:02:30.120 --> 00:02:32.560] what is generative AI?[00:02:32.560 --> 00:02:34.760] How would you define that for the average person?[00:02:34.760 --> 00:02:37.880] - So I think everyone said of kind of the concepts[00:02:37.880 --> 00:02:40.560] of big data 'cause the whole of the internet previously[00:02:40.560 --> 00:02:41.680] was on big data.[00:02:41.680 --> 00:02:44.640] Large, large models built by Google and Facebook[00:02:44.640 --> 00:02:47.000] and others to basically target you ads[00:02:47.000 --> 00:02:49.200] ads were the main part of that.[00:02:49.200 --> 00:02:51.640] And these models extended.[00:02:51.640 --> 00:02:54.480] So how to generalize model of what a person was like[00:02:54.480 --> 00:02:56.600] and then your specific interests,[00:02:56.600 --> 00:03:00.480] like EMAD likes green hoodies or Logan lights black jumpers.[00:03:00.480 --> 00:03:05.520] That then extended the previous to what the next thing was.[00:03:05.520 --> 00:03:08.080] They're like extension models inferring what was there.[00:03:08.080 --> 00:03:09.560] Gerative models are a bit different[00:03:09.560 --> 00:03:12.160] in that they learn principles from structured[00:03:12.160 --> 00:03:16.080] and unstructured data and then they can generate new things[00:03:16.080 --> 00:03:17.400] based on those principles.[00:03:17.400 --> 00:03:21.680] So you could ask it to write an essay about bubble sort[00:03:21.680 --> 00:03:24.840] or a solid about Shakespeare or the TTH[00:03:24.840 --> 00:03:26.840] which is digital so you can do that.[00:03:26.840 --> 00:03:29.960] Or in the case of some of the work that we're most famous for[00:03:29.960 --> 00:03:32.800] you enter in a labradoodle with a hat[00:03:32.800 --> 00:03:34.960] and a stained glass window and it understands that[00:03:34.960 --> 00:03:37.200] and then creates that in a few seconds.[00:03:37.200 --> 00:03:39.360] So I'd say that's probably the biggest difference[00:03:39.360 --> 00:03:40.880] between this new type of generative AI[00:03:40.880 --> 00:03:43.040] and then that old type of AI.[00:03:43.040 --> 00:03:44.880] So the way that I also say that we've moved[00:03:44.880 --> 00:03:47.720] from a big data area to more a big model era[00:03:47.720 --> 00:03:50.600] 'cause these models are very difficult to create, train[00:03:50.600 --> 00:03:53.480] which is why only a few companies such as ours do it.[00:03:53.480 --> 00:...
More episodes of the podcast swyx backup pod
yitay techinasia interview
03/07/2024
RWKV test epsiode
29/08/2023
Twitter space on the Agent Protocol
25/07/2023
code interpreter spaces
09/07/2023
The Rise of the AI Engineer (Twitter Space)
05/07/2023
Jensen Huang - NTU commencement speech
28/05/2023
ermergency pod - google memo
04/05/2023
latent space - replit
02/05/2023
Latent Space - Sharif
26/04/2023
ZARZA We are Zarza, the prestigious firm behind major projects in information technology.