{"id":118266,"date":"2026-03-12T21:15:48","date_gmt":"2026-03-12T21:15:48","guid":{"rendered":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/"},"modified":"2026-03-12T21:17:09","modified_gmt":"2026-03-12T21:17:09","slug":"sharon-zhou-on-post-training-oreilly","status":"publish","type":"post","link":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/","title":{"rendered":"Sharon Zhou on Post-Training \u2013 O\u2019Reilly"},"content":{"rendered":"<p> <a href=\"https:\/\/go.fiverr.com\/visit\/?bta=1052423&nci=17043\" Target=\"_Top\"><img loading=\"lazy\" decoding=\"async\" border=\"0\" src=\"https:\/\/mailinvest.blog\/wp-content\/themes\/breek\/assets\/images\/transparent.gif\" data-lazy=\"true\" data-src=\"https:\/\/fiverr.ck-cdn.com\/tn\/serve\/?cid=40081059\"  width=\"601\" height=\"201\"><\/a>\n<br \/><img decoding=\"async\" src=\"https:\/\/mailinvest.blog\/wp-content\/themes\/breek\/assets\/images\/transparent.gif\" data-lazy=\"true\" data-src=\"https:\/\/www.oreilly.com\/radar\/wp-content\/uploads\/sites\/3\/2024\/01\/Podcast_Cover_GenAI_in_the_Real_World-scaled.png\" \/><\/p>\n<div id=\"postContent-content\">\n<figure class=\"wp-block-video\"><video controls=\"\" src=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4\"\/><\/figure>\n<p>Publish-training will get your mannequin to behave the way in which you need it to. As AMD VP of AI Sharon Zhou explains to Ben on this episode, the frontier labs are satisfied, however the common developer continues to be determining how post-training works below the hood and why they need to care. Of their centered dialogue, Sharon and Ben get into the method and trade-offs, strategies like supervised fine-tuning, reinforcement studying, in-context studying, and RAG, and why we nonetheless want post-training within the age of brokers. (It\u2019s methods to get the agent to truly work.) Test it out.<\/p>\n<p>Concerning the <em>Generative AI within the Actual World<\/em> podcast: In 2023, ChatGPT put AI on everybody\u2019s agenda. In 2026, the problem can be turning these agendas into actuality. In Generative AI within the Actual World, Ben Lorica interviews leaders who&#8217;re constructing with AI. Study from their expertise to assist put AI to work in your enterprise.<\/p>\n<p>Take a look at different episodes of this podcast <a href=\"https:\/\/learning.oreilly.com\/playlists\/42123a72-1108-40f1-91c0-adbfb9f4983b\/?_gl=1*pra1u5*_gcl_au*Mzc5ODUxNDEzLjE3NzI3NDUyNzk.*_ga*NjI3OTAzNjIzLjE3NzI0NzYxMzg.*_ga_092EL089CH*czE3NzMwODg2NjgkbzI3JGcwJHQxNzczMDg4NjY4JGo2MCRsMCRoMA..\" target=\"_blank\" rel=\"noreferrer noopener\">on the O\u2019Reilly learning platform<\/a> or observe us on <a href=\"https:\/\/www.youtube.com\/playlist?list=PL055Epbe6d5YcJUhZbsVW9dlMueIuOxK_\" target=\"_blank\" rel=\"noreferrer noopener\">YouTube<\/a>, <a href=\"https:\/\/open.spotify.com\/show\/5C9oof8TFkP65lDUcEy5jT\" target=\"_blank\" rel=\"noreferrer noopener\">Spotify<\/a>, <a href=\"https:\/\/podcasts.apple.com\/us\/podcast\/generative-ai-in-the-real-world\/id1835476293\" target=\"_blank\" rel=\"noreferrer noopener\">Apple<\/a>, or wherever you get your podcasts.<\/p>\n<p><em>This transcript was created with the assistance of AI and has been calmly edited for readability.<\/em><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=0\" target=\"_blank\" rel=\"noreferrer noopener\">00.00<\/a><br \/><strong>In the present day we have now a VP of AI at AMD and previous pal, Sharon Zhou. And we\u2019re going to speak about post-training primarily. However clearly we\u2019ll cowl different subjects of curiosity in AI. So Sharon, welcome to the podcast.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=17\" target=\"_blank\" rel=\"noreferrer noopener\">00.17<\/a><br \/>Thanks a lot for having me, Ben.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=19\" target=\"_blank\" rel=\"noreferrer noopener\">00.19<\/a><br \/><strong>All proper. So post-training.\u00a0.\u00a0. For our listeners, let\u2019s begin on the very fundamentals right here. Give us your one- to four-sentence definition of what post-training is even at a excessive degree?\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=35\" target=\"_blank\" rel=\"noreferrer noopener\">00.35<\/a><br \/>Yeah, at a excessive degree, post-training is a kind of coaching of a language mannequin that will get it to behave in the way in which that you really want it to. For instance, getting the mannequin to speak, just like the chat in ChatGPT was carried out by post-training.<\/p>\n<p>So principally educating the mannequin to not simply have an enormous quantity of data however truly have the ability to have a dialogue with you, for it to make use of instruments, hit APIs, use reasoning and assume by way of issues step-by-step earlier than giving a solution\u2014a extra correct reply, hopefully. So post-training actually makes the fashions usable. And never only a piece of uncooked intelligence, however extra, I&#8217;d say, usable intelligence and sensible intelligence.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=74\" target=\"_blank\" rel=\"noreferrer noopener\">01.14<\/a><br \/><strong>So we\u2019re two or three years into this generative AI period. Do you assume at this level, Sharon, you continue to have to persuade those that they need to do post-training, or that\u2019s carried out; they\u2019re already satisfied?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=91\" target=\"_blank\" rel=\"noreferrer noopener\">01.31<\/a><br \/>Oh, they\u2019re already satisfied as a result of I feel the largest shift in generative AI was brought on by post-training ChatGPT. The rationale why ChatGPT was wonderful was truly not due to pretraining or getting all that data into ChatGPT. It was about making it usable in order that you would truly chat with it, proper?<\/p>\n<p>So the frontier labs are doing a ton of post-training. Now, by way of convincing, I&#8217;d say that for the frontier labs, the brand new labs, they don\u2019t want any convincing for post-training. However I feel for the common developer, there may be, you realize, one thing to consider on post-training. There are trade-offs, proper? So I feel it\u2019s actually essential to be taught concerning the course of as a result of then you possibly can truly perceive the place the longer term goes with these frontier fashions.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=135\" target=\"_blank\" rel=\"noreferrer noopener\">02.15<\/a><br \/>However I feel there&#8217;s a query of how a lot it is best to do by yourself, versus, us[ing] the present instruments which are on the market.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=143\" target=\"_blank\" rel=\"noreferrer noopener\">02.23<\/a><br \/><strong>So by convincing, I imply not the frontier labs and even the tech-forward firms however your mother and pop.\u00a0.\u00a0. Not mother and pop.\u00a0.\u00a0. I suppose your common enterprise, proper?<\/strong><\/p>\n<p><strong>At this level, I\u2019m assuming they already know that the fashions are nice, however they might not be fairly usable off the shelf for his or her very particular enterprise utility or workflow. So is that basically what\u2019s driving the curiosity proper now\u2014that persons are truly making an attempt to make use of these fashions off the shelf, they usually can\u2019t make them work off the shelf?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=184\" target=\"_blank\" rel=\"noreferrer noopener\">03.04<\/a><br \/>Nicely, I hoped to have the ability to discuss my neighborhood pizza retailer post-training. However I feel, truly, on your common enterprise, my advice is much less so making an attempt to do plenty of the post-training by yourself\u2014as a result of there\u2019s plenty of infrastructure work to do at scale to run on a ton of GPUs, for instance, in a really secure means, and to have the ability to iterate very successfully.<\/p>\n<p>I feel it\u2019s essential to find out about this course of, nevertheless, as a result of I feel there are plenty of methods to affect post-training in order that your finish goal can occur in these frontier fashions or inside an open mannequin, for instance, to work with individuals who have that infrastructure arrange. So some examples may embrace: You possibly can design your individual RL atmosphere, and what that&#8217;s is a bit of sandbox atmosphere for the mannequin to go be taught a brand new kind of talent\u2014for instance, studying to code. That is how the mannequin learns to code or learns math, for instance. And it\u2019s a bit of atmosphere that you just\u2019re capable of arrange and design. And you then can provide that to the totally different mannequin suppliers or, for instance, APIs will help you with post-training these fashions. And I feel that\u2019s actually helpful as a result of that will get the capabilities into the mannequin that you really want, that you just care about on the finish of the day.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=259\" target=\"_blank\" rel=\"noreferrer noopener\">04.19<\/a><br \/><strong>So a couple of years in the past, there was this normal pleasure about supervised fine-tuning. After which all of the sudden there have been all these companies that made it useless easy. All you needed to do is give you labeled examples. Granted, that that may get tedious, proper? However when you do this, you add your labeled examples, exit to lunch, come again, you might have an endpoint that\u2019s fine-trained, fine-tuned. So what occurred to that? Is that one thing that folks ended up persevering with down that path, or are they abandoning it, or are they nonetheless utilizing it however with different issues?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=300\" target=\"_blank\" rel=\"noreferrer noopener\">05.00<\/a><br \/>Yeah. So I feel it\u2019s a bit cut up. Some folks have discovered that doing in-context studying\u2014basically placing plenty of data into the immediate context, into the immediate examples, into the immediate\u2014has been pretty efficient for his or her use case. And others have discovered that that\u2019s not sufficient, and that truly, doing supervised fine-tuning on the mannequin can get you higher outcomes, and you are able to do so on a smaller mannequin that you would be able to make non-public and make very low latency. And likewise like successfully free if in case you have it by yourself {hardware}, proper?<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=330\" target=\"_blank\" rel=\"noreferrer noopener\">05.30<\/a><br \/>So I feel these are type of the trade-offs that persons are pondering by way of. It\u2019s clearly very a lot simpler basically to do in-context studying. And it may truly be less expensive when you\u2019re solely hitting that API a couple of occasions. Your context is kind of small.<\/p>\n<p>And the host and fashions like, for instance, like <a href=\"https:\/\/www.anthropic.com\/claude\/haiku\" target=\"_blank\" rel=\"noreferrer noopener\">Haiku<\/a>, a really small mannequin, are fairly low-cost and low latency already. So I feel there\u2019s principally that trade-off. And with all of machine studying, with all of AI, that is one thing that you need to take a look at empirically.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=363\" target=\"_blank\" rel=\"noreferrer noopener\">06.03<\/a><br \/>So I&#8217;d say the largest factor is persons are testing these items empirically, the variations between them and people trade-offs. And I\u2019ve seen a little bit of a cut up, and I actually assume it comes all the way down to experience. So the extra you understand how to truly tune the fashions, the extra success you\u2019ll get out of it instantly with a really small timeline. And also you\u2019ll perceive how lengthy one thing will take versus when you don\u2019t have that have, you&#8217;ll wrestle and also you won&#8217;t have the ability to get to the best lead to the best time-frame, to make sense from an ROI perspective.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=395\" target=\"_blank\" rel=\"noreferrer noopener\">06.35<\/a><br \/><strong>So the place does retrieval-augmented technology fall into the spectrum of the instruments within the toolbox?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=404\" target=\"_blank\" rel=\"noreferrer noopener\">06.44<\/a><br \/>Yeah. I feel RAG is a method to truly immediate the mannequin and use search principally to go looking by way of a bunch of paperwork and selectively add issues into the context, whether or not or not it&#8217;s the context is simply too small, so like, it could actually solely deal with a specific amount of data, otherwise you don\u2019t need to distract the mannequin with a bunch of irrelevant data, solely the related data from retrieval.<\/p>\n<p>I feel retrieval is a really highly effective search device. And I feel it\u2019s essential to know that when you use it at inference time fairly a bit, that is one thing you train the mannequin to make use of higher. It\u2019s a device that the mannequin must learn to use, and it may be taught in post-training for the mannequin to truly do retrieval, do RAG, extraordinarily successfully, in various kinds of RAG as properly.<\/p>\n<p>So I feel figuring out that&#8217;s truly pretty essential. For instance, within the RL environments that I create, and the fine-tuning type of information that I create, I embrace RAG examples as a result of I need the mannequin to have the ability to be taught that and have the ability to use RAG successfully.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=466\" target=\"_blank\" rel=\"noreferrer noopener\">07.46<\/a><br \/><strong>So in addition to supervised fine-tuning, the opposite class of strategies, broadly talking, falls below reinforcement studying for post-training. However the impression I get\u2014and I\u2019m a giant RL fan, and I\u2019m a cheerleader of RL\u2014however it appears at all times simply across the nook, past the grasp of standard enterprise. It looks like a category of instruments that the labs, the neo labs and the AI labs, can do properly, however it simply looks like the tooling just isn&#8217;t there to make it, you realize.\u00a0.\u00a0. Like I describe supervised fine-tuning as largely solved if in case you have a service. There\u2019s no equal factor for RL, proper?\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=515\" target=\"_blank\" rel=\"noreferrer noopener\">08.35<\/a><br \/>That\u2019s proper. And I feel SFT (supervised fine-tuning) got here first, so then it has been allowed to mature over time. And so proper now RL is type of seeing that second as properly. It was a really thrilling yr final yr, after we used a bunch of RL at test-time compute, educating a mannequin to purpose, and that was actually thrilling with RL. And so I feel that\u2019s ramped up extra, however we don\u2019t have as many companies as we speak which are capable of assist with that. I feel it\u2019s solely a matter of time, although.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=544\" target=\"_blank\" rel=\"noreferrer noopener\">09.04<\/a><br \/><strong>So that you stated earlier, it\u2019s essential for enterprises to know that these strategies exist, that there\u2019s firms who will help you with these strategies, however it is perhaps an excessive amount of of a raise to attempt to do it your self.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=560\" target=\"_blank\" rel=\"noreferrer noopener\">09.20<\/a><br \/>I feel perhaps totally finish to finish, it&#8217;s difficult as an enterprise. I feel there are particular person builders who&#8217;re in a position to do that and truly get plenty of worth from it. For instance, for imaginative and prescient language fashions or for fashions that generate photographs, persons are doing plenty of bits and items of fine-tuning, and getting very customized outcomes that they want from these fashions.<\/p>\n<p>So I feel it depends upon who you might be and what you\u2019re surrounded by. The <a href=\"https:\/\/tinker-docs.thinkingmachines.ai\/\" target=\"_blank\" rel=\"noreferrer noopener\">Tinker API<\/a> from Considering Machines is absolutely fascinating to me as a result of that permits one other set of individuals to have the ability to entry it. I\u2019m not fairly positive it\u2019s fairly on the enterprise degree, however I do know researchers at universities now have entry to distributed compute, like doing post-training on distributed compute, and fairly massive clusters\u2014which is kind of difficult to do for them. And in order that makes it truly attainable for a minimum of that section of the market and that consumer base to truly get began.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=621\" target=\"_blank\" rel=\"noreferrer noopener\">10.21<\/a><br \/><strong>Yeah. So for our listeners who&#8217;re aware of simply plain inference, the OpenAI API has turn into type of the de facto API for inference. After which the concept is that this Tinker API would possibly play that function for fine-tuning inputs, appropriate? It\u2019s not type of the entire venture that\u2019s there.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=643\" target=\"_blank\" rel=\"noreferrer noopener\">10.43<\/a><br \/>Right. Yeah, that\u2019s their intention. And to do it in a heavy like distributed means.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=649\" target=\"_blank\" rel=\"noreferrer noopener\">10.49<\/a><br \/><strong>So then, if I\u2019m CTO at an enterprise and I&#8217;ve an AI staff and, you realize, we\u2019re lower than pace on post-training, what are the steps to try this? Will we usher in consultants they usually clarify to us, right here\u2019s your choices and these are the distributors, or.\u00a0.\u00a0.? What\u2019s the best playbook?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=675\" target=\"_blank\" rel=\"noreferrer noopener\">11.15<\/a><br \/>Nicely, the technique I&#8217;d make use of is, given these fashions change their capabilities continually, I&#8217;d clearly have groups testing the boundaries of the most recent iteration of mannequin at inference. After which from a post-training perspective, I&#8217;d even be testing that. I&#8217;d have a small, hopefully elite staff that&#8217;s wanting into what I can do with these fashions, particularly the open ones. And after I post-train, what truly comes from that. And I&#8217;d take into consideration my use instances and the specified issues I&#8217;d need to see from the mannequin given my understanding of post-training.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=708\" target=\"_blank\" rel=\"noreferrer noopener\">11.48<\/a><br \/>So hopefully you find out about post-training by way of this e book with O\u2019Reilly. However you\u2019re additionally capable of now grasp like, What are the kinds of capabilities I can add into the mannequin? And because of this, what sorts of issues can I then add into the ecosystem such that they get included into the subsequent technology of mannequin as properly?<\/p>\n<p>For instance, I used to be at an occasion just lately and somebody stated, oh, you realize, these fashions are so scary. Once you threaten the mannequin, you may get higher outcomes. So is that even moral? You realize, the mannequin will get scared and will get you a greater end result. And I stated, truly, you possibly can post-train that out of the mannequin. The place if you threaten it, it truly doesn\u2019t provide you with a greater end result. That\u2019s not truly like a legitimate mannequin habits. You may change that habits of the mannequin. So understanding these instruments can lend that perspective of, oh, I can change this habits as a result of I can change what output given this enter. Like how the mannequin reacts to such a enter. And I understand how.\u00a0<\/p>\n<p>I additionally know the instruments proper. One of these information. So perhaps I must be releasing such a information extra. I must be releasing a majority of these tutorials extra that truly helps the mannequin be taught at totally different ranges of problem. And I must be releasing a majority of these information, a majority of these instruments, a majority of these MCPs and expertise such that the mannequin truly does choose that up.<\/p>\n<p>And that can be throughout all various kinds of fashions, whether or not that be a frontier lab taking a look at your information or your inner staff that&#8217;s doing a little post-training with that data.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=800\" target=\"_blank\" rel=\"noreferrer noopener\">13.20<\/a><br \/><strong>Let\u2019s say I\u2019m certainly one of these enterprises, and we have already got some primary purposes that use RAG, and you realize, I hear this podcast and say, OK, let\u2019s do this, attempt to go down the trail of post-training. So we have already got some familiarity with methods to do eval for RAG or another primary AI utility. How does my eval pipeline change in gentle of post-training? Do I&#8217;ve to vary something there?\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=843\" target=\"_blank\" rel=\"noreferrer noopener\">14.03<\/a><br \/>Sure and no. I feel you possibly can increase on what you might have proper now. And I feel your present eval\u2014hopefully it\u2019s a very good eval. There\u2019s additionally greatest practices round evals. However basically let\u2019s say it\u2019s only a record of attainable inputs and outputs, a method to grade these outputs, for the mannequin. And it covers a good distribution over the duties you care about. Then, sure, you possibly can prolong that to post-training.\u00a0<\/p>\n<p>For fine-tuning, it\u2019s a reasonably simple type of extension. You do want to consider basically the distribution of what you\u2019re evaluating such that you would be able to belief that the mannequin\u2019s actually higher at your duties. After which for RL, you&#8217;ll take into consideration, How do I successfully grade this at each step of the way in which, and have the ability to perceive has the mannequin carried out properly or not and have the ability to catch the place the mannequin is, for instance, reward hacking when it\u2019s dishonest, so to talk?<\/p>\n<p>So I feel you possibly can take what you might have proper now. And that\u2019s type of the great thing about it. You may take what you might have after which you possibly can increase it for post-training.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=910\" target=\"_blank\" rel=\"noreferrer noopener\">15.10<\/a><br \/><strong>So, Sharon, ought to folks consider one thing like supervised fine-tuning as one thing you do for one thing very slender? In different phrases, as you realize, one of many challenges with supervised fine-tuning is that to start with, you need to give you the dataset, and let\u2019s say you are able to do that, you then do the supervised fine-tuning, and it really works, however it solely works for type of that information distribution in some way. And so in different phrases, you shouldn\u2019t anticipate miracles, proper?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=944\" target=\"_blank\" rel=\"noreferrer noopener\">15.44<\/a><br \/>Sure, truly one thing I do suggest is pondering by way of what you need to do this supervised fine-tuning on. And actually, I feel it must be habits adaptation. So for instance, in pretraining, that\u2019s when the mannequin is studying from an enormous quantity of knowledge, for instance, from the web, curated. And it\u2019s simply gaining uncooked intelligence throughout plenty of totally different duties and plenty of totally different domains. And it\u2019s simply gaining that data, predicting that subsequent token. However it doesn\u2019t actually have any of these behavioral parts to it.\u00a0<\/p>\n<p>Now, let\u2019s say it\u2019s solely discovered about model certainly one of some library. If in fine-tuning, so if in post-training, you now give it examples of chatting with the mannequin, then it\u2019s in a position to have the ability to chat over model one and model zero. (Let\u2019s say there\u2019s a model zero.) And also you solely gave it examples of chatting with model one, however it\u2019s capable of generalize that model zero. Nice. That\u2019s precisely what you need. That\u2019s a habits change that you just\u2019re making within the mannequin. However we\u2019ve additionally seen points the place, when you for instance now give the mannequin in fine-tuning examples of \u201coh, right here\u2019s one thing with model two,\u201d however the base mannequin, the pretrained mannequin didn&#8217;t ever see something about model two, it&#8217;s going to be taught this habits of constructing issues up. And so that may generalize as properly. And that might truly damage the mannequin.\u00a0<\/p>\n<p>So one thing that I actually encourage folks to consider is the place to place every step of data. And it\u2019s attainable that sure quantities of data are greatest carried out as extra of a pretraining step. So I\u2019ve seen folks take a pretrained mannequin, do some continued pretraining\u2014perhaps you name it midtraining, I\u2019m undecided. However like one thing there\u2014and you then do this fine-tuning step of habits modification on prime.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1056\" target=\"_blank\" rel=\"noreferrer noopener\">17.36<\/a><br \/><strong>In your earlier startup, you of us talked about one thing.\u00a0.\u00a0. I neglect. I\u2019m making an attempt to recollect. One thing referred to as reminiscence tuning, is that proper?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1066\" target=\"_blank\" rel=\"noreferrer noopener\">17.46<\/a><br \/>Yeah. A combination of reminiscence specialists.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1068\" target=\"_blank\" rel=\"noreferrer noopener\">17.48<\/a><br \/><strong>Yeah, yeah. Is it honest to solid that as a type of post-training?\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1074\" target=\"_blank\" rel=\"noreferrer noopener\">17.54<\/a><br \/>Sure, that&#8217;s completely a type of post-training. We have been doing it within the adapter house.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1079\" target=\"_blank\" rel=\"noreferrer noopener\">17.59<\/a><br \/><strong>Yeah. And it is best to describe for our viewers what that&#8217;s.<\/strong>\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1082\" target=\"_blank\" rel=\"noreferrer noopener\">18.02<\/a><br \/>Okay. Yeah. So we invented one thing referred to as combination of reminiscence specialists. And basically, you possibly can hear just like the phrases, apart from the phrase \u201creminiscence,\u201d it\u2019s a mix of specialists. So it\u2019s a kind of MOE. MOEs are usually carried out within the base layer of a mannequin. And what it principally means is like there are a bunch of various specialists, and for explicit requests, for a specific enter immediate, it routes to solely a type of specialists or solely a few these specialists as a substitute of the entire mannequin.<\/p>\n<p>And this makes latency actually low and makes it actually environment friendly. And the bottom fashions are sometimes MOEs as we speak for the frontier fashions. However what we have been doing was serious about, properly, what if we froze your base mannequin, your base pretrained mannequin, and for post-training, we may do an MOE on prime? And particularly, we may do an MOE on prime by way of the adapters. So by way of your LoRA adapters. And so as a substitute of only one LoRA adopter, you would have a mix of those LoRA adopters. And they might successfully have the ability to be taught a number of totally different duties on prime of your base mannequin such that you&#8217;d have the ability to preserve your base mannequin utterly frozen and have the ability to, routinely in a discovered means, change between these adapters.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1152\" target=\"_blank\" rel=\"noreferrer noopener\">19.12<\/a><br \/><strong>So the consumer expertise or developer expertise is much like supervised fine-tuning: I&#8217;ll want labeled datasets for this one, one other set of labeled datasets for this one, and so forth.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1169\" target=\"_blank\" rel=\"noreferrer noopener\">19.29<\/a><br \/>So truly, yeah. Just like supervised fine-tuning, you&#8217;ll simply have.\u00a0.\u00a0. Nicely, you would put it into one big dataset, and it might learn to determine which adapters to allocate it to. So let\u2019s say you had 256 adapters or 1024 adapters. It will be taught what the optimum routing is.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1187\" target=\"_blank\" rel=\"noreferrer noopener\">19.47<\/a><br \/><strong>And you then of us tried to elucidate this within the context of neural plasticity, as I recall.<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1195\" target=\"_blank\" rel=\"noreferrer noopener\">19.55<\/a><br \/>Did we? I don\u2019t know.\u00a0.\u00a0.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1198\" target=\"_blank\" rel=\"noreferrer noopener\">19.58<\/a><br \/><strong>The concept being that, due to this strategy, your mannequin could be far more dynamic.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1208\" target=\"_blank\" rel=\"noreferrer noopener\">20.08<\/a><br \/>Yeah. I do assume there\u2019s a distinction between inference, so simply going forwards within the mannequin, versus with the ability to go backwards indirectly, whether or not that be by way of the complete mannequin or by way of adapters, however indirectly with the ability to be taught one thing by way of backprop.<\/p>\n<p>So I do assume there&#8217;s a fairly basic distinction between these two kinds of methods to have interaction with a mannequin. And arguably at inference time, your weights are frozen, so the mannequin\u2019s \u201cmind\u201d is totally frozen, proper? And so you possibly can\u2019t actually closely adapt something in the direction of a unique goal. It\u2019s frozen. So with the ability to frequently modify what the mannequin\u2019s goal and pondering and steering and habits is, I feel it\u2019s helpful now.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1254\" target=\"_blank\" rel=\"noreferrer noopener\">20.54<\/a><br \/>I feel there are extra approaches to this as we speak, however from a consumer expertise perspective, some folks have discovered it simpler to only load plenty of issues into the context. And I feel there\u2019s.\u00a0.\u00a0. I\u2019ve truly just lately had this debate with a couple of folks round whether or not in-context studying actually is someplace in between simply frozen inference forwards and backprop. Clearly it\u2019s not doing backprop immediately, however there are methods to imitate sure issues. However perhaps that&#8217;s what we\u2019re doing as a human all through the day. After which I&#8217;ll backprop at night time after I\u2019m sleeping.\u00a0<\/p>\n<p>So I feel persons are taking part in with these concepts and making an attempt to grasp what\u2019s happening with the mannequin. I don\u2019t assume it\u2019s definitive but. However we do see some properties, when simply taking part in with the enter immediate. However there I feel, for sure, there are 100% basic variations when you&#8217;ll be able to backprop into the weights.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1309\" target=\"_blank\" rel=\"noreferrer noopener\">21.49<\/a><br \/><strong>So perhaps for our listeners, briefly outline in-context studying.<\/strong>\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1315\" target=\"_blank\" rel=\"noreferrer noopener\">21.55<\/a><br \/>Oh, yeah. Sorry. So in-context studying is a misleading time period as a result of the phrase \u201cstudying\u201d doesn\u2019t truly.\u00a0.\u00a0. Backprop doesn\u2019t occur. All it&#8217;s is definitely placing examples into the immediate of the mannequin and also you simply run inference. However on condition that immediate, the mannequin appears to be taught from these examples and have the ability to be nudged by these examples to a unique reply.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1337\" target=\"_blank\" rel=\"noreferrer noopener\">22.17<\/a><br \/><strong>By the way in which, now we have now frameworks like DSPy, which comes with instruments like GEPA which may optimize your prompts. I do know a couple of years in the past, you of us have been telling folks [that] prompting your means by way of an issue just isn&#8217;t the best strategy. However now we have now extra principled methods, Sharon, of growing the best prompts? So how do instruments like that influence post-training?\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1371\" target=\"_blank\" rel=\"noreferrer noopener\">22.51<\/a><br \/>Oh, yeah. Instruments like that influence post-training, as a result of you possibly can train the mannequin in post-training to make use of these instruments extra successfully. Particularly if they assist with optimizing the immediate and optimizing the understanding of what somebody is placing into the mannequin.<\/p>\n<p>For instance, let me simply give a distinction of how far we\u2019ve gotten. So post-training makes the mannequin extra resilient to totally different prompts and have the ability to deal with various kinds of prompts and to have the ability to get the intention from the consumer. In order an excessive instance, earlier than ChatGPT, after I was utilizing GPT-3 again in 2020, if I actually put an area accidentally on the finish of my immediate\u2014like after I stated, \u201cHow are you?\u201d however I unintentionally pressed Area after which Enter, the mannequin utterly freaked out. And that\u2019s due to the way in which issues have been tokenized, and that simply would mess issues up. However there are plenty of totally different bizarre sensitivities within the mannequin such that it might simply utterly freak out, and by freak out I imply it might simply repeat the identical factor again and again, or simply go off the rails about one thing utterly irrelevant.<\/p>\n<p>And in order that\u2019s what the state of issues have been, and the mannequin was not post-trained to.\u00a0.\u00a0. Nicely, it wasn\u2019t fairly post-trained then, however it additionally wasn\u2019t typically post-trained to be resilient to any kind of immediate, versus now as we speak, I don\u2019t find out about you, however the way in which I code is I simply spotlight one thing and simply put a query mark into the immediate.<\/p>\n<p>I\u2019m so lazy, or like simply put the error in and it\u2019s capable of deal with it\u2014perceive that you just\u2019re making an attempt to repair this error as a result of why else would you be speaking to it. And so it\u2019s simply far more resilient as we speak to various things within the immediate.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1466\">24.26<\/a><br \/>Bear in mind Google \u201cDid you imply this?\u201d It\u2019s type of an excessive model of that, the place you kind one thing utterly misspelled into Google, and it\u2019s capable of type of determine what you truly meant and provide the outcomes.<\/p>\n<p>It\u2019s the identical factor, much more excessive, like tremendous Google, so to talk. However, yeah, it\u2019s resilient to that immediate. However that must be carried out by way of post-training\u2014that&#8217;s taking place in post-training for lots of those fashions. It\u2019s displaying the mannequin, hey, for these attainable inputs which are simply gross and tousled, you possibly can nonetheless give the consumer a very well-defined output and perceive their intention.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1505\" target=\"_blank\" rel=\"noreferrer noopener\">25.05<\/a><br \/><strong>So the new factor as we speak, in fact, is brokers. And brokers now, persons are utilizing issues like device calling, proper? So MCP servers.\u00a0.\u00a0. You\u2019re not as depending on this monolithic mannequin to unravel all the things for you. So you possibly can simply use a mannequin to orchestrate a bunch of little specialised specialist brokers.<\/strong><\/p>\n<p><strong>So do I nonetheless want post-training?\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1539\" target=\"_blank\" rel=\"noreferrer noopener\">25.39<\/a><br \/>Oh, completely. You utilize post-training to get the agent to truly work.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1543\" target=\"_blank\" rel=\"noreferrer noopener\">25.43<\/a><br \/><strong>So get the agent to drag all the best instruments.\u00a0.\u00a0.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1546\" target=\"_blank\" rel=\"noreferrer noopener\">25.46<\/a><br \/>Yeah, truly, an enormous purpose why hallucinations have been, like, a lot better than earlier than is as a result of now, below the hood, they\u2019ve taught the mannequin to perhaps use a calculator device as a substitute of simply output, you realize, math by yourself, or have the ability to use the search API as a substitute of make issues up out of your pretraining information.<\/p>\n<p>So this device calling is absolutely, actually efficient, however you do want to show the mannequin to make use of it successfully. And I truly assume what\u2019s fascinating.\u00a0.\u00a0. So MCPs have managed to create a terrific middleman layer to assist fashions have the ability to name various things, use various kinds of instruments with a constant interface. Nonetheless, I&#8217;ve discovered that because of in all probability a bit of bit lack of post-training on MCPs, or not as a lot as, say, a Python API, if in case you have a Python perform declaration or a Python API, that\u2019s truly the fashions truly are inclined to do empirically, a minimum of for me, higher on it as a result of fashions have seen so many extra examples of that. In order that\u2019s an instance of, oh, truly in post-training I did see extra of that than MCPs.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1612\" target=\"_blank\" rel=\"noreferrer noopener\">26.52<\/a><br \/>So weirdly, it\u2019s higher utilizing Python APIs on your identical device than an MCP of your individual device, empirically as we speak. And so I feel it actually depends upon what it\u2019s been post-trained on. And understanding that post-training course of and likewise what goes into that may assist you to perceive why these variations happen. And likewise why we&#8217;d like a few of these instruments to assist us, as a result of it\u2019s a bit of bit chicken-egg, however just like the mannequin is able to sure issues, calling totally different instruments, and many others. However having an MCP layer is a means to assist everybody manage round a single interface such that we are able to then do post-training on these fashions such that they will then do properly on it.<\/p>\n<p>I don\u2019t know if that is smart, however yeah, that\u2019s why it\u2019s so essential.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1661\" target=\"_blank\" rel=\"noreferrer noopener\">27.41<\/a><br \/><strong>Yeah, yeah. Within the areas I\u2019m fascinated with, which I imply, the info engineering, DevOps type of purposes, it looks like there\u2019s new instruments like Dex, open supply instruments, which let you type of save pipelines or playbooks that work so that you just don\u2019t continually must reinvent the wheel, you realize, simply because principally, that\u2019s how these items perform anyway, proper? So somebody will get one thing to work after which everybody type of advantages from that. However then when you\u2019re continually ranging from scratch, and also you immediate after which the agent has to relearn all the things from scratch when it turns on the market\u2019s already a recognized means to do that downside, it\u2019s simply not environment friendly, proper?\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1710\" target=\"_blank\" rel=\"noreferrer noopener\">28.30<\/a><br \/>Oh, I additionally assume one other thrilling frontier that\u2019s type of within the zeitgeist of as we speak is, you realize, given Moltbook or OpenClaw stuff, multi-agent has been talked about far more. And that\u2019s additionally by way of post-training for the mannequin, to launch subagents and to have the ability to interface with different brokers successfully. These are all kinds of habits that we have now to show the mannequin to have the ability to deal with. It\u2019s capable of do plenty of this out of the field, identical to GPT-3 was capable of chat with you when you give it the best nudging prompts, and many others., however ChatGPT is so a lot better at chatting with you.<\/p>\n<p>So it\u2019s the identical factor. Like now persons are, you realize, including to their post-training combine this multi-agent workflow or subagent workflow. And that\u2019s actually, actually essential for these fashions to be efficient at with the ability to do this. To be each the principle agent, the unified agent on the prime, but in addition to be the subagent to have the ability to launch its personal subagents as properly.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1766\" target=\"_blank\" rel=\"noreferrer noopener\">29.26<\/a><br \/><strong>One other development just lately is the emergence of those multimodal fashions and even, persons are beginning to discuss world fashions. I do know these are early, however I feel even simply within the space of multimodality, visible language fashions, and so forth, what&#8217;s the state of post-training exterior of simply LLMs? Simply totally different sorts of this far more multimodal basis fashions? Are folks doing the post-training in these frontier fashions as properly?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1804\" target=\"_blank\" rel=\"noreferrer noopener\">30.04<\/a><br \/>Oh, completely. I truly assume one actually enjoyable one\u2014I suppose that is largely a language mannequin, however they&#8217;re possible tokenizing very in another way\u2014are people who find themselves taking a look at, for instance, life sciences and post-training basis fashions for that.<\/p>\n<p>So there you&#8217;ll need to adapt the tokenizer, since you wished to have the ability to put various kinds of tokens in and tokens out, and have the mannequin be very environment friendly at that. And so that you\u2019re doing that in post-training, in fact, to have the ability to train that new tokenizer. However you\u2019re additionally serious about what different suggestions loops you are able to do.<\/p>\n<p>So persons are automating issues like, I don\u2019t know, the pipetting and testing out the totally different, you realize, molecules, mixing them collectively and with the ability to get a end result from that. After which, you realize, utilizing that as a reward sign again into the mannequin. In order that\u2019s a very highly effective different kind of area that\u2019s perhaps adjoining to how we take into consideration language fashions, however tokenized in another way, and has discovered an fascinating area of interest the place we are able to get good, verifiable rewards again into the mannequin that&#8217;s fairly totally different from how we take into consideration, for instance, coding or math, and even normal human preferences. It\u2019s touching the true world or bodily world\u2014so it\u2019s in all probability all actual, however the bodily world a bit of bit extra.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1885\" target=\"_blank\" rel=\"noreferrer noopener\">31.25<\/a><br \/><strong>So in closing, let\u2019s get your very fast takes on a couple of of those AI sizzling subjects. First one, reinforcement studying. When will it turn into mainstream?\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1898\" target=\"_blank\" rel=\"noreferrer noopener\">31.38<\/a><br \/>Mainstream? How is it not mainstream?\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1900\" target=\"_blank\" rel=\"noreferrer noopener\">31.40<\/a><br \/><strong>No, no, I imply, for normal enterprises to have the ability to do it themselves.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1907\" target=\"_blank\" rel=\"noreferrer noopener\">31.47<\/a><br \/>This yr. Folks have gotten to be sprinting. Come on.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1910\" target=\"_blank\" rel=\"noreferrer noopener\">31.50<\/a><br \/><strong>You assume? Do you assume there can be instruments on the market in order that I don\u2019t want in-house expertise in RL to do it myself?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1919\" target=\"_blank\" rel=\"noreferrer noopener\">31.59<\/a><br \/>Sure. Yeah.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1921\" target=\"_blank\" rel=\"noreferrer noopener\">32.01<\/a><br \/><strong>Secondly, scaling. Is scaling nonetheless the way in which to go? The frontier labs appear to assume so. They assume that greater is best. So are you listening to something within the analysis frontiers that let you know, hey, perhaps there\u2019s alternate options to only pure scaling?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1940\" target=\"_blank\" rel=\"noreferrer noopener\">32.20<\/a><br \/>I nonetheless consider in scaling. I consider we\u2019ve not met a restrict but. Not seen a plateau but. I feel the factor folks want to acknowledge is that it\u2019s at all times been a \u201c10X compute for 2X intelligence\u201d kind of curve. So it\u2019s not precisely like 10X-10X. However yeah, I nonetheless consider in scaling, and we haven\u2019t actually seen an empirical plateau on that but.<\/p>\n<p>That being stated, I\u2019m actually enthusiastic about individuals who problem it. As a result of I feel it might be actually wonderful if we may problem it and get an enormous quantity of intelligence with much less pure {dollars}, particularly now as we begin to hit up on trillions of {dollars} in among the frontier labs, of like that\u2019s the subsequent degree of scale that they\u2019ll be seeing. Nonetheless, at a compute firm, I\u2019m okay with this buy. Come spend trillions! [laughs]\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=1993\" target=\"_blank\" rel=\"noreferrer noopener\">33.13<\/a><br \/><strong>By the way in which, with respect to scaling, so that you assume the fashions we have now now, even when you cease progress, there\u2019s plenty of adaptation that enterprises can do. And there\u2019s plenty of advantages from the fashions we have already got as we speak?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2010\" target=\"_blank\" rel=\"noreferrer noopener\">33.30<\/a><br \/>Right. Sure. We\u2019re not even scratching the floor, I feel.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2014\" target=\"_blank\" rel=\"noreferrer noopener\">33.34<\/a><br \/><strong>The third matter I wished to select your mind fast is \u201copen\u201d: open supply, open weights, no matter. So, there\u2019s nonetheless a spot, I feel.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2029\" target=\"_blank\" rel=\"noreferrer noopener\">33.49<\/a><br \/>There are contenders within the US who need to be an open supply DeepSeek competitor however American, to make it extra amenable when promoting into.\u00a0.\u00a0.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2042\" target=\"_blank\" rel=\"noreferrer noopener\">34.02<\/a><br \/><strong>They don\u2019t exist, proper? I imply, there\u2019s <\/strong><a href=\"https:\/\/allenai.org\/\" target=\"_blank\" rel=\"noreferrer noopener\"><strong>Allen<\/strong><\/a><strong>.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2046\" target=\"_blank\" rel=\"noreferrer noopener\">34.06<\/a><br \/>Oh, like Ai2 for <a href=\"https:\/\/allenai.org\/olmo\" target=\"_blank\" rel=\"noreferrer noopener\">Olmo<\/a>\u2026 Their startup\u2019s doing a little stuff. I don\u2019t know in the event that they\u2019ve introduced issues but, however yeah hopefully we\u2019ll hear from them quickly.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2055\" target=\"_blank\" rel=\"noreferrer noopener\">34.15<\/a><br \/><strong>Yeah yeah yeah.\u00a0<\/strong><\/p>\n<p><strong>One other fascinating factor about these Chinese language AI groups is clearly, you might have the massive firms like Tencent, Baidu, Alibaba\u2014so that they\u2019re doing their factor. However then there\u2019s this wave of startups. Put aside DeepSeek. So the opposite startups on this house, it looks like they\u2019re focusing on the West as properly, proper? As a result of principally it\u2019s arduous to monetize in China, as a result of folks have a tendency to not pay, particularly the enterprises. [laughs]<\/strong><\/p>\n<p><strong>I\u2019m simply noticing plenty of them are incorporating in Singapore after which making an attempt to construct options for outdoor of China.<\/strong>\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2100\" target=\"_blank\" rel=\"noreferrer noopener\">35.00<\/a><br \/>Nicely, the TAM is kind of massive right here, so.\u00a0.\u00a0. It\u2019s fairly massive in each locations.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2107\" target=\"_blank\" rel=\"noreferrer noopener\">35.07<\/a><br \/><strong>So it\u2019s the ultimate query. So we\u2019ve talked about post-training. We talked about the advantages, however we additionally talked concerning the challenges. And so far as I can inform, one of many challenges is, as you identified, to do it finish to finish requires a bit of experience. To begin with, take into consideration simply the info. You would possibly want the best information platform or information infrastructure to prep your information to do no matter it&#8217;s that you just\u2019re doing for post-training. And you then get into RL.\u00a0<\/strong><\/p>\n<p><strong>So what are among the key foundational issues that enterprises ought to spend money on to set themselves up for post-training\u2014to get actually good at publish coaching? So I discussed a knowledge platform, perhaps spend money on the info. What else?<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2161\" target=\"_blank\" rel=\"noreferrer noopener\">36.01<\/a><br \/>I feel the kind of information platform issues. I\u2019m undecided if I completely am purchased into how CIOs are approaching it as we speak. I feel what issues at that infrastructure layer is definitely ensuring you deeply perceive what duties you need these fashions to do. And never solely that, however then codifying it indirectly\u2014whether or not that be inputs and outputs and, you realize, desired outputs, whether or not that be a method to grade outputs, whether or not that be the best atmosphere to have the agent in. Having the ability to articulate that&#8217;s extraordinarily highly effective and I feel is the one of many key methods of getting that job that you really want this agent to do, for instance, to be truly inside the mannequin. Whether or not it\u2019s you doing post-training or another person doing post-training, it doesn&#8217;t matter what, when you construct that, that can be one thing that provides a excessive ROI, as a result of anybody will have the ability to take that and have the ability to embed it and also you\u2019ll have the ability to get that functionality sooner than anybody else.\u00a0<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2223\" target=\"_blank\" rel=\"noreferrer noopener\">37.03<\/a><br \/><strong>And on the {hardware} facet, one fascinating factor that comes out of this dialogue is that if RL actually turns into mainstream, then you must have a wholesome mixture of CPUs and GPUs as properly.\u00a0<\/strong><\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2237\" target=\"_blank\" rel=\"noreferrer noopener\">37.17<\/a><br \/>That\u2019s proper. And you realize, AMD makes each.\u00a0.\u00a0.<\/p>\n<p><a href=\"https:\/\/cdn.oreillystatic.com\/radar\/generative-ai-real-world-podcast\/sharon_zhou_genai_podcast_v2.mp4#t=2245\" target=\"_blank\" rel=\"noreferrer noopener\">37.25<\/a><br \/><strong>It\u2019s nice at each of these.<\/strong><\/p>\n<p><strong>And with that thanks, Sharon.<\/strong><\/p>\n<\/div>\n<iframe data-lazy=\"true\" data-src=\"https:\/\/www.fiverr.com\/gig_widgets?id=U2FsdGVkX18x7XQvttUTrv1oEqmGNGTgvvCUiUoJ\/AP4z\/UyMz8lXGOLpu15jIMxBbTR0gmD5uBoFvhC4KWeALQRp3h\/X\/AwcVD0K8Wj9H\/ZzYKzcCNHosB9oS4SCJJFWiN85P9ICAc4OgCoE\/wHKIY7CDkf2\/DQ1vqGvk4smVe5cRDEmrLPCWi4FC8p40VUhSmWQ5udCm0zoJtorgWv3vbDQw0kKYkwn39ozAnQXDe+YvWMxkLFWA+O3TFwkJvdkIK+\/AUSnRssPKt5WHY0FhNOxnSPcLslEL4G4\/RfP95ve99U+kRnDy3X+KtzdQLY+u935ghON\/o3UE4IMv9oN6JX9RnxzL\/LRcOgnHigxStSGPKsZYtnz8RWNVT\/rOLAibqiWJadC5MYHRbekF3eg6FOGrQGkXYbsn0+a5aovnlLCbLwIqY9fcS17UX8J235iQ6cdmHNbrPeS84CMm34RA==&affiliate_id=1052423&strip_google_tagmanager=true\" loading=\"lazy\" data-with-title=\"true\" class=\"fiverr_nga_frame\" frameborder=\"0\" height=\"350\" width=\"100%\" referrerpolicy=\"no-referrer-when-downgrade\" data-mode=\"random_gigs\" onload=\" var frame = this; var script = document.createElement('script'); script.addEventListener('load', function() { window.FW_SDK.register(frame); }); script.setAttribute('src', 'https:\/\/www.fiverr.com\/gig_widgets\/sdk'); document.body.appendChild(script); \" ><\/iframe>\n<br \/><a href=\"https:\/\/www.oreilly.com\/radar\/podcast\/generative-ai-in-the-real-world-sharon-zhou-on-post-training\/\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Publish-training will get your mannequin to behave the way in which you need it to. As AMD VP of AI Sharon Zhou explains to Ben&#8230;<\/p>\n","protected":false},"author":1,"featured_media":88939,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[7],"tags":[],"class_list":["post-118266","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-tech-universe"],"acf":[],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Sharon Zhou on Post-Training \u2013 O\u2019Reilly - mailinvest.blog<\/title>\n<meta name=\"description\" content=\"Technology is forever changing, and there are always new pieces of technology to replace obsolete ones. Tons of people enjoy reading tech blogs on a daily basis.mailinvest.blog tracks all the latest consumer technology breakthroughs and shows you what&#039;s new, what matters and how technology can enrich your life. mailinvest.blog also provides the information, tools, and advice that helps when deciding what to buy.\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Sharon Zhou on Post-Training \u2013 O\u2019Reilly - mailinvest.blog\" \/>\n<meta property=\"og:description\" content=\"Technology is forever changing, and there are always new pieces of technology to replace obsolete ones. Tons of people enjoy reading tech blogs on a daily basis.mailinvest.blog tracks all the latest consumer technology breakthroughs and shows you what&#039;s new, what matters and how technology can enrich your life. mailinvest.blog also provides the information, tools, and advice that helps when deciding what to buy.\" \/>\n<meta property=\"og:url\" content=\"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/\" \/>\n<meta property=\"og:site_name\" content=\"mailinvest.blog\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/freelanceracademic\/\" \/>\n<meta property=\"article:published_time\" content=\"2026-03-12T21:15:48+00:00\" \/>\n<meta property=\"article:modified_time\" content=\"2026-03-12T21:17:09+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/mailinvest.blog\/wp-content\/uploads\/2025\/08\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png\" \/>\n\t<meta property=\"og:image:width\" content=\"1600\" \/>\n\t<meta property=\"og:image:height\" content=\"1600\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/png\" \/>\n<meta name=\"author\" content=\"admin@mailinvest.blog\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"admin@mailinvest.blog\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"34 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/\"},\"author\":{\"name\":\"admin@mailinvest.blog\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#\\\/schema\\\/person\\\/012701c4c204d4e4ebd34f926cfd31a4\"},\"headline\":\"Sharon Zhou on Post-Training \u2013 O\u2019Reilly\",\"datePublished\":\"2026-03-12T21:15:48+00:00\",\"dateModified\":\"2026-03-12T21:17:09+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/\"},\"wordCount\":6852,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#organization\"},\"image\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/mailinvest.blog\\\/wp-content\\\/uploads\\\/2025\\\/08\\\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png\",\"articleSection\":[\"Tech Universe\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/\",\"url\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/\",\"name\":\"Sharon Zhou on Post-Training \u2013 O\u2019Reilly - mailinvest.blog\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#website\"},\"primaryImageOfPage\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/#primaryimage\"},\"image\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/#primaryimage\"},\"thumbnailUrl\":\"https:\\\/\\\/mailinvest.blog\\\/wp-content\\\/uploads\\\/2025\\\/08\\\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png\",\"datePublished\":\"2026-03-12T21:15:48+00:00\",\"dateModified\":\"2026-03-12T21:17:09+00:00\",\"description\":\"Technology is forever changing, and there are always new pieces of technology to replace obsolete ones. Tons of people enjoy reading tech blogs on a daily basis.mailinvest.blog tracks all the latest consumer technology breakthroughs and shows you what's new, what matters and how technology can enrich your life. mailinvest.blog also provides the information, tools, and advice that helps when deciding what to buy.\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/\"]}]},{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/#primaryimage\",\"url\":\"https:\\\/\\\/mailinvest.blog\\\/wp-content\\\/uploads\\\/2025\\\/08\\\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png\",\"contentUrl\":\"https:\\\/\\\/mailinvest.blog\\\/wp-content\\\/uploads\\\/2025\\\/08\\\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png\",\"width\":1600,\"height\":1600},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/2026\\\/03\\\/12\\\/sharon-zhou-on-post-training-oreilly\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/mailinvest.blog\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Sharon Zhou on Post-Training \u2013 O\u2019Reilly\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#website\",\"url\":\"https:\\\/\\\/mailinvest.blog\\\/\",\"name\":\"mailinvest.blog\",\"description\":\"Technology is forever changing, and there are always new pieces of technology to replace obsolete ones. Tons of people enjoy reading tech blogs on a daily basis. mailinvest.blog tracks all the latest consumer technology breakthroughs and shows you what&#039;s new, what matters and how technology can enrich your life. mailinvest.blog also provides the information, tools, and advice that helps when deciding what to buy.\",\"publisher\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/mailinvest.blog\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#organization\",\"name\":\"mailinvest\",\"url\":\"https:\\\/\\\/mailinvest.blog\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/mailinvest.blog\\\/wp-content\\\/uploads\\\/2022\\\/01\\\/default.png\",\"contentUrl\":\"https:\\\/\\\/mailinvest.blog\\\/wp-content\\\/uploads\\\/2022\\\/01\\\/default.png\",\"width\":1000,\"height\":1000,\"caption\":\"mailinvest\"},\"image\":{\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/www.facebook.com\\\/freelanceracademic\\\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/mailinvest.blog\\\/#\\\/schema\\\/person\\\/012701c4c204d4e4ebd34f926cfd31a4\",\"name\":\"admin@mailinvest.blog\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/98ed217bd0f3d6a6dcae2d9b0c76e305b049a07275e315e1407e19ec8b08e139?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/98ed217bd0f3d6a6dcae2d9b0c76e305b049a07275e315e1407e19ec8b08e139?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/98ed217bd0f3d6a6dcae2d9b0c76e305b049a07275e315e1407e19ec8b08e139?s=96&d=mm&r=g\",\"caption\":\"admin@mailinvest.blog\"},\"sameAs\":[\"https:\\\/\\\/mailinvest.blog\",\"admin@mailinvest.blog\"],\"url\":\"https:\\\/\\\/mailinvest.blog\\\/index.php\\\/author\\\/adminmailinvest-blog\\\/\"}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Sharon Zhou on Post-Training \u2013 O\u2019Reilly - mailinvest.blog","description":"Technology is forever changing, and there are always new pieces of technology to replace obsolete ones. Tons of people enjoy reading tech blogs on a daily basis.mailinvest.blog tracks all the latest consumer technology breakthroughs and shows you what's new, what matters and how technology can enrich your life. mailinvest.blog also provides the information, tools, and advice that helps when deciding what to buy.","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/","og_locale":"en_US","og_type":"article","og_title":"Sharon Zhou on Post-Training \u2013 O\u2019Reilly - mailinvest.blog","og_description":"Technology is forever changing, and there are always new pieces of technology to replace obsolete ones. Tons of people enjoy reading tech blogs on a daily basis.mailinvest.blog tracks all the latest consumer technology breakthroughs and shows you what's new, what matters and how technology can enrich your life. mailinvest.blog also provides the information, tools, and advice that helps when deciding what to buy.","og_url":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/","og_site_name":"mailinvest.blog","article_publisher":"https:\/\/www.facebook.com\/freelanceracademic\/","article_published_time":"2026-03-12T21:15:48+00:00","article_modified_time":"2026-03-12T21:17:09+00:00","og_image":[{"width":1600,"height":1600,"url":"https:\/\/mailinvest.blog\/wp-content\/uploads\/2025\/08\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png","type":"image\/png"}],"author":"admin@mailinvest.blog","twitter_card":"summary_large_image","twitter_misc":{"Written by":"admin@mailinvest.blog","Est. reading time":"34 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/#article","isPartOf":{"@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/"},"author":{"name":"admin@mailinvest.blog","@id":"https:\/\/mailinvest.blog\/#\/schema\/person\/012701c4c204d4e4ebd34f926cfd31a4"},"headline":"Sharon Zhou on Post-Training \u2013 O\u2019Reilly","datePublished":"2026-03-12T21:15:48+00:00","dateModified":"2026-03-12T21:17:09+00:00","mainEntityOfPage":{"@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/"},"wordCount":6852,"commentCount":0,"publisher":{"@id":"https:\/\/mailinvest.blog\/#organization"},"image":{"@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/#primaryimage"},"thumbnailUrl":"https:\/\/mailinvest.blog\/wp-content\/uploads\/2025\/08\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png","articleSection":["Tech Universe"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/","url":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/","name":"Sharon Zhou on Post-Training \u2013 O\u2019Reilly - mailinvest.blog","isPartOf":{"@id":"https:\/\/mailinvest.blog\/#website"},"primaryImageOfPage":{"@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/#primaryimage"},"image":{"@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/#primaryimage"},"thumbnailUrl":"https:\/\/mailinvest.blog\/wp-content\/uploads\/2025\/08\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png","datePublished":"2026-03-12T21:15:48+00:00","dateModified":"2026-03-12T21:17:09+00:00","description":"Technology is forever changing, and there are always new pieces of technology to replace obsolete ones. Tons of people enjoy reading tech blogs on a daily basis.mailinvest.blog tracks all the latest consumer technology breakthroughs and shows you what's new, what matters and how technology can enrich your life. mailinvest.blog also provides the information, tools, and advice that helps when deciding what to buy.","breadcrumb":{"@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/"]}]},{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/#primaryimage","url":"https:\/\/mailinvest.blog\/wp-content\/uploads\/2025\/08\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png","contentUrl":"https:\/\/mailinvest.blog\/wp-content\/uploads\/2025\/08\/Podcast_Cover_GenAI_in_the_Real_World-1600x1600.png","width":1600,"height":1600},{"@type":"BreadcrumbList","@id":"https:\/\/mailinvest.blog\/index.php\/2026\/03\/12\/sharon-zhou-on-post-training-oreilly\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/mailinvest.blog\/"},{"@type":"ListItem","position":2,"name":"Sharon Zhou on Post-Training \u2013 O\u2019Reilly"}]},{"@type":"WebSite","@id":"https:\/\/mailinvest.blog\/#website","url":"https:\/\/mailinvest.blog\/","name":"mailinvest.blog","description":"Technology is forever changing, and there are always new pieces of technology to replace obsolete ones. Tons of people enjoy reading tech blogs on a daily basis. mailinvest.blog tracks all the latest consumer technology breakthroughs and shows you what&#039;s new, what matters and how technology can enrich your life. mailinvest.blog also provides the information, tools, and advice that helps when deciding what to buy.","publisher":{"@id":"https:\/\/mailinvest.blog\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/mailinvest.blog\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/mailinvest.blog\/#organization","name":"mailinvest","url":"https:\/\/mailinvest.blog\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/mailinvest.blog\/#\/schema\/logo\/image\/","url":"https:\/\/mailinvest.blog\/wp-content\/uploads\/2022\/01\/default.png","contentUrl":"https:\/\/mailinvest.blog\/wp-content\/uploads\/2022\/01\/default.png","width":1000,"height":1000,"caption":"mailinvest"},"image":{"@id":"https:\/\/mailinvest.blog\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/freelanceracademic\/"]},{"@type":"Person","@id":"https:\/\/mailinvest.blog\/#\/schema\/person\/012701c4c204d4e4ebd34f926cfd31a4","name":"admin@mailinvest.blog","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/98ed217bd0f3d6a6dcae2d9b0c76e305b049a07275e315e1407e19ec8b08e139?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/98ed217bd0f3d6a6dcae2d9b0c76e305b049a07275e315e1407e19ec8b08e139?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/98ed217bd0f3d6a6dcae2d9b0c76e305b049a07275e315e1407e19ec8b08e139?s=96&d=mm&r=g","caption":"admin@mailinvest.blog"},"sameAs":["https:\/\/mailinvest.blog","admin@mailinvest.blog"],"url":"https:\/\/mailinvest.blog\/index.php\/author\/adminmailinvest-blog\/"}]}},"_links":{"self":[{"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/posts\/118266","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/comments?post=118266"}],"version-history":[{"count":1,"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/posts\/118266\/revisions"}],"predecessor-version":[{"id":118267,"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/posts\/118266\/revisions\/118267"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/media\/88939"}],"wp:attachment":[{"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/media?parent=118266"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/categories?post=118266"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/mailinvest.blog\/index.php\/wp-json\/wp\/v2\/tags?post=118266"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}