<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:media="http://search.yahoo.com/mrss/"><channel><title><![CDATA[Honest AI]]></title><description><![CDATA[Learn about the future of AI and ML.]]></description><link>https://honest-ai.com/</link><image><url>https://honest-ai.com/favicon.png</url><title>Honest AI</title><link>https://honest-ai.com/</link></image><generator>Ghost 5.70</generator><lastBuildDate>Mon, 13 Apr 2026 18:32:00 GMT</lastBuildDate><atom:link href="https://honest-ai.com/rss/" rel="self" type="application/rss+xml"/><ttl>60</ttl><item><title><![CDATA[Honest-02: Should you build with GPT-3?]]></title><description><![CDATA[<p>Happy Friday everyone,</p><h3 id="updates">Updates</h3><p>This last week was filled with interesting concepts, from the update on Neuralink to discussions of the metaverse. Honest AI appeared in a call-in with Betaworks last Friday, joining panelists discussing how NLP is currently the hottest trend for startups, even if you exclude GPT-3. More</p>]]></description><link>https://honest-ai.com/honest-02-should-you-build-with-gpt-3/</link><guid isPermaLink="false">5f5283a8e83e633e34493bb9</guid><category><![CDATA[Newsletter]]></category><dc:creator><![CDATA[Tyler Lastovich]]></dc:creator><pubDate>Fri, 04 Sep 2020 19:06:39 GMT</pubDate><media:content url="https://honest-ai.com/content/images/2020/09/honest-2newsletter.png" medium="image"/><content:encoded><![CDATA[<img src="https://honest-ai.com/content/images/2020/09/honest-2newsletter.png" alt="Honest-02: Should you build with GPT-3?"><p>Happy Friday everyone,</p><h3 id="updates">Updates</h3><p>This last week was filled with interesting concepts, from the update on Neuralink to discussions of the metaverse. Honest AI appeared in a call-in with Betaworks last Friday, joining panelists discussing how NLP is currently the hottest trend for startups, even if you exclude GPT-3. More improvements to the website have also been shipped, including a new article design. (love it? hate it?)</p><p>Meanwhile, the community has launched! While it is still in it&apos;s earliest form, the first brave users signed up last week. I will continue to add topics, links, and answer questions there so please join in and invite some friends or colleagues.</p><p><strong>I encourage you to <a href="https://honest-ai.com/early-adopter">Sign up to the community for free!</a></strong></p><h1 id="should-you-build-with-gpt-3">Should you build with GPT-3?</h1><p>In the first honest article, I break down the pros and cons of using the OpenAI API as the core part of a business. As with any new piece of technology it is important to understand the risks ahead of time, without getting too caught up in the hype.</p><p><strong><a href="https://honest-ai.com/should-you-build-invest-in-gpt-3/">Read the article here</a></strong></p><h3 id="pricing">Pricing</h3><p>OpenAI debuted their pricing model for GPT-3 this week, which resulted in some pained grimaces. With pricing based per-token, it looks like many fixed-price apps will be rendered DOA. This limits the fun and creative cases along with the daily-helper apps. I am not saying they are wrong in pricing it high; If you have the best, you can charge the most. Based on my personal usage I would already be in the $400/month category. Notably this is just for the beta period, and the largest engine, davinci. We will have to wait and see for more details later. Pricing is enacted starting 10/1.</p><h3 id="deployment">Deployment</h3><p>New guidelines were also released as to which applications can actually use GPT-3. Tiered into 3 categories based in risk, they are taking an active stance to limit bad press. Notably they are banning their API&apos;s use for generating synthetic articles for SEO, pretty much anything to do with politics or medicine, and open-ended chatbots. Every usage will be hand-reviewed by the OpenAI team, which should help to ease fears that the internet will be soon overrun by intelligent bots.</p><p>It is safe to assume that with these updates the crazy hype seen over the last 3 months will start to cool off.</p><h1 id="interesting-links-this-week-">Interesting links this week &#x2013;</h1><ul><li><a href="https://docs.aitextgen.io/?ref=honest-ai.com">aitextgen is a smaller, simpler GPT-2 model you can train for free</a></li></ul><p>If the pricing for GPT-3 has you a little disappointed (&#x1F64B;&#x200D;) then maybe it is time to take a look at a more scaleable option. I came across Max Woolf&apos;s project aitextgen on Twitter and it looks quite interesting for specific use cases. Featuring a teeny-tiny model (1,400 times smaller than the flagship GPT-3), it should be very fast when <a href="https://colab.research.google.com/drive/15qBZx5y9rdaQSyWpsreMDnTiZ5IlN0zD?usp=sharing&amp;ref=honest-ai.com#scrollTo=H7LoMj4GA4n_">trained for specific text generation</a>. Just don&apos;t expect the same level of fake smarts shown by others.</p><ul><li><a href="https://www.youtube.com/watch?v=DVvmgjBL74w&amp;ref=honest-ai.com">Neuralink&apos;s summer 2020 update. &#xA0;Brains + AI = &#x2728;(Youtube)</a></li></ul><p>Neuralink remains one of the most interesting companies in the world. While the aspirational, world-changing goals remain years or decades away, it is hard to not get caught up in the what-ifs. &#xA0;Machine learning is a core component of making sense of the complex waveforms emitted by the brain. If you haven&apos;t seen the video demo yet, check it out.</p><ul><li><a href="https://youtu.be/13CZPWmke6A?ref=honest-ai.com">Who is behind OpenAI? (Youtube)</a></li></ul><p>Do you want to understand a bit more about who designs these complex AI systems? While a few months old, this is a great conversation between Lex Fridman and Ilya Sutskever (co-founder of OpenAI) to see logical thinking in action. </p><h2 id="coming-up-">Coming up &#x2013;</h2><ul><li>The first feature essay! A lengthy piece on AI, APIs, and the Metaverse. This is one you won&apos;t want to miss. </li><li>A dive into why synthetic personas and other virtual media are going to be huge in the near future.</li><li>More info on a sister project to Honest AI, a site that makes it very easy to use transformers (GPT2/3)!</li></ul><p><br>Have a good weekend,</p><p>-Tyler</p>]]></content:encoded></item><item><title><![CDATA[Should you build/invest in GPT-3?]]></title><description><![CDATA[With so much media attention being given to GPT-3 it is easy to get caught up in the commotion. Let's dive a little deeper and go over the pros and cons of the language model as it exists today. Since timing is important, should you start a business around the new language model now, or wait?]]></description><link>https://honest-ai.com/should-you-build-invest-in-gpt-3/</link><guid isPermaLink="false">5f489005e83e633e34493b7a</guid><category><![CDATA[Article]]></category><dc:creator><![CDATA[Tyler Lastovich]]></dc:creator><pubDate>Fri, 28 Aug 2020 07:40:33 GMT</pubDate><media:content url="https://honest-ai.com/content/images/2020/08/invest_GPT-1.png" medium="image"/><content:encoded><![CDATA[<img src="https://honest-ai.com/content/images/2020/08/invest_GPT-1.png" alt="Should you build/invest in GPT-3?"><p>The hype around GPT-3 right now is <em>frothy</em>. It feels like a &apos;game-changer&apos; to many and the potential it unlocks is stated at every turn. But should you really build (or invest in) a company based on GPT-3? Note: these are solely my opinions. I have used GPT-3 quite extensively now, for a number of different tasks and to build demos.</p><p>My take: <strong>Not yet</strong>. &#xA0; <em>(well, maybe)</em></p><p>Let&apos;s back up for a second. &#xA0;Here is what we know about GPT-3.</p><ul><li>It is a huge language model (350gb, 175B parameters)</li><li>It has a 2048 token (~word) attention span, with a maximum output of 512 tokens</li><li>It is in limited, invite-only beta right now. (the slack group is ~1k total members)</li><li>It is run in the Azure cloud and you can only access it via their API.</li><li>When doing complex tasks it is relatively slow to use.</li><li>It is relatively expensive (~2-16&#xA2; per API call) *Edit: see chart below</li><li>In order to use it in production you need to go through a review process and actively limit toxic outputs.</li></ul><p>That all sounds a little scary, right? It should &#xA0;&#x2013;these are non-trivial platform risks that would likely be critical paths for a GPT-3 based business. Are there other competitors to it? &#xA0;Sure. &#xA0;But none match the ease of use, and general quality of output that the huge &apos;davinci&apos; engine returns.</p><h1 id="risks">Risks</h1><h3 id="you-are-at-their-mercy-for-pricing">You are at their mercy for pricing</h3><p>Even if the initial pricing is cheap to start with, without immediate competition they have serious pricing-power. &#xA0;You will be a price taker. &#xA0;Good luck with your chat bot though (sorry).</p><p>**Edit 9/1/20: We now have an idea of how expensive API calls will be. In short it is fairly expensive for an API, costing as much as $0.16 per call. &#xA0;Practically speaking, this means that most applications using GPT-3 will need to bill users based on usage rather than a monthly subscription price. I don&apos;t believe this price point is unfair, but does limit the practical usage of the API. In the personas demo I made, each page load requires 5 api calls (averaging ~1200 tokens each), then another n calls for chat messages. &#xA0;So the rough math looks like 6,000 + 500(n) tokens per generation. I can safely say that just in playing with GPT-3 I have used millions of tokens. Ouch.</p><figure class="kg-card kg-image-card kg-card-hascaption"><img src="https://honest-ai.com/content/images/2020/09/image.png" class="kg-image" alt="Should you build/invest in GPT-3?" loading="lazy" width="1980" height="840" srcset="https://honest-ai.com/content/images/size/w600/2020/09/image.png 600w, https://honest-ai.com/content/images/size/w1000/2020/09/image.png 1000w, https://honest-ai.com/content/images/size/w1600/2020/09/image.png 1600w, https://honest-ai.com/content/images/2020/09/image.png 1980w" sizes="(min-width: 720px) 720px"><figcaption>Beta period pricing, goes into effect 10/1</figcaption></figure><h3 id="you-are-at-their-mercy-for-latency">You are at their mercy for latency</h3><p>GPT-3 is not a tiny little model. &#xA0;It will take some time for results to be returned. With the caveat of <em>it is just in beta testing</em> the model does already show slowdowns under high traffic times (as mentioned in the slack group). This means that the speed might be different in the middle of the day vs at night, or that traffic spikes could be impactful. While smaller prompts for chat have worked quickly in my testing, complicated prompts that are generating written prose take much longer. This lead to problems because....</p><h3 id="the-models-attention-span-is-short-">The models attention span is short.</h3><p>You need to be able to perform all of your work in roughly 2000 words. If &#xA0;you cannot do that you will need to add more complex execution methods such as context-stuffing to help GPT-3 to retain some understanding of what you are trying to accomplish across multiple API calls. This means that you will be making calls in serial to the API, considerably slowing your app down. In a demo I built, I make 5 serial calls to GPT-3 and it takes ~50-80 seconds for the main process to complete. Ouch. I might be able to shave a little time off of this, but not much. &#xA0;That is a long time to wait for a 3rd party API to return. (I remain hopeful the execution speed will get much faster by the end of the beta period!)</p><p>What does this all mean? It means you have to understand what you can reasonably create based on how long operations take.</p><p>Consider an app that is trying to summarize a long input of text (book, article) using basic context-stuffing. You can get more complicated with this, but bear with for a moment.</p><p>You start with 2048 total tokens. &#xA0;From that you need to remove:</p><ul><li>At least one sentence to set initial context (assume 48 tokens for simplicity)</li><li>0-512 tokens to recap previously summarized context</li><li>0-512 tokens for output</li></ul><p>That leaves <code>~1000-1800</code> new tokens that can be summarized per call. This equates to roughly <code>2 - 4</code> written pages or &#xA0;<code>00:07:30 - 00:13:50</code> of transcribed audio.</p><p>Meaning, you can summarize a book with roughly <code>4:1 - 10:1</code> compression in one pass in <code>(pages/4)*5s</code>. &#xA0;I am being generous with the completion time, but a 100 page book would take <code>00:02:05</code>, and a 500-pager would be over 10 minutes. That firmly puts long-form summarization in a &apos;we will email you when your task is done&apos; category.</p><p>**Edit 9/1/20: In the recently published pricing, both the input prompt text and output text count against the charged token count. This makes the use-case of summarization as described above exceedingly expensive. &#xA0;By some napkin maths this means you are looking at a few dollars to summarize a podcast or short book. ~$100 to summarize a Harry Potter book.</p><h3 id="toxic-outputs">Toxic outputs</h3><p>I don&apos;t want to dwell too much on this as I know that OpenAI is actively working on making this a great user experience. That said, right now the early content filter they have in place just isn&apos;t very good. &apos;Toxic&apos; output is just incredibly hard to characterize as context and nuance matters. I was writing movie scenes and it flagged nearly every output as toxic. I am very hopeful they improve on this quickly, but for now I would recommend every company looking to use GPT-3 in user-facing apps needs to implement their own filters. Which is a downer as it significantly reduces the <em>it just works</em> nature of the OpenAI API.</p><h3 id="everyone-has-the-same-engines">Everyone has the same engines</h3><p>As it stands right now, everyone has access to the same four pre-trained engines: ada, babbage, curie, and davinci (ranging from small to large). This means that both you and your competitor will have the ability to get the same information back from the model. All you have the power to modify is the input prompt, and what you do with the outputs. Simply put, the value of your business will not be built using a commodity (the API), it will be built on all the other surrounding processes. For example, there have been many designer or no-code plugins demoed. I promise there is a lot of work happening in the background to show you the end results that have nothing to do with GPT-3. These processes are actually what will make your company or product special, rather than the fact it is &apos;powered by OpenAI&apos;. &#xA0;</p><h1 id="benefits">Benefits</h1><h3 id="it-is-seriously-awesome">It is seriously <em>awesome</em></h3><p>I can&apos;t stress this enough. The OpenAI API is just a great piece of engineering all around and I am very grateful that so many worked hard to produce it &#x1F44F;. It has sparked the interest of so many people to care about AI and in a way is the reason this project (Honest AI) was started now.</p><h3 id="it-is-incredibly-easy-to-use-">It is incredibly easy to use.</h3><p>Using GPT-3 for simple tasks is about as easy as writing your first &apos;hello world&apos;. This is part of the reason that it feels like magic and has allowed so many people to build weekend demos. As it is a transformer, the input you feed it determines the output. This means that instead of engineering code, you have to creatively come up with input prompts. It is really amazing how small differences in prompt setup have material impact on how well the API accomplishes your desired task. This means that creativity is highly valuable when working with GPT-3.</p><h3 id="the-results-are-consistently-great">The results are consistently great</h3><p>The largest engine (davinci) is the first language model that I have ever been really blown away by. &#xA0;The results it gives are truly amazing. It is creative in ways that most people aren&apos;t. It forms relationships between objects, people, themes, and narratives that are enjoyable to work with. &#xA0;It also works very well for straight forward NLP tasks, such as picking out characteristics from an input sentence.</p><h3 id="they-allow-for-fine-tuning">They allow for fine-tuning</h3><p>OpenAI has already started giving people access to fine-tune specific engines for more accurate results. This has the potential to drastically improve response quality and latency. It is worth noting that they do not allow for fine-tuning of the largest &apos;davinci&apos; model, so it will be important to have sufficient input data to allow for the smaller models to make accurate predictions. There is an additional waitlist for this, even for the limited number of people that already have beta access. It will be many months before this will be commonly available (via slack).</p><h3 id="this-is-the-future-directionally-correct-">This is the future (&apos;directionally correct&apos;)</h3><p>Whether you are a founder, engineer, or investor it is clear that this type of abstract problem solving will only become more prevalent as machine learning advances. Regardless of the risks, history has shown that people front-running major trends** have been overall successful, if only in understanding the risks more clearly than others for future ventures.</p><p><em>**This is not to say that companies started now will succeed. It could very well be too early yet.</em></p><h3 id="it-works-perfectly-as-a-human-helper">It works perfectly as a human helper</h3><p>While anecdotal, I have started leaving a tab with the GPT-3 playground open, just to feed it things I come across during the day. It will easily summarize emails and small documents, answer questions, or just provide some entertaining ideas. This will be a serious use case for OpenAI to pursue.</p><h3 id="it-has-the-hype">It has the hype</h3><p>Hype is a real benefit. Everyone is talking about GPT-3, in a way that I have never seen in AI before. For press coverage or pitching technical users, &#xA0;is always easier to go with the flow than against it. Just be sure the product really uses GPT-3 rather than only slapping the label on.</p><h1 id="so-should-you-build-or-invest">So, should you build or invest?</h1><p>It is simply too early to bank an entire startup or product line on the usage of GPT-3. In order to be comfortable with adding it as a critical path you should think through:</p><ul><li>How much it will cost. What are your unit costs? Unless each API call returns $1 I would wait until they release the pricing details in early September to get serious.</li><li>Could end-users be exposed to toxic outputs? If the answer is yes, your startup will be a little harder. If nothing else, you will face increased scrutiny during the production review.</li></ul><p><strong>Exceptions</strong>:</p><ul><li>You are using it to build end-products, where one generation will lead to many sales. ex. Writing a book or movie script. Summarizing public documents. Creative works. </li><li>You are building infrastructure products to support future use-cases. There is plenty of green space for picks and shovels businesses here.</li></ul><p></p><h3 id="my-2-">My 2&#xA2;</h3><p></p><p><strong>Investors - </strong><br>You should be taking all the NLP/GPT-3 meetings they can. &#xA0;Learn about this stuff early rather than later.</p><p><strong>Founders - </strong><br>You should think about how to build a business that is enhanced by GPT-3, but not dependent upon it.<br><br>I am very bullish on the API overall. I think the OpenAI team is working quickly, very responsive to feedback, and has users best intentions at heart. For their very first product, this has been quite the launch! Even with the risks present, I am currently making a few different apps myself (both fall into the exceptions above). &#xA0;<br>Building something or investing in GPT-3? Feel free to contact me at tyler@honest-ai.com.</p>]]></content:encoded></item><item><title><![CDATA[Honest 01: Hello and Welcome!]]></title><description><![CDATA[<p>First off, <strong>thank you</strong> for being one of the very first to believe in Honest AI!</p><p>It has been a fast and furious few weeks here, but I have been happily surprised by the reception this project has received at every turn! I pushed four significant updates to the website</p>]]></description><link>https://honest-ai.com/newsletter-01/</link><guid isPermaLink="false">5f3f0afae83e633e34493a74</guid><category><![CDATA[Newsletter]]></category><dc:creator><![CDATA[Tyler Lastovich]]></dc:creator><pubDate>Mon, 24 Aug 2020 23:25:40 GMT</pubDate><media:content url="https://honest-ai.com/content/images/2020/08/newsletter01.png" medium="image"/><content:encoded><![CDATA[<img src="https://honest-ai.com/content/images/2020/08/newsletter01.png" alt="Honest 01: Hello and Welcome!"><p>First off, <strong>thank you</strong> for being one of the very first to believe in Honest AI!</p><p>It has been a fast and furious few weeks here, but I have been happily surprised by the reception this project has received at every turn! I pushed four significant updates to the website code this week and have few more ready to go &#x2013;this truly is just the beginning.</p><h1 id="the-community-is-almost-ready-">The community is almost ready!</h1><p>The centerpiece of Honest will be a community where technical and non-technical people work together to create positive applications with AI. There gap between what people read in the media and what really needs to be understood is getting wider, let&apos;s work to fix that!</p><p>This week involved creating the structure for the community using the brand new <a href="https://circle.so/?ref=honest-ai.com">Circle</a> platform. &#xA0;After a great deal of research I found this is clean piece of software to be the clear leader &#x2013; I hope you will really enjoy it.</p><p>This will be a paid community to keep the quality high, but as early believer you and anyone you refer in the near future gets <strong><em>free access for life!</em></strong> &#xA0;I will be sending out invites this week, so be on the lookout &#x1F440;.</p><h3 id="what-is-in-the-community">What is in the community?</h3><ul><li><strong>Show Us Something</strong> - Product Hunt for AI and demos. Promote your project! Posts are public and indexed by Google. &#x1F4A5;</li><li><strong>Introductions</strong> - Say hello and build connections in the AI space</li><li><strong>Ideas &amp; Honest Feedback</strong> - Have an idea? Post it here for clear, no-punches-held feedback.</li><li><strong>Jobs</strong> - A free to post place to find jobs in AI and ML.</li><li><strong>Creators Chat</strong> - Building something? Publish a worklog and get help when you need it.</li></ul><figure class="kg-card kg-image-card"><img src="https://honest-ai.com/content/images/2020/08/image-1.png" class="kg-image" alt="Honest 01: Hello and Welcome!" loading="lazy" width="2000" height="1029" srcset="https://honest-ai.com/content/images/size/w600/2020/08/image-1.png 600w, https://honest-ai.com/content/images/size/w1000/2020/08/image-1.png 1000w, https://honest-ai.com/content/images/size/w1600/2020/08/image-1.png 1600w, https://honest-ai.com/content/images/2020/08/image-1.png 2126w" sizes="(min-width: 720px) 720px"></figure><h1 id="demo-day">Demo Day</h1><p>Last Wednesday, Honest made its first public appearance when I was selected to present at the <a href="https://pioneer.app/gpt3?ref=honest-ai.com">Pioneer GPT-3 Demo Day</a>. Despite some jokes about creating a bot army, I am happy to say that my demo on synthetic personas was voted to second place by the audience! You can see a short example of the demo in action below. </p><figure class="kg-card kg-embed-card"><iframe width="612" height="344" src="https://www.youtube.com/embed/fy78gYxroAY?feature=oembed" frameborder="0" allow="accelerometer; autoplay; encrypted-media; gyroscope; picture-in-picture" allowfullscreen></iframe></figure><h1 id="gpt-3">GPT-3</h1><p>Are you tired of GPT-3 yet? I sure hope not! The last few months have be abuzz with talk of the language model from OpenAI. I expect GPT-3 to continue to dominate over every other AI conversation for a foreseeable future. With that in mind, I have a number of different demo apps in-progress to continue to push its limits and show off the real-world utility it can offer. I will be discussing these in-depth in the community, so if you are interested, &#xA0;join in. :)</p><p>OpenAI has been continually changing details behind the scenes, so I have chosen to make h<a href="https://honest-ai.com/gpt-3">onest-ai.com/gpt-3</a> a living document that will be updated regularly. Check back if you are ever wondering about specifics.</p><p>Last week OpenAI decided to extend the free beta period for the API until the first week of September and put a pause on taking new production applications until 8/31. They also are slowly opening up the capability to fine-tune your own model, with the wait time for access stated as up to &apos;a couple of months&apos;. &#xA0;The model has performed slower than usual as they continue to allow new users to access to the API. By the numbers, the OpenAI beta Slack group now counts just over a thousand members, 2x more than last month.</p><h1 id="you-can-help-">You can help!</h1><p>A community is only as strong as its members. If you like the concept of Honest AI, I would truly appreciate it if you shared it with someone.</p><p>Follow Honest&apos;s new profiles around the web:</p><p><a href="https://twitter.com/honest_ai?ref=honest-ai.com">Twitter</a><br><a href="https://www.youtube.com/channel/UCEVaxiDnYboWORN4-pavQiQ/?ref=honest-ai.com">Youtube</a><br><a href="https://www.linkedin.com/company/honestai?ref=honest-ai.com">LinkedIn</a><br><a href="https://instagram.com/honestai?ref=honest-ai.com">Instagram</a><br><br>Thank you,<br>-Tyler<br><br>PS: If you ever have any feedback just hit reply, I will respond to all comments.</p>]]></content:encoded></item><item><title><![CDATA[Synthetic Personas]]></title><description><![CDATA[The virtual presence, from 'Her' to Alexa has never been closer. Today we have nearly all the parts to generate a full synthetic identity.]]></description><link>https://honest-ai.com/synthetic-personas/</link><guid isPermaLink="false">5f30c4ade83e633e34493a4c</guid><category><![CDATA[Synthetic Personas]]></category><category><![CDATA[Article]]></category><dc:creator><![CDATA[Tyler Lastovich]]></dc:creator><pubDate>Mon, 10 Aug 2020 03:56:28 GMT</pubDate><media:content url="https://honest-ai.com/content/images/2020/08/synthetic-Personas-1.png" medium="image"/><content:encoded><![CDATA[<img src="https://honest-ai.com/content/images/2020/08/synthetic-Personas-1.png" alt="Synthetic Personas"><p>Can we really build virtual people from scratch?</p><p>Synthetic media is simply described as media that is not real. This can include CGI (computer generated imagery) characters we see in movies, virtual assistants we talk to, or content generated entirely by computers.</p><p>While this is not a new concept, the technology to produce it is rapidly advancing. From generating images using GANs (generative adversarial networks) to writing backstories and personality with GPT-3, the virtual identity is closing in on reality.</p><p><br><br></p><figure class="kg-card kg-embed-card"><iframe width="612" height="344" src="https://www.youtube.com/embed/fy78gYxroAY?feature=oembed" frameborder="0" allow="accelerometer; autoplay; encrypted-media; gyroscope; picture-in-picture" allowfullscreen></iframe></figure>]]></content:encoded></item><item><title><![CDATA[GPT-3]]></title><description><![CDATA[GPT-3 explained in common language. Learn why the OpenAI API has captured the minds of developers and business leaders the world over. ]]></description><link>https://honest-ai.com/gpt-3/</link><guid isPermaLink="false">5f29a91de83e633e344938ec</guid><category><![CDATA[GPT-3]]></category><category><![CDATA[Article]]></category><dc:creator><![CDATA[Tyler Lastovich]]></dc:creator><pubDate>Mon, 10 Aug 2020 03:42:43 GMT</pubDate><media:content url="https://honest-ai.com/content/images/2020/08/gpt3.png" medium="image"/><content:encoded><![CDATA[<h2 id="what-is-it">What is it?</h2><img src="https://honest-ai.com/content/images/2020/08/gpt3.png" alt="GPT-3"><p>GPT-3 is a machine learning language model created by OpenAI, a leader in artificial intelligence. &#xA0;In short, it is a system that has consumed enough text (nearly a trillion words) that it is able to make sense of text, and output text in a way that appears human-like. I use &apos;text&apos; here specifically, as GPT-3 itself has no intelligence &#x2013;it simply knows how to predict the next word (called a token) &#xA0;in a sentence, paragraph, or text block. It does this <em>exceedingly</em> <em>well</em>.</p><p>As an analogy you can think of GPT-3 like a freshly hired intern, who is well read, opinionated, and has a poor short-term memory. It is clever and offers fresh perspectives on how to solve problems, yet you don&apos;t really trust it to run your company or talk directly to customers.</p><p>GPT-3 works by taking a section of input text, and predicting the next section of text that should follow directly after. When hearing this people often compare it to autocorrect. The biggest difference being creativity. &#xA0;When you use GPT-3, you supply your input, and a few options. &#xA0;The most important of which is called temperature, the measure of how creative the outputs will be. Computers are not designed to be creative, so effectively this is option gives GPT-3 the freedom to questionable choices.</p><p>If you read everything there is to read, and stored how likely words are to appear together, in context, then you should be able to &apos;guess&apos; how a sentence or story will sound. This is hard to conceptualize because we as humans don&apos;t process information like this.</p><blockquote><em><em>GPT-3 is like a freshly-hired intern, who is well read, opinionated, and has a poor short-term memory</em></em></blockquote><p>&#x200C;<br>&#x200C;More specifically GPT-3 stands for &apos;Generative Pretrained Transformer 3&apos;, with transformer representing a type of machine learning model that deals with sequential data.</p><h3 id="why-is-fancy-autocorrect-interesting">Why is fancy autocorrect interesting?</h3><p>Well, it turns out that given enough input data, an AI like GPT-3 is able to repeatably perform non-trivial tasks. If you supply it well-structured input text you can get GPT-3 to respond very naturally, often appearing as if a person was generating the answers. &#xA0;This makes GPT-3 well suited for tasks such as creative writing, summarization, classification, and transactional messaging. </p><h2 id="how-can-i-use-it-for-my-business">How can I use it for my business?</h2><h3 id="access">Access</h3><p>Right now GPT-3 is only available through OpenAI&apos;s API product offering. They will likely never release the full model to the public as they did with previous versions (<a href="https://openai.com/blog/gpt-2-1-5b-release/?ref=honest-ai.com">GPT-2</a>). The OpenAI API is the companies first public product and is still relatively early in its development cycle. As of June 2020, OpenAI has opened up a waitlist to be invited to join the beta. They claims 10s of thousands of people have asked to be invited, so if you have a special use-case you will likely have to email a an OpenAI employee with a specific request. They have been slowly on-boarding people over the course of the last month.</p><blockquote><em><em><a href="https://forms.office.com/Pages/ResponsePage.aspx?id=VsqMpNrmTkioFJyEllK8s0v5E5gdyQhOuZCXNuMR8i1UQjFWVTVUVEpGNkg3U1FNRDVVRFg3U0w4Vi4u&amp;ref=honest-ai.com">Join the waitlist</a></em></em></blockquote><h3 id="product-offerings">Product offerings</h3><p>There are two types of API endpoints that companies can access, namely <em>completions</em> and <em>search</em>.</p><p>&#x200C;<br>&#x200C;<strong><strong>Completions</strong></strong>: The headline product, a completion lets GPT-3 take in an input prompt (which we will cover in detail below), and complete that to return a result. &#x200C;<br>&#x200C;&#x200C;<br>&#x200C;<strong><strong>Search</strong></strong>: GPT-3 is very competent at parsing natural language inputs, making it ideal for building search tools. &#x200C;<br>&#x200C;&#x200C;<br>&#x200C;<strong><strong>Models: </strong></strong>OpenAI currently offers 4 models (ada, babbage, curie, davinci), each of a different size. The flagship model is &apos;davinci&apos;, representing the 175 billion parameters touted in the media. Model size directly correlates to how fast the API calls are handled, so for time-sensitive requests davinci will likely be too slow. See &apos;Speed&apos; below for more.</p><p><strong><strong>Fine-tuning</strong></strong>: OpenAI will offer to train a model specifically for you, based on a dataset that you supply. &#xA0;Fine-tuning will be ideal for companies to produce extremely accurate outputs for tasks with more than a few dozen input prompts, or that require a very specific output structure. Read more in Fine-tuning below.</p><h3 id="playground">Playground</h3><p>Beyond just making calls using code to the API, OpenAI also offers a user interface called the playground. &#xA0;This lets developers quickly test ideas and refine input prompts before committing to create an application. It is easy to foresee that some variant of this playground could be made in a public-facing manner for anyone to get help with smaller problems.</p><h3 id="pricing">Pricing</h3><p>This API is currently in beta testing stage and was free to use until October. Starting 10/1 the following prices are in place:</p><p></p><figure class="kg-card kg-image-card kg-width-wide"><img src="https://honest-ai.com/content/images/2020/09/image-1.png" class="kg-image" alt="GPT-3" loading="lazy" width="2000" height="1391" srcset="https://honest-ai.com/content/images/size/w600/2020/09/image-1.png 600w, https://honest-ai.com/content/images/size/w1000/2020/09/image-1.png 1000w, https://honest-ai.com/content/images/size/w1600/2020/09/image-1.png 1600w, https://honest-ai.com/content/images/2020/09/image-1.png 2266w" sizes="(min-width: 1200px) 1200px"></figure><p><br>While there is a free &apos;Explore&apos; tier offered, it exists more as an interest builder than a useful allotment of tokens. It is quite easy to use up 100k tokens in an afternoon. Each engine varies greatly in size, and this variance is reflected in the &apos;token per credit system&apos;. The smaller models are significantly more economical to run, though they will lack the serendipity and creativity of the largest Davinci engine. <br><br>Assuming 1,000 tokens used per request (prompt+output) a subscriber on the create tier would be allowed to use: </p><p>~$0.05/call for davinci (2,000)<br>~$0.005/call for curie (20,000)<br>~$0.001/call for babbage (100,000)<br>~$0.0007/call for ada (150,000)</p><p> &#x200C;<br>It is important to understand that GPT-3 has no direct competition for creative output, meaning they can easily price the API as they wish going forward.</p><h2 id="limitations-and-warnings">Limitations and Warnings</h2><h3 id="stability">Stability</h3><p>The OpenAI API is very much still a beta product, and has already had a few periods of unplanned downtime since I have used it. It is not yet suitable for critical business use cases. The OpenAI team has been communicative and very fast to respond to these outages, so I have no doubt they will be ironed out before a general release happens.</p><h3 id="speed">Speed</h3><p>In my usage the latency of the API has been poor to OK. It is certainly one of the slowest APIs I have recently used. This is likely one of the main areas OpenAI will need to improve on before going live to a large audience. Serving a model the size of <em>davinci</em> (~350gb) efficiently is a non-trivial problem!</p><h3 id="content-moderation">Content Moderation</h3><p>As it stands API calls are made to the &apos;raw model&apos;. This means that the outputs are unfiltered and may contain explicit, hateful, or racist content. &apos;davinci&apos; was trained using <a href="https://commoncrawl.org/?ref=honest-ai.com">Common Crawl</a>, a service that records the content of web pages -including reddit, forums, and other</p><p>OpenAI is upfront with this and clearly states that applications should not let the user interact directly with the raw model. They currently require applications to go through an approval process before going into production.</p><h3 id="limited-inputs-outputs">Limited Inputs &amp; Outputs</h3><p>Currently the OpenAI API only allows for an input prompt to be 1000 tokens long (or around 500 words). &#xA0;This means that you must fit all of the context you are supplying in that prompt, quite a challenge.</p><p><br></p><h2 id="should-i-build-a-gpt-3-based-company">Should I build a GPT-3-based company?</h2><p>tl;dr &#x2013; <em>Maybe</em>. There are some specific concerns to be aware of though.</p><p>Almost everyone I have talked to who plays with GPT-3 has that <em>ah-ha! </em>moment where they think of a great project to use it on. The play&#x200C;</p><h4 id="you-are-taking-a-significant-platform-risk-"><em>You are taking a significant platform risk.</em></h4><ul><li>OpenAI owns the API and they are free to do whatever they want with it, regardless of how that impacts your business.</li><li>GPT-3 has no direct competitors (yet). This gives OpenAI incredible pricing power, access control, and the ability to add/enforce policies.</li><li>Your service is tied to GPT-3 working correctly. GPT-3 is a very large ML model, that requires tangible compute resources to operate, which means it has the potential to run into scaling issues. (as alerady seen during the beta period)</li><li>The speed of your application will be directly dependent on the OpenAI API.</li></ul><h4 id="differentiation-is-hard"><em>Differentiation is hard</em></h4><ul><li>GPT-3 is a few-shot model, therefore it is relatively trivial to reverse-engineer an input prompt. (you can read more in <a href="https://arxiv.org/pdf/2005.14165.pdf?ref=honest-ai.com">the scientific paper</a>)</li><li>Be careful that your app itself provides value, and doesn&apos;t simply let the OpenAI API do all the &apos;real work&apos;.</li></ul><h2 id="garbage-in-garbage-out">Garbage in &gt;&gt; Garbage out</h2><p>In working with GPT-3 one thing becomes very clear, the inputs you feed it matter &#x2013;a lot! As a few-shot model, it is extrapolating what it thinks the right response is exclusively from the up to 1000 tokens of input you provide.</p><h2 id="where-do-language-models-go-from-here">Where do language models go from here?</h2><p>This OpenAI is still an early implementation of AI. &#xA0;As compute power and algorithm optimization continue to increase we will be able to effectively train much larger models. Based on estimates, GPT-3 cost roughly $4.6M worth of compute time to train.&#x200C;<br>&#x200C;&#x200C;<br>&#x200C;GPT-2: 1.5 billion parameters&#x200C;<br>&#x200C;GPT-3: 175 billion parameters&#x200C;<br>&#x200C;GPT-4: ???&#x200C;<br>&#x200C;Human brain: 100 trillion synapses (analygous to parameter)&#x200C;</p><figure class="kg-card kg-embed-card kg-card-hascaption"><iframe width="612" height="344" src="https://www.youtube.com/embed/kpiY_LemaTc?feature=oembed" frameborder="0" allow="accelerometer; autoplay; encrypted-media; gyroscope; picture-in-picture" allowfullscreen></iframe><figcaption>A short little comparison between GPT-3 and the brain by Lex Fridman.</figcaption></figure>]]></content:encoded></item></channel></rss>