{"version": "https://jsonfeed.org/version/1.1", "title": "Sam Altman", "feed_url": "https://blog.samaltman.com", "home_page_url": "54", "description": null, "icon": null, "favicon": null, "authors": [{"name": "Sam Altman", "url": null, "avatar": null}], "language": "en-US", "expired": null, "hub": null, "items": [{"id": "tag:blog.samaltman.com,2013:Post/2279512", "url": "https://blog.samaltman.com/2279512", "external_url": null, "title": "-", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>Here is a photo of my family. I love them more than anything.<br /><br />        <div class=\"posthaven-gallery\" id=\"posthaven_gallery[2305120]\">\n                  <p class=\"posthaven-file posthaven-file-image posthaven-file-state-processed\">\n          <img class=\"posthaven-gallery-image\" src=\"https://phaven-prod.s3.amazonaws.com/files/image_part/asset/3439526/q4Xtf4iMYA_SCrMc8hMfnF_k3UE/medium_WhatsApp_Image_2026-02-28_at_15.26.14__4_.jpeg\" />\n        </p>\n\n        </div>\n</p><p>Images have power, I hope. Normally we try to be pretty private, but in this case I am sharing a photo in the hopes that it might dissuade the next person from throwing a Molotov cocktail at our house, no matter what they think about me.</p><p>The first person did it last night, at 3:45 am in the morning. Thankfully it bounced off the house and no one got hurt.</p><p>Words have power too. There was an incendiary article about me a few days ago. Someone said to me yesterday they thought it was coming at a time of great anxiety about AI and that it made things more dangerous for me. I brushed it aside.</p><p>Now I am awake in the middle of the night and pissed, and thinking that I have underestimated the power of words and narratives. This seems like as good of a time as any to address a few things.\u00a0</p><p><b>First</b>, what I believe.</p><p>*Working towards prosperity for everyone, empowering all people, and advancing science and technology are moral obligations for me.</p><p>*AI will be the most powerful tool for expanding human capability and potential that anyone has ever seen. Demand for this tool will be essentially uncapped, and people will do incredible things with it. The world deserves huge amounts of AI and we must figure out how to make it happen.</p><p>*It will not all go well. The fear and anxiety about AI is justified; we are in the process of witnessing the largest change to society in a long time, and perhaps ever. We have to get safety right, which is not just about aligning a model\u2014we urgently need a society-wide response to be resilient to new threats. This includes things like new policy to help navigate through a difficult economic transition in order to get to a much better future.</p><p>\u00a0*AI has to be democratized; power cannot be too concentrated. Control of the future belongs to all people and their institutions. AI needs to empower people individually, and we need to make decisions about our future and the new rules collectively. I do not think it is right that a few AI labs would make the most consequential decisions about the shape of our future.</p><p>*Adaptability is critical. We are all learning about something new very quickly; some of our beliefs will be right and some will be wrong, and sometimes we will need to change our mind quickly as the technology develops and society evolves. No one understands the impacts of superintelligence yet, but they will be immense.</p><p>\u00a0</p><p><b>Second</b>, some personal reflections.</p><p>As I reflect on my own work in the first decade of OpenAI, I can point to a lot of things I\u2019m proud of and a bunch of mistakes.</p><p>I was thinking about our upcoming trial with Elon and remembering how much I held the line on not being willing to agree to the unilateral control he wanted over OpenAI. I\u2019m proud of that, and the narrow path we navigated then to allow the continued existence of OpenAI, and all the achievements that followed.</p><p>I am not proud of being conflict-averse, which has caused great pain for me and OpenAI. I am not proud of handling myself badly in a conflict with our previous board that led to a huge mess for the company. I have made many other mistakes throughout the insane trajectory of OpenAI; I am a flawed person in the center of an exceptionally complex situation, trying to get a little better each year, always working for the mission. We knew going into this how huge the stakes of AI were, and that the personal disagreements between well-meaning people I cared about would be amplified greatly. But it\u2019s another thing to live through these bitter conflicts and often to have to arbitrate them, and the costs have been serious. I am sorry to people I\u2019ve hurt and wish I had learned more faster.</p><p>I am also very aware that OpenAI is now a major platform, not a scrappy startup, and we need to operate in a more predictable way now. It has been an extremely intense, chaotic, and high-pressure few years.\u00a0</p><p>Mostly though, I am extremely proud that we are delivering on our mission, which seemed incredibly unlikely when we started. Against all odds, we figured out how to build very powerful AI, figured out how to amass enough capital to build the infrastructure to deliver it, figured out how to build a product company and business, figured out how to deliver reasonably safe and robust services at a massive scale, and much more.<br /><br />A lot of companies say they are going to change the world; we actually did.</p><p><b><br /></b></p><p><b>Third</b>, some thoughts about the industry.</p><p>My personal takeaway from the last several years, and take on why there has been so much Shakespearean drama between the companies in our field, comes down to this: \u201cOnce you see AGI you can\u2019t unsee it.\u201d It has a real \"ring of power\u201d dynamic to it, and makes people do crazy things. I don\u2019t mean that AGI is the ring itself, but instead the totalizing philosophy of \u201cbeing the one to control AGI\u201d.</p><p>The only solution I can come up with is to orient towards sharing the technology with people broadly, and for no one to have the ring. The two obvious ways to do this are individual empowerment and making sure democratic system stays in control.</p><p>It is important that the democratic process remains more powerful than companies. Laws and norms are going to change, but we have to work within the democratic process, even though it will be messy and slower than we\u2019d like. We want to be a voice and a stakeholder, but not to have all the power.</p><p>A lot of the criticism of our industry comes from sincere concern about the incredibly high stakes of this technology. This is quite valid, and we welcome good-faith criticism and debate. I empathize with anti-technology sentiments and clearly technology isn\u2019t always good for everyone. But overall, I believe technological progress can make the future unbelievably good, for your family and mine.\u00a0</p><p></p><p>While we have that debate, we should de-escalate the rhetoric and tactics and try to have fewer explosions in fewer homes, figuratively and literally.</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>Here is a photo of my family. I love them more than anything.<br /><br />        <div class=\"posthaven-gallery\" id=\"posthaven_gallery[2305120]\">\n                  <p class=\"posthaven-file posthaven-file-image posthaven-file-state-processed\">\n          <img class=\"posthaven-gallery-image\" src=\"https://phaven-prod.s3.amazonaws.com/files/image_part/asset/3439526/q4Xtf4iMYA_SCrMc8hMfnF_k3UE/medium_WhatsApp_Image_2026-02-28_at_15.26.14__4_.jpeg\" />\n        </p>\n\n        </div>\n</p><p>Images have power, I hope. Normally we try to be pretty private, but in this case I am sharing a photo in the hopes that it might dissuade the next person from throwing a Molotov cocktail at our house, no matter what they think about me.</p><p>The first person did it last night, at 3:45 am in the morning. Thankfully it bounced off the house and no one got hurt.</p><p>Words have power too. There was an incendiary article about me a few days ago. Someone said to me yesterday they thought it was coming at a time of great anxiety about AI and that it made things more dangerous for me. I brushed it aside.</p><p>Now I am awake in the middle of the night and pissed, and thinking that I have underestimated the power of words and narratives. This seems like as good of a time as any to address a few things.\u00a0</p><p><b>First</b>, what I believe.</p><p>*Working towards prosperity for everyone, empowering all people, and advancing science and technology are moral obligations for me.</p><p>*AI will be the most powerful tool for expanding human capability and potential that anyone has ever seen. Demand for this tool will be essentially uncapped, and people will do incredible things with it. The world deserves huge amounts of AI and we must figure out how to make it happen.</p><p>*It will not all go well. The fear and anxiety about AI is justified; we are in the process of witnessing the largest change to society in a long time, and perhaps ever. We have to get safety right, which is not just about aligning a model\u2014we urgently need a society-wide response to be resilient to new threats. This includes things like new policy to help navigate through a difficult economic transition in order to get to a much better future.</p><p>\u00a0*AI has to be democratized; power cannot be too concentrated. Control of the future belongs to all people and their institutions. AI needs to empower people individually, and we need to make decisions about our future and the new rules collectively. I do not think it is right that a few AI labs would make the most consequential decisions about the shape of our future.</p><p>*Adaptability is critical. We are all learning about something new very quickly; some of our beliefs will be right and some will be wrong, and sometimes we will need to change our mind quickly as the technology develops and society evolves. No one understands the impacts of superintelligence yet, but they will be immense.</p><p>\u00a0</p><p><b>Second</b>, some personal reflections.</p><p>As I reflect on my own work in the first decade of OpenAI, I can point to a lot of things I\u2019m proud of and a bunch of mistakes.</p><p>I was thinking about our upcoming trial with Elon and remembering how much I held the line on not being willing to agree to the unilateral control he wanted over OpenAI. I\u2019m proud of that, and the narrow path we navigated then to allow the continued existence of OpenAI, and all the achievements that followed.</p><p>I am not proud of being conflict-averse, which has caused great pain for me and OpenAI. I am not proud of handling myself badly in a conflict with our previous board that led to a huge mess for the company. I have made many other mistakes throughout the insane trajectory of OpenAI; I am a flawed person in the center of an exceptionally complex situation, trying to get a little better each year, always working for the mission. We knew going into this how huge the stakes of AI were, and that the personal disagreements between well-meaning people I cared about would be amplified greatly. But it\u2019s another thing to live through these bitter conflicts and often to have to arbitrate them, and the costs have been serious. I am sorry to people I\u2019ve hurt and wish I had learned more faster.</p><p>I am also very aware that OpenAI is now a major platform, not a scrappy startup, and we need to operate in a more predictable way now. It has been an extremely intense, chaotic, and high-pressure few years.\u00a0</p><p>Mostly though, I am extremely proud that we are delivering on our mission, which seemed incredibly unlikely when we started. Against all odds, we figured out how to build very powerful AI, figured out how to amass enough capital to build the infrastructure to deliver it, figured out how to build a product company and business, figured out how to deliver reasonably safe and robust services at a massive scale, and much more.<br /><br />A lot of companies say they are going to change the world; we actually did.</p><p><b><br /></b></p><p><b>Third</b>, some thoughts about the industry.</p><p>My personal takeaway from the last several years, and take on why there has been so much Shakespearean drama between the companies in our field, comes down to this: \u201cOnce you see AGI you can\u2019t unsee it.\u201d It has a real \"ring of power\u201d dynamic to it, and makes people do crazy things. I don\u2019t mean that AGI is the ring itself, but instead the totalizing philosophy of \u201cbeing the one to control AGI\u201d.</p><p>The only solution I can come up with is to orient towards sharing the technology with people broadly, and for no one to have the ring. The two obvious ways to do this are individual empowerment and making sure democratic system stays in control.</p><p>It is important that the democratic process remains more powerful than companies. Laws and norms are going to change, but we have to work within the democratic process, even though it will be messy and slower than we\u2019d like. We want to be a voice and a stakeholder, but not to have all the power.</p><p>A lot of the criticism of our industry comes from sincere concern about the incredibly high stakes of this technology. This is quite valid, and we welcome good-faith criticism and debate. I empathize with anti-technology sentiments and clearly technology isn\u2019t always good for everyone. But overall, I believe technological progress can make the future unbelievably good, for your family and mine.\u00a0</p><p></p><p>While we have that debate, we should de-escalate the rhetoric and tactics and try to have fewer explosions in fewer homes, figuratively and literally.</p></div>", "image": null, "banner_image": null, "date_published": "2026-04-10T14:55:13Z", "date_modified": "2026-04-11T05:39:26Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2228126", "url": "https://blog.samaltman.com/sora-update-number-1", "external_url": null, "title": "Sora update #1", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>We have been learning quickly from how people are using Sora and taking feedback from users, rightsholders, and other interested groups. We of course spent a lot of time discussing this before launch, but now that we have a product out we can do more than just theorize.<br /><br />We are going to make two changes soon (and many more to come).<br /><br />First, we will give rightsholders more granular control over generation of characters, similar to the opt-in model for likeness but with additional controls.</p><p>We are hearing from a lot of rightsholders who are very excited for this new kind of \"interactive fan fiction\" and think this new kind of engagement will accrue a lot of value to them, but want the ability to specify how their characters can be used (including not at all). We assume different people will try very different approaches and will figure out what works for them. But we want to apply the same standard towards everyone, and let rightsholders decide how to proceed (our aim of course is to make it so compelling that many people want to). There may be some edge cases of generations that get through that shouldn't, and getting our stack to work well will take some iteration.\u00a0<br /><br />In particular, we'd like to acknowledge the remarkable creative output of Japan--we are struck by how deep the connection between users and Japanese content is!<br /><br />Second, we are going to have to somehow make money for video generation. People are generating much more than we expected per user, and a lot of videos are being generated for very small audiences. We are going to try sharing some of this revenue with rightsholders who want their characters generated by users. The exact model will take some trial and error to figure out, but we plan to start very soon. Our hope is that the new kind of engagement is even more valuable than the revenue share, but of course we we want both to be valuable.<br /></p><p>Please expect a very high rate of change from us; it reminds me of the early days of ChatGPT. We will make some good decisions and some missteps, but we will take feedback and try to fix the missteps very quickly. We plan to do our iteration on different approaches in Sora, but then apply it consistently across our products.<br /></p></div>", "summary": "<div class=\"posthaven-post-body\"><p>We have been learning quickly from how people are using Sora and taking feedback from users, rightsholders, and other interested groups. We of course spent a lot of time discussing this before launch, but now that we have a product out we can do more than just theorize.<br /><br />We are going to make two changes soon (and many more to come).<br /><br />First, we will give rightsholders more granular control over generation of characters, similar to the opt-in model for likeness but with additional controls.</p><p>We are hearing from a lot of rightsholders who are very excited for this new kind of \"interactive fan fiction\" and think this new kind of engagement will accrue a lot of value to them, but want the ability to specify how their characters can be used (including not at all). We assume different people will try very different approaches and will figure out what works for them. But we want to apply the same standard towards everyone, and let rightsholders decide how to proceed (our aim of course is to make it so compelling that many people want to). There may be some edge cases of generations that get through that shouldn't, and getting our stack to work well will take some iteration.\u00a0<br /><br />In particular, we'd like to acknowledge the remarkable creative output of Japan--we are struck by how deep the connection between users and Japanese content is!<br /><br />Second, we are going to have to somehow make money for video generation. People are generating much more than we expected per user, and a lot of videos are being generated for very small audiences. We are going to try sharing some of this revenue with rightsholders who want their characters generated by users. The exact model will take some trial and error to figure out, but we plan to start very soon. Our hope is that the new kind of engagement is even more valuable than the revenue share, but of course we we want both to be valuable.<br /></p><p>Please expect a very high rate of change from us; it reminds me of the early days of ChatGPT. We will make some good decisions and some missteps, but we will take feedback and try to fix the missteps very quickly. We plan to do our iteration on different approaches in Sora, but then apply it consistently across our products.<br /></p></div>", "image": null, "banner_image": null, "date_published": "2025-10-03T16:37:50Z", "date_modified": "2026-04-11T05:30:06Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2227392", "url": "https://blog.samaltman.com/sora-2", "external_url": null, "title": "Sora 2", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>We are launching a new app called Sora. This is a combination of a new model called Sora 2, and a new product that makes it easy to create, share, and view videos.</p><p>This feels to many of us like the \u201cChatGPT for creativity\u201d moment, and it feels fun and new. There is something great about making it really easy and fast to go from idea to result, and the new social dynamics that emerge.</p><p>Creativity could be about to go through a Cambrian explosion, and along with it, the quality of art and entertainment can drastically increase. Even in the very early days of playing with Sora, it\u2019s been striking to many of us how open the playing field suddenly feels.</p><p>In particular, the ability to put yourself and your friends into a video\u2014the team worked very hard on character consistency\u2014with the cameo feature is something we have really enjoyed during testing, and is to many of us a surprisingly compelling new way to connect.</p><p>We also feel some trepidation. Social media has had some good effects on the world, but it\u2019s also had some bad ones. We are aware of how addictive a service like this could become, and we can imagine many ways it could be used for bullying.</p><p>It is easy to imagine the degenerate case of AI video generation that ends up with us all being sucked into an RL-optimized slop feed. The team has put great care and thought into trying to figure out how to make a delightful product that doesn\u2019t fall into that trap, and has come up with a number of promising ideas. We will experiment in the early days of the product with different approaches.</p><p><span>In addition to the mitigations we have already put in place (which include things like mitigations to prevent someone from misusing someone\u2019s likeness in deepfakes, safeguards for disturbing or illegal content, periodic checks on how Sora is impacting users\u2019 mood and wellbeing, and more) we are sure we will discover new things we need to do if Sora becomes very successful. To help guide us towards more of the good and less of the bad, here are some principles we have for this product:</span></p><p><br /></p><p><b>*Optimize for long-term user satisfaction.\u00a0</b>The majority of users, looking back on the past 6 months, should feel that their life is better for using Sora that it would have been if they hadn\u2019t. If that\u2019s not the case, we will make significant changes (and if we can\u2019t fix it, we would discontinue offering the service). \u00a0</p><p>*<b>Encourage users to control their feed.</b>\u00a0You should be able to tell Sora what you want\u2014do you want to see videos that will make you more relaxed, or more energized? Or only videos that fit a specific interest? Or only for a certain about of time? Eventually as our technology progresses, you will be should to the tell Sora what you want in detail in natural language. (However, parental controls for teens include the ability to opt out of a personalized feed, and other things like turning off DMs.)</p><p>*<b>Prioritize creation.</b>\u00a0We want to make it easy and rewarding for everyone to participate in the creation process; we believe people are natural-born creators, and creating is important to our satisfaction.</p><p>*<b>Help users achieve their long-term goals.\u00a0</b>We want to understand a user\u2019s true goals, and help them achieve them. If you want to be more connected to your friends, we will try to help you with that. If you want to get fit, we can show you fitness content that will motivate you. If you want to start a business, we want to help teach you the skills you need. And if you truly just want to doom scroll and be angry, then ok, we\u2019ll help you with that (although we want users to spend time using the app if they think it\u2019s time well spent, we don\u2019t want to be paternalistic about what that means to them).<br /></p></div>", "summary": "<div class=\"posthaven-post-body\"><p>We are launching a new app called Sora. This is a combination of a new model called Sora 2, and a new product that makes it easy to create, share, and view videos.</p><p>This feels to many of us like the \u201cChatGPT for creativity\u201d moment, and it feels fun and new. There is something great about making it really easy and fast to go from idea to result, and the new social dynamics that emerge.</p><p>Creativity could be about to go through a Cambrian explosion, and along with it, the quality of art and entertainment can drastically increase. Even in the very early days of playing with Sora, it\u2019s been striking to many of us how open the playing field suddenly feels.</p><p>In particular, the ability to put yourself and your friends into a video\u2014the team worked very hard on character consistency\u2014with the cameo feature is something we have really enjoyed during testing, and is to many of us a surprisingly compelling new way to connect.</p><p>We also feel some trepidation. Social media has had some good effects on the world, but it\u2019s also had some bad ones. We are aware of how addictive a service like this could become, and we can imagine many ways it could be used for bullying.</p><p>It is easy to imagine the degenerate case of AI video generation that ends up with us all being sucked into an RL-optimized slop feed. The team has put great care and thought into trying to figure out how to make a delightful product that doesn\u2019t fall into that trap, and has come up with a number of promising ideas. We will experiment in the early days of the product with different approaches.</p><p><span>In addition to the mitigations we have already put in place (which include things like mitigations to prevent someone from misusing someone\u2019s likeness in deepfakes, safeguards for disturbing or illegal content, periodic checks on how Sora is impacting users\u2019 mood and wellbeing, and more) we are sure we will discover new things we need to do if Sora becomes very successful. To help guide us towards more of the good and less of the bad, here are some principles we have for this product:</span></p><p><br /></p><p><b>*Optimize for long-term user satisfaction.\u00a0</b>The majority of users, looking back on the past 6 months, should feel that their life is better for using Sora that it would have been if they hadn\u2019t. If that\u2019s not the case, we will make significant changes (and if we can\u2019t fix it, we would discontinue offering the service). \u00a0</p><p>*<b>Encourage users to control their feed.</b>\u00a0You should be able to tell Sora what you want\u2014do you want to see videos that will make you more relaxed, or more energized? Or only videos that fit a specific interest? Or only for a certain about of time? Eventually as our technology progresses, you will be should to the tell Sora what you want in detail in natural language. (However, parental controls for teens include the ability to opt out of a personalized feed, and other things like turning off DMs.)</p><p>*<b>Prioritize creation.</b>\u00a0We want to make it easy and rewarding for everyone to participate in the creation process; we believe people are natural-born creators, and creating is important to our satisfaction.</p><p>*<b>Help users achieve their long-term goals.\u00a0</b>We want to understand a user\u2019s true goals, and help them achieve them. If you want to be more connected to your friends, we will try to help you with that. If you want to get fit, we can show you fitness content that will motivate you. If you want to start a business, we want to help teach you the skills you need. And if you truly just want to doom scroll and be angry, then ok, we\u2019ll help you with that (although we want users to spend time using the app if they think it\u2019s time well spent, we don\u2019t want to be paternalistic about what that means to them).<br /></p></div>", "image": null, "banner_image": null, "date_published": "2025-09-30T09:13:47Z", "date_modified": "2026-04-11T05:27:00Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2225757", "url": "https://blog.samaltman.com/abundant-intelligence", "external_url": null, "title": "Abundant Intelligence", "content_text": "", "content_html": "<div class=\"posthaven-post-body\">Growth in the use of AI services has been astonishing; we expect it to be even more astonishing going forward.<br /><br />As AI gets smarter, access to AI will be a fundamental driver of the economy, and maybe eventually something we consider a fundamental human right. Almost everyone will want more AI working on their behalf.<br /><br />To be able to deliver what the world needs\u2014for inference compute to run these models, and for training compute to keep making them better and better\u2014we are putting the groundwork in place to be able to significantly expand our ambitions for building out AI infrastructure.<br /><br />If AI stays on the trajectory that we think it will, then amazing things will be possible. Maybe with 10 gigawatts of compute, AI can figure out how to cure cancer. Or with 10 gigawatts of compute, AI can figure out how to provide customized tutoring to every student on earth. If we are limited by compute, we\u2019ll have to choose which one to prioritize; no one wants to make that choice, so let\u2019s go build.<br /><br />Our vision is simple: we want to create a factory that can produce a gigawatt of new AI infrastructure every week. The execution of this will be extremely difficult; it will take us years to get to this milestone and it will require innovation at every level of the stack, from chips to power to building to robotics. But we have been hard at work on this and believe it is possible. In our opinion, it will be the coolest and most important infrastructure project ever. We are particularly excited to build a lot of this in the US; right now, other countries are building things like chips fabs and new energy production much faster than we are, and we want to help turn that tide.<p>Over the next couple of months, we\u2019ll be talking about some of our plans and the partners we are working with to make this a reality. Later this year, we\u2019ll talk about how we are financing it; given how increasing compute is the literal key to increasing revenue, we have some interesting new ideas.<br /></p></div>", "summary": "<div class=\"posthaven-post-body\">Growth in the use of AI services has been astonishing; we expect it to be even more astonishing going forward.<br /><br />As AI gets smarter, access to AI will be a fundamental driver of the economy, and maybe eventually something we consider a fundamental human right. Almost everyone will want more AI working on their behalf.<br /><br />To be able to deliver what the world needs\u2014for inference compute to run these models, and for training compute to keep making them better and better\u2014we are putting the groundwork in place to be able to significantly expand our ambitions for building out AI infrastructure.<br /><br />If AI stays on the trajectory that we think it will, then amazing things will be possible. Maybe with 10 gigawatts of compute, AI can figure out how to cure cancer. Or with 10 gigawatts of compute, AI can figure out how to provide customized tutoring to every student on earth. If we are limited by compute, we\u2019ll have to choose which one to prioritize; no one wants to make that choice, so let\u2019s go build.<br /><br />Our vision is simple: we want to create a factory that can produce a gigawatt of new AI infrastructure every week. The execution of this will be extremely difficult; it will take us years to get to this milestone and it will require innovation at every level of the stack, from chips to power to building to robotics. But we have been hard at work on this and believe it is possible. In our opinion, it will be the coolest and most important infrastructure project ever. We are particularly excited to build a lot of this in the US; right now, other countries are building things like chips fabs and new energy production much faster than we are, and we want to help turn that tide.<p>Over the next couple of months, we\u2019ll be talking about some of our plans and the partners we are working with to make this a reality. Later this year, we\u2019ll talk about how we are financing it; given how increasing compute is the literal key to increasing revenue, we have some interesting new ideas.<br /></p></div>", "image": null, "banner_image": null, "date_published": "2025-09-23T05:41:02Z", "date_modified": "2026-04-11T01:17:25Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2222747", "url": "https://blog.samaltman.com/jakub-and-szymon", "external_url": null, "title": "Jakub and Szymon", "content_text": "", "content_html": "<div class=\"posthaven-post-body\">AI has gotten remarkably better in recent years; ChatGPT can do amazing things that we take for granted. This is as it should be, and is the story of human progress. But behind the blinking circle, nicely abstracted away, is the greatest story of human ingenuity I have ever seen. A lot of people have worked unbelievably hard to discover how to build something that most experts thought was impossible on this timeframe, and to build a company to deliver products at massive scale to let people benefit from it. Most people who use ChatGPT will never think about the people that put so much work into it, which is totally ok, but just to take a minute of your time\u2026<br /><br />There are two people I'd like to mention that OpenAI would not be OpenAI without: Jakub Pachocki and Szymon Sidor. Time and again, they combine research and engineering to solve impossible problems. They have not gotten enough public credit, but they decided to scale up RL as a baseline to see where it broke when the conventional wisdom was that it didn't scale which led to our Dota result, built much of the infrastructure that enabled a lot of our scientific discoveries, led GPT-4 pretraining, drove together with Ilya and Lukasz the initial ideas that led to the reasoning breakthrough, and have made significant progress exploring new paradigms.<p>Jakub is our chief scientist. He once described Szymon as \u201cindefatigable\u201d, which is as perfect of a use of that word as I have ever heard. OpenAI has not yet thrown a problem at them they have not been able to solve; I have heard about partnerships like there is research labs of the past where two people are able to complement each other so well, but it is very special to get to watch it unfold over the years.<br /></p></div>", "summary": "<div class=\"posthaven-post-body\">AI has gotten remarkably better in recent years; ChatGPT can do amazing things that we take for granted. This is as it should be, and is the story of human progress. But behind the blinking circle, nicely abstracted away, is the greatest story of human ingenuity I have ever seen. A lot of people have worked unbelievably hard to discover how to build something that most experts thought was impossible on this timeframe, and to build a company to deliver products at massive scale to let people benefit from it. Most people who use ChatGPT will never think about the people that put so much work into it, which is totally ok, but just to take a minute of your time\u2026<br /><br />There are two people I'd like to mention that OpenAI would not be OpenAI without: Jakub Pachocki and Szymon Sidor. Time and again, they combine research and engineering to solve impossible problems. They have not gotten enough public credit, but they decided to scale up RL as a baseline to see where it broke when the conventional wisdom was that it didn't scale which led to our Dota result, built much of the infrastructure that enabled a lot of our scientific discoveries, led GPT-4 pretraining, drove together with Ilya and Lukasz the initial ideas that led to the reasoning breakthrough, and have made significant progress exploring new paradigms.<p>Jakub is our chief scientist. He once described Szymon as \u201cindefatigable\u201d, which is as perfect of a use of that word as I have ever heard. OpenAI has not yet thrown a problem at them they have not been able to solve; I have heard about partnerships like there is research labs of the past where two people are able to complement each other so well, but it is very special to get to watch it unfold over the years.<br /></p></div>", "image": null, "banner_image": null, "date_published": "2025-09-08T16:10:22Z", "date_modified": "2026-04-07T11:01:46Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2202875", "url": "https://blog.samaltman.com/the-gentle-singularity", "external_url": null, "title": "The Gentle Singularity", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>We are past the event horizon; the takeoff has started. Humanity is close to building digital superintelligence, and at least so far it\u2019s much less weird than it seems like it should be.</p><p>Robots are not yet walking the streets, nor are most of us talking to AI all day. People still die of disease, we still can\u2019t easily go to space, and there is a lot about the universe we don\u2019t understand.</p><p>And yet, we have recently built systems that are smarter than people in many ways, and are able to significantly amplify the output of people using them. The least-likely part of the work is behind us; the scientific insights that got us to systems like GPT-4 and o3 were hard-won, but will take us very far.</p><p>AI will contribute to the world in many ways, but the gains to quality of life from AI driving faster scientific progress and increased productivity will be enormous; the future can be vastly better than the present. Scientific progress is the biggest driver of overall progress; it\u2019s hugely exciting to think about how much more we could have.</p><p>In some big sense, ChatGPT is already more powerful than any human who has ever lived. Hundreds of millions of people rely on it every day and for increasingly important tasks; a small new capability can create a hugely positive impact; a small misalignment multiplied by hundreds of millions of people can cause a great deal of negative impact.</p><p>2025 has seen the arrival of agents that can do real cognitive work; writing computer code will never be the same. 2026 will likely see the arrival of systems that can figure out novel insights. 2027 may see the arrival of robots that can do tasks in the real world.</p><p>A lot more people will be able to create software, and art. But the world wants a lot more of both, and experts will probably still be much better than novices, as long as they embrace the new tools. Generally speaking, the ability for one person to get much more done in 2030 than they could in 2020 will be a striking change, and one many people will figure out how to benefit from.</p><p>In the most important ways, the 2030s may not be wildly different. People will still love their families, express their creativity, play games, and swim in lakes.<br /></p><p>But in still-very-important-ways, the 2030s are likely going to be wildly different from any time that has come before. We do not know how far beyond human-level intelligence we can go, but we are about to find out.</p><p>In the 2030s, intelligence and energy\u2014ideas, and the ability to make ideas happen\u2014are going to become wildly abundant. These two have been the fundamental limiters on human progress for a long time; with abundant intelligence and energy (and good governance), we can theoretically have anything else.</p><p>Already we live with incredible digital intelligence, and after some initial shock, most of us are pretty used to it. Very quickly we go from being amazed that AI can generate a beautifully-written paragraph to wondering when it can generate a beautifully-written novel; or from being amazed that it can make live-saving medical diagnoses to wondering when it can develop the cures; or from being amazed it can create a small computer program to wondering when it can create an entire new company. This is how the singularity goes: wonders become routine, and then table stakes.</p><p>We already hear from scientists that they are two or three times more productive than they were before AI. Advanced AI is interesting for many reasons, but perhaps nothing is quite as significant as the fact that we can use it to do faster AI research. We may be able to discover new computing substrates, better algorithms, and who knows what else. If we can do a decade\u2019s worth of research in a year, or a month, then the rate of progress will obviously be quite different.</p><p>From here on, the tools we have already built will help us find further scientific insights and aid us in creating better AI systems. Of course this isn\u2019t the same thing as an AI system completely autonomously updating its own code, but nevertheless this is a larval version of recursive self-improvement.</p><p>There are other self-reinforcing loops at play. The economic value creation has started a flywheel of compounding infrastructure buildout to run these increasingly-powerful AI systems. And robots that can build other robots (and in some sense, datacenters that can build other datacenters) aren\u2019t that far off.\u00a0</p><p>If we have to make the first million humanoid robots the old-fashioned way, but then they can operate the entire supply chain\u2014digging and refining minerals, driving trucks, running factories, etc.\u2014to build more robots, which can build more chip fabrication facilities, data centers, etc, then the rate of progress will obviously be quite different.</p><p>As datacenter production gets automated, the cost of intelligence should eventually converge to near the cost of electricity. (People are often curious about how much energy a ChatGPT query uses; the average query uses about 0.34 watt-hours, about what an oven would use in a little over one second, or a high-efficiency lightbulb would use in a couple of minutes. It also uses about 0.000085 gallons of water; roughly one fifteenth of a teaspoon.)</p><p>The rate of technological progress will keep accelerating, and it will continue to be the case that people are capable of adapting to almost anything. There will be very hard parts like whole classes of jobs going away, but on the other hand the world will be getting so much richer so quickly that we\u2019ll be able to seriously entertain new policy ideas we never could before. We probably won\u2019t adopt a new social contract all at once, but when we look back in a few decades, the gradual changes will have amounted to something big.</p><p>If history is any guide, we will figure out new things to do and new things to want, and assimilate new tools quickly (job change after the industrial revolution is a good recent example). Expectations will go up, but capabilities will go up equally quickly, and we\u2019ll all get better stuff. We will build ever-more-wonderful things for each other. People have a long-term important and curious advantage over AI: we are hard-wired to care about other people and what they think and do, and we don\u2019t care very much about machines.</p><p>A subsistence farmer from a thousand years ago would look at what many of us do and say we have fake jobs, and think that we are just playing games to entertain ourselves since we have plenty of food and unimaginable luxuries. I hope we will look at the jobs a thousand years in the future and think they are very fake jobs, and I have no doubt they will feel incredibly important and satisfying to the people doing them.</p><p>The rate of new wonders being achieved will be immense. It\u2019s hard to even imagine today what we will have discovered by 2035; maybe we will go from solving high-energy physics one year to beginning space colonization the next year; or from a major materials science breakthrough one year to true high-bandwidth brain-computer interfaces the next year. Many people will choose to live their lives in much the same way, but at least some people will probably decide to \u201cplug in\u201d.</p><p>Looking forward, this sounds hard to wrap our heads around. But probably living through it will feel impressive but manageable. From a relativistic perspective, the singularity happens bit by bit, and the merge happens slowly. We are climbing the long arc of exponential technological progress; it always looks vertical looking forward and flat going backwards, but it\u2019s one smooth curve. (Think back to 2020, and what it would have sounded like to have something close to AGI by 2025, versus what the last 5 years have actually been like.)</p><p>There are serious challenges to confront along with the huge upsides. We do need to solve the safety issues, technically and societally, but then it\u2019s critically important to widely distribute access to superintelligence given the economic implications. The best path forward might be something like:</p><ol>\n<li><p>Solve the alignment problem, meaning that we can robustly guarantee that we get AI systems to learn and act towards what we collectively really want over the long-term (social media feeds are an example of misaligned AI; the algorithms that power those are incredible at getting you to keep scrolling and clearly understand your short-term preferences, but they do so by exploiting something in your brain that overrides your long-term preference).</p></li>\n<li><p>Then focus on making superintelligence cheap, widely available, and not too concentrated with any person, company, or country. Society is resilient, creative, and adapts quickly. If we can harness the collective will and wisdom of people, then although we\u2019ll make plenty of mistakes and some things will go really wrong, we will learn and adapt quickly and be able to use this technology to get maximum upside and minimal downside. Giving users a lot of freedom, within broad bounds society has to decide on, seems very important. The sooner the world can start a conversation about what these broad bounds are and how we define collective alignment, the better.</p></li>\n</ol><p>We (the whole industry, not just OpenAI) are building a brain for the world. It will be extremely personalized and easy for everyone to use; we will be limited by good ideas. For a long time, technical people in the startup industry have made fun of \u201cthe idea guys\u201d; people who had an idea and were looking for a team to build it. It now looks to me like they are about to have their day in the sun.</p><p>OpenAI is a lot of things now, but before anything else, we are a superintelligence research company. We have a lot of work in front of us, but most of the path in front of us is now lit, and the dark areas are receding fast. We feel extraordinarily grateful to get to do what we do.</p><p>Intelligence too cheap to meter is well within grasp. This may sound crazy to say, but if we told you back in 2020 we were going to be where we are today, it probably sounded more crazy than our current predictions about 2030.</p><p>May we scale smoothly, exponentially and uneventfully through superintelligence.</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>We are past the event horizon; the takeoff has started. Humanity is close to building digital superintelligence, and at least so far it\u2019s much less weird than it seems like it should be.</p><p>Robots are not yet walking the streets, nor are most of us talking to AI all day. People still die of disease, we still can\u2019t easily go to space, and there is a lot about the universe we don\u2019t understand.</p><p>And yet, we have recently built systems that are smarter than people in many ways, and are able to significantly amplify the output of people using them. The least-likely part of the work is behind us; the scientific insights that got us to systems like GPT-4 and o3 were hard-won, but will take us very far.</p><p>AI will contribute to the world in many ways, but the gains to quality of life from AI driving faster scientific progress and increased productivity will be enormous; the future can be vastly better than the present. Scientific progress is the biggest driver of overall progress; it\u2019s hugely exciting to think about how much more we could have.</p><p>In some big sense, ChatGPT is already more powerful than any human who has ever lived. Hundreds of millions of people rely on it every day and for increasingly important tasks; a small new capability can create a hugely positive impact; a small misalignment multiplied by hundreds of millions of people can cause a great deal of negative impact.</p><p>2025 has seen the arrival of agents that can do real cognitive work; writing computer code will never be the same. 2026 will likely see the arrival of systems that can figure out novel insights. 2027 may see the arrival of robots that can do tasks in the real world.</p><p>A lot more people will be able to create software, and art. But the world wants a lot more of both, and experts will probably still be much better than novices, as long as they embrace the new tools. Generally speaking, the ability for one person to get much more done in 2030 than they could in 2020 will be a striking change, and one many people will figure out how to benefit from.</p><p>In the most important ways, the 2030s may not be wildly different. People will still love their families, express their creativity, play games, and swim in lakes.<br /></p><p>But in still-very-important-ways, the 2030s are likely going to be wildly different from any time that has come before. We do not know how far beyond human-level intelligence we can go, but we are about to find out.</p><p>In the 2030s, intelligence and energy\u2014ideas, and the ability to make ideas happen\u2014are going to become wildly abundant. These two have been the fundamental limiters on human progress for a long time; with abundant intelligence and energy (and good governance), we can theoretically have anything else.</p><p>Already we live with incredible digital intelligence, and after some initial shock, most of us are pretty used to it. Very quickly we go from being amazed that AI can generate a beautifully-written paragraph to wondering when it can generate a beautifully-written novel; or from being amazed that it can make live-saving medical diagnoses to wondering when it can develop the cures; or from being amazed it can create a small computer program to wondering when it can create an entire new company. This is how the singularity goes: wonders become routine, and then table stakes.</p><p>We already hear from scientists that they are two or three times more productive than they were before AI. Advanced AI is interesting for many reasons, but perhaps nothing is quite as significant as the fact that we can use it to do faster AI research. We may be able to discover new computing substrates, better algorithms, and who knows what else. If we can do a decade\u2019s worth of research in a year, or a month, then the rate of progress will obviously be quite different.</p><p>From here on, the tools we have already built will help us find further scientific insights and aid us in creating better AI systems. Of course this isn\u2019t the same thing as an AI system completely autonomously updating its own code, but nevertheless this is a larval version of recursive self-improvement.</p><p>There are other self-reinforcing loops at play. The economic value creation has started a flywheel of compounding infrastructure buildout to run these increasingly-powerful AI systems. And robots that can build other robots (and in some sense, datacenters that can build other datacenters) aren\u2019t that far off.\u00a0</p><p>If we have to make the first million humanoid robots the old-fashioned way, but then they can operate the entire supply chain\u2014digging and refining minerals, driving trucks, running factories, etc.\u2014to build more robots, which can build more chip fabrication facilities, data centers, etc, then the rate of progress will obviously be quite different.</p><p>As datacenter production gets automated, the cost of intelligence should eventually converge to near the cost of electricity. (People are often curious about how much energy a ChatGPT query uses; the average query uses about 0.34 watt-hours, about what an oven would use in a little over one second, or a high-efficiency lightbulb would use in a couple of minutes. It also uses about 0.000085 gallons of water; roughly one fifteenth of a teaspoon.)</p><p>The rate of technological progress will keep accelerating, and it will continue to be the case that people are capable of adapting to almost anything. There will be very hard parts like whole classes of jobs going away, but on the other hand the world will be getting so much richer so quickly that we\u2019ll be able to seriously entertain new policy ideas we never could before. We probably won\u2019t adopt a new social contract all at once, but when we look back in a few decades, the gradual changes will have amounted to something big.</p><p>If history is any guide, we will figure out new things to do and new things to want, and assimilate new tools quickly (job change after the industrial revolution is a good recent example). Expectations will go up, but capabilities will go up equally quickly, and we\u2019ll all get better stuff. We will build ever-more-wonderful things for each other. People have a long-term important and curious advantage over AI: we are hard-wired to care about other people and what they think and do, and we don\u2019t care very much about machines.</p><p>A subsistence farmer from a thousand years ago would look at what many of us do and say we have fake jobs, and think that we are just playing games to entertain ourselves since we have plenty of food and unimaginable luxuries. I hope we will look at the jobs a thousand years in the future and think they are very fake jobs, and I have no doubt they will feel incredibly important and satisfying to the people doing them.</p><p>The rate of new wonders being achieved will be immense. It\u2019s hard to even imagine today what we will have discovered by 2035; maybe we will go from solving high-energy physics one year to beginning space colonization the next year; or from a major materials science breakthrough one year to true high-bandwidth brain-computer interfaces the next year. Many people will choose to live their lives in much the same way, but at least some people will probably decide to \u201cplug in\u201d.</p><p>Looking forward, this sounds hard to wrap our heads around. But probably living through it will feel impressive but manageable. From a relativistic perspective, the singularity happens bit by bit, and the merge happens slowly. We are climbing the long arc of exponential technological progress; it always looks vertical looking forward and flat going backwards, but it\u2019s one smooth curve. (Think back to 2020, and what it would have sounded like to have something close to AGI by 2025, versus what the last 5 years have actually been like.)</p><p>There are serious challenges to confront along with the huge upsides. We do need to solve the safety issues, technically and societally, but then it\u2019s critically important to widely distribute access to superintelligence given the economic implications. The best path forward might be something like:</p><ol>\n<li><p>Solve the alignment problem, meaning that we can robustly guarantee that we get AI systems to learn and act towards what we collectively really want over the long-term (social media feeds are an example of misaligned AI; the algorithms that power those are incredible at getting you to keep scrolling and clearly understand your short-term preferences, but they do so by exploiting something in your brain that overrides your long-term preference).</p></li>\n<li><p>Then focus on making superintelligence cheap, widely available, and not too concentrated with any person, company, or country. Society is resilient, creative, and adapts quickly. If we can harness the collective will and wisdom of people, then although we\u2019ll make plenty of mistakes and some things will go really wrong, we will learn and adapt quickly and be able to use this technology to get maximum upside and minimal downside. Giving users a lot of freedom, within broad bounds society has to decide on, seems very important. The sooner the world can start a conversation about what these broad bounds are and how we define collective alignment, the better.</p></li>\n</ol><p>We (the whole industry, not just OpenAI) are building a brain for the world. It will be extremely personalized and easy for everyone to use; we will be limited by good ideas. For a long time, technical people in the startup industry have made fun of \u201cthe idea guys\u201d; people who had an idea and were looking for a team to build it. It now looks to me like they are about to have their day in the sun.</p><p>OpenAI is a lot of things now, but before anything else, we are a superintelligence research company. We have a lot of work in front of us, but most of the path in front of us is now lit, and the dark areas are receding fast. We feel extraordinarily grateful to get to do what we do.</p><p>Intelligence too cheap to meter is well within grasp. This may sound crazy to say, but if we told you back in 2020 we were going to be where we are today, it probably sounded more crazy than our current predictions about 2030.</p><p>May we scale smoothly, exponentially and uneventfully through superintelligence.</p></div>", "image": null, "banner_image": null, "date_published": "2025-06-10T13:12:47Z", "date_modified": "2026-04-11T01:19:36Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2174132", "url": "https://blog.samaltman.com/three-observations", "external_url": null, "title": "Three Observations", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>Our mission is to ensure that AGI (Artificial General Intelligence) benefits all of humanity.\u00a0</p><p>Systems that start to point to AGI* are coming into view, and so we think it\u2019s important to understand the moment we are in. AGI is a weakly defined term, but generally speaking we mean it to be a system that can tackle increasingly complex problems, at human level, in many fields.</p><p>People are tool-builders with an inherent drive to understand and create, which leads to the world getting better for all of us. Each new generation builds upon the discoveries of the generations before to create even more capable tools\u2014electricity, the transistor, the computer, the internet, and soon AGI.</p><p>Over time, in fits and starts, the steady march of human innovation has brought previously unimaginable levels of prosperity and improvements to almost every aspect of people\u2019s lives.</p><p>In some sense, AGI is just another tool in this ever-taller scaffolding of human progress we are building together. In another sense, it is the beginning of something for which it\u2019s hard not to say \u201cthis time it\u2019s different\u201d; the economic growth in front of us looks astonishing, and we can now imagine a world where we cure all diseases, have much more time to enjoy with our families, and can fully realize our creative potential.</p><p>In a decade, perhaps everyone on earth will be capable of accomplishing more than the most impactful person can today.</p><p>We continue to see rapid progress with AI development. Here are three observations about the economics of AI:</p><p>1. <b>The intelligence of an AI model roughly equals the log of the resources used to train and run it.</b> These resources are chiefly training compute, data, and inference compute. It appears that you can spend arbitrary amounts of money and get continuous and predictable gains; the scaling laws that predict this are accurate over many orders of magnitude.</p><p>2. <b>The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use.</b> You can see this in the token cost from GPT-4 in early 2023 to GPT-4o in mid-2024, where the price per token dropped about 150x in that time period. Moore\u2019s law changed the world at 2x every 18 months; this is unbelievably stronger.\u00a0</p><p>3. <b>The socioeconomic value of linearly increasing intelligence is super-exponential in nature.</b> A consequence of this is that we see no reason for exponentially increasing investment to stop in the near future.</p><p>If these three observations continue to hold true, the impacts on society will be significant.</p><p>We are now starting to roll out AI agents, which will eventually feel like virtual co-workers.</p><p>Let\u2019s imagine the case of a software engineering agent, which is an agent that we expect to be particularly important. Imagine that this agent will eventually be capable of doing most things a software engineer at a top company with a few years of experience could do, for tasks up to a couple of days long. It will not have the biggest new ideas, it will require lots of human supervision and direction, and it will be great at some things but surprisingly bad at others.</p><p>Still, imagine it as a real-but-relatively-junior virtual coworker. Now imagine 1,000 of them. Or 1 million of them. Now imagine such agents in every field of knowledge work.</p><p>In some ways, AI may turn out to be like the transistor economically\u2014a big scientific discovery that scales well and that seeps into almost every corner of the economy. We don\u2019t think much about transistors, or transistor companies, and the gains are very widely distributed. But we do expect our computers, TVs, cars, toys, and more to perform miracles.</p><p>The world will not change all at once; it never does. Life will go on mostly the same in the short run, and people in 2025 will mostly spend their time in the same way they did in 2024. We will still fall in love, create families, get in fights online, hike in nature, etc.</p><p>But the future will be coming at us in a way that is impossible to ignore, and the long-term changes to our society and economy will be huge. We will find new things to do, new ways to be useful to each other, and new ways to compete, but they may not look very much like the jobs of today.\u00a0</p><p>Agency, willfulness, and determination will likely be extremely valuable. Correctly deciding what to do and figuring out how to navigate an ever-changing world will have huge value; resilience and adaptability will be helpful skills to cultivate. AGI will be the biggest lever ever on human willfulness, and enable individual people to have more impact than ever before, not less.</p><p>We expect the impact of AGI to be uneven. Although some industries will change very little, scientific progress will likely be much faster than it is today; this impact of AGI may surpass everything else.</p><p>The price of many goods will eventually fall dramatically (right now, the cost of intelligence and the cost of energy constrain a lot of things), and the price of luxury goods and a few inherently limited resources like land may rise even more dramatically.</p><p>Technically speaking, the road in front of us looks fairly clear. But public policy and collective opinion on how we should integrate AGI into society matter a lot; one of our reasons for launching products early and often is to give society and the technology time to co-evolve.</p><p>AI will seep into all areas of the economy and society; we will expect everything to be smart. Many of us expect to need to give people more control over the technology than we have historically, including open-sourcing more, and accept that there is a balance between safety and individual empowerment that will require trade-offs.</p><p>While we never want to be reckless and there will likely be some major decisions and limitations related to AGI safety that will be unpopular, directionally, as we get closer to achieving AGI, we believe that trending more towards individual empowerment is important; the other likely path we can see is AI being used by authoritarian governments to control their population through mass surveillance and loss of autonomy.</p><p>Ensuring that the benefits of AGI are broadly distributed is critical. The historical impact of technological progress suggests that most of the metrics we care about (health outcomes, economic prosperity, etc.) get better on average and over the long-term, but increasing equality does not seem technologically determined and getting this right may require new ideas.</p><p>In particular, it does seem like the balance of power between capital and labor could easily get messed up, and this may require early intervention. We are open to strange-sounding ideas like giving some \u201ccompute budget\u201d to enable everyone on Earth to use a lot of AI, but we can also see a lot of ways where just relentlessly driving the cost of intelligence as low as possible has the desired effect.</p><p>Anyone in 2035 should be able to marshall the intellectual capacity equivalent to everyone in 2025; everyone should have access to unlimited genius to direct however they can imagine. There is a great deal of talent right now without the resources to fully express itself, and if we change that, the resulting creative output of the world will lead to tremendous benefits for us all.</p><br /><br /><br /><p><br /></p><br /><p>Thanks especially to Josh Achiam, Boaz Barak and Aleksander Madry for reviewing drafts of this.</p><p>*By using the term AGI here, we aim to communicate clearly, and we do not intend to alter or interpret the definitions and processes that define our relationship with Microsoft. We fully expect to be partnered with Microsoft for the long term. This footnote seems silly, but on the other hand we know some journalists will try to get clicks by writing something silly so here we are pre-empting the silliness\u2026</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>Our mission is to ensure that AGI (Artificial General Intelligence) benefits all of humanity.\u00a0</p><p>Systems that start to point to AGI* are coming into view, and so we think it\u2019s important to understand the moment we are in. AGI is a weakly defined term, but generally speaking we mean it to be a system that can tackle increasingly complex problems, at human level, in many fields.</p><p>People are tool-builders with an inherent drive to understand and create, which leads to the world getting better for all of us. Each new generation builds upon the discoveries of the generations before to create even more capable tools\u2014electricity, the transistor, the computer, the internet, and soon AGI.</p><p>Over time, in fits and starts, the steady march of human innovation has brought previously unimaginable levels of prosperity and improvements to almost every aspect of people\u2019s lives.</p><p>In some sense, AGI is just another tool in this ever-taller scaffolding of human progress we are building together. In another sense, it is the beginning of something for which it\u2019s hard not to say \u201cthis time it\u2019s different\u201d; the economic growth in front of us looks astonishing, and we can now imagine a world where we cure all diseases, have much more time to enjoy with our families, and can fully realize our creative potential.</p><p>In a decade, perhaps everyone on earth will be capable of accomplishing more than the most impactful person can today.</p><p>We continue to see rapid progress with AI development. Here are three observations about the economics of AI:</p><p>1. <b>The intelligence of an AI model roughly equals the log of the resources used to train and run it.</b> These resources are chiefly training compute, data, and inference compute. It appears that you can spend arbitrary amounts of money and get continuous and predictable gains; the scaling laws that predict this are accurate over many orders of magnitude.</p><p>2. <b>The cost to use a given level of AI falls about 10x every 12 months, and lower prices lead to much more use.</b> You can see this in the token cost from GPT-4 in early 2023 to GPT-4o in mid-2024, where the price per token dropped about 150x in that time period. Moore\u2019s law changed the world at 2x every 18 months; this is unbelievably stronger.\u00a0</p><p>3. <b>The socioeconomic value of linearly increasing intelligence is super-exponential in nature.</b> A consequence of this is that we see no reason for exponentially increasing investment to stop in the near future.</p><p>If these three observations continue to hold true, the impacts on society will be significant.</p><p>We are now starting to roll out AI agents, which will eventually feel like virtual co-workers.</p><p>Let\u2019s imagine the case of a software engineering agent, which is an agent that we expect to be particularly important. Imagine that this agent will eventually be capable of doing most things a software engineer at a top company with a few years of experience could do, for tasks up to a couple of days long. It will not have the biggest new ideas, it will require lots of human supervision and direction, and it will be great at some things but surprisingly bad at others.</p><p>Still, imagine it as a real-but-relatively-junior virtual coworker. Now imagine 1,000 of them. Or 1 million of them. Now imagine such agents in every field of knowledge work.</p><p>In some ways, AI may turn out to be like the transistor economically\u2014a big scientific discovery that scales well and that seeps into almost every corner of the economy. We don\u2019t think much about transistors, or transistor companies, and the gains are very widely distributed. But we do expect our computers, TVs, cars, toys, and more to perform miracles.</p><p>The world will not change all at once; it never does. Life will go on mostly the same in the short run, and people in 2025 will mostly spend their time in the same way they did in 2024. We will still fall in love, create families, get in fights online, hike in nature, etc.</p><p>But the future will be coming at us in a way that is impossible to ignore, and the long-term changes to our society and economy will be huge. We will find new things to do, new ways to be useful to each other, and new ways to compete, but they may not look very much like the jobs of today.\u00a0</p><p>Agency, willfulness, and determination will likely be extremely valuable. Correctly deciding what to do and figuring out how to navigate an ever-changing world will have huge value; resilience and adaptability will be helpful skills to cultivate. AGI will be the biggest lever ever on human willfulness, and enable individual people to have more impact than ever before, not less.</p><p>We expect the impact of AGI to be uneven. Although some industries will change very little, scientific progress will likely be much faster than it is today; this impact of AGI may surpass everything else.</p><p>The price of many goods will eventually fall dramatically (right now, the cost of intelligence and the cost of energy constrain a lot of things), and the price of luxury goods and a few inherently limited resources like land may rise even more dramatically.</p><p>Technically speaking, the road in front of us looks fairly clear. But public policy and collective opinion on how we should integrate AGI into society matter a lot; one of our reasons for launching products early and often is to give society and the technology time to co-evolve.</p><p>AI will seep into all areas of the economy and society; we will expect everything to be smart. Many of us expect to need to give people more control over the technology than we have historically, including open-sourcing more, and accept that there is a balance between safety and individual empowerment that will require trade-offs.</p><p>While we never want to be reckless and there will likely be some major decisions and limitations related to AGI safety that will be unpopular, directionally, as we get closer to achieving AGI, we believe that trending more towards individual empowerment is important; the other likely path we can see is AI being used by authoritarian governments to control their population through mass surveillance and loss of autonomy.</p><p>Ensuring that the benefits of AGI are broadly distributed is critical. The historical impact of technological progress suggests that most of the metrics we care about (health outcomes, economic prosperity, etc.) get better on average and over the long-term, but increasing equality does not seem technologically determined and getting this right may require new ideas.</p><p>In particular, it does seem like the balance of power between capital and labor could easily get messed up, and this may require early intervention. We are open to strange-sounding ideas like giving some \u201ccompute budget\u201d to enable everyone on Earth to use a lot of AI, but we can also see a lot of ways where just relentlessly driving the cost of intelligence as low as possible has the desired effect.</p><p>Anyone in 2035 should be able to marshall the intellectual capacity equivalent to everyone in 2025; everyone should have access to unlimited genius to direct however they can imagine. There is a great deal of talent right now without the resources to fully express itself, and if we change that, the resulting creative output of the world will lead to tremendous benefits for us all.</p><br /><br /><br /><p><br /></p><br /><p>Thanks especially to Josh Achiam, Boaz Barak and Aleksander Madry for reviewing drafts of this.</p><p>*By using the term AGI here, we aim to communicate clearly, and we do not intend to alter or interpret the definitions and processes that define our relationship with Microsoft. We fully expect to be partnered with Microsoft for the long term. This footnote seems silly, but on the other hand we know some journalists will try to get clicks by writing something silly so here we are pre-empting the silliness\u2026</p></div>", "image": null, "banner_image": null, "date_published": "2025-02-09T13:05:32Z", "date_modified": "2026-04-11T01:02:58Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2163678", "url": "https://blog.samaltman.com/reflections", "external_url": null, "title": "Reflections", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>The second birthday of ChatGPT was only a little over a month ago, and now we have transitioned into the next paradigm of models that can do complex reasoning. New years get people in a reflective mood, and I wanted to share some personal thoughts about how it has gone so far, and some of the things I\u2019ve learned along the way.</p><p>As we get closer to AGI, it feels like an important time to look at the progress of our company. There is still so much to understand, still so much we don\u2019t know, and it\u2019s still so early. But we know a lot more than we did when we started.</p><p>We started OpenAI almost nine years ago because we believed that AGI was possible, and that it could be the most impactful technology in human history. We wanted to figure out how to build it and make it broadly beneficial; we were excited to try to make our mark on history. Our ambitions were extraordinarily high and so was our belief that the work might benefit society in an equally extraordinary way.</p><p>At the time, very few people cared, and if they did, it was mostly because they thought we had no chance of success.</p><p>In 2022, OpenAI was a quiet research lab working on something temporarily called \u201cChat With GPT-3.5\u201d. (We are much better at research than we are at naming things.) We had been watching people use the playground feature of our API and knew that developers were really enjoying talking to the model. We thought building a demo around that experience would show people something important about the future and help us make our models better and safer.</p><p>We ended up mercifully calling it ChatGPT instead, and launched it on November 30th of 2022.</p><p>We always knew, abstractly, that at some point we would hit a tipping point and the AI revolution would get kicked off. But we didn\u2019t know what the moment would be. To our surprise, it turned out to be this.</p><p>The launch of ChatGPT kicked off a growth curve like nothing we have ever seen\u2014in our company, our industry, and the world broadly. We are finally seeing some of the massive upside we have always hoped for from AI, and we can see how much more will come soon.</p><p></p><hr /><p></p><p>It hasn\u2019t been easy. The road hasn\u2019t been smooth and the right choices haven\u2019t been obvious.</p><p>In the last two years, we had to build an entire company, almost from scratch, around this new technology. There is no way to train people for this except by doing it, and when the technology category is completely new, there is no one at all who can tell you exactly how it should be done.</p><p>Building up a company at such high velocity with so little training is a messy process. It\u2019s often two steps forward, one step back (and sometimes, one step forward and two steps back). Mistakes get corrected as you go along, but there aren\u2019t really any handbooks or guideposts when you\u2019re doing original work. Moving at speed in uncharted waters is an incredible experience, but it is also immensely stressful for all the players. Conflicts and misunderstanding abound.</p><p>These years have been the most rewarding, fun, best, interesting, exhausting, stressful, and\u2014particularly for the last two\u2014unpleasant years of my life so far. The overwhelming feeling is gratitude; I know that someday I\u2019ll be retired at our ranch watching the plants grow, a little bored, and will think back at how cool it was that I got to do the work I dreamed of since I was a little kid. I try to remember that on any given Friday, when seven things go badly wrong by 1 pm.</p><p></p><hr /><p></p><p>A little over a year ago, on one particular Friday, the main thing that had gone wrong that day was that I got fired by surprise on a video call, and then right after we hung up the board published a blog post about it. I was in a hotel room in Las Vegas. It felt, to a degree that is almost impossible to explain, like a dream gone wrong.</p><p>Getting fired in public with no warning kicked off a really crazy few hours, and a pretty crazy few days. The \u201cfog of war\u201d was the strangest part. None of us were able to get satisfactory answers about what had happened, or why.\u00a0</p><p>The whole event was, in my opinion, a big failure of governance by well-meaning people, myself included. Looking back, I certainly wish I had done things differently, and I\u2019d like to believe I\u2019m a better, more thoughtful leader today than I was a year ago.</p><p>I also learned the importance of a board with diverse viewpoints and broad experience in managing a complex set of challenges. Good governance requires a lot of trust and credibility. I appreciate the way so many people worked together to build a stronger system of governance for OpenAI that enables us to pursue our mission of ensuring that AGI benefits all of humanity.</p><p>My biggest takeaway is how much I have to be thankful for and how many people I owe gratitude towards: to everyone who works at OpenAI and has chosen to spend their time and effort going after this dream, to friends who helped us get through the crisis moments, to our partners and customers who supported us and entrusted us to enable their success, and to the people in my life who showed me how much they cared. [1]</p><p>We all got back to the work in a more cohesive and positive way and I\u2019m very proud of our focus since then. We have done what is easily some of our best research ever. We grew from about 100 million weekly active users to more than 300 million. Most of all, we have continued to put technology out into the world that people genuinely seem to love and that solves real problems.</p><p></p><hr /><p></p><p>Nine years ago, we really had no idea what we were eventually going to become; even now, we only sort of know. AI development has taken many twists and turns and we expect more in the future.</p><p>Some of the twists have been joyful; some have been hard. It\u2019s been fun watching a steady stream of research miracles occur, and a lot of naysayers have become true believers. We\u2019ve also seen some colleagues split off and become competitors. Teams tend to turn over as they scale, and OpenAI scales really fast. I think some of this is unavoidable\u2014startups usually see a lot of turnover at each new major level of scale, and at OpenAI numbers go up by orders of magnitude every few months. The last two years have been like a decade at a normal company. When any company grows and evolves so fast, interests naturally diverge. And when any company in an important industry is in the lead, lots of people attack it for all sorts of reasons, especially when they are trying to compete with it.</p><p>Our vision won\u2019t change; our tactics will continue to evolve. For example, when we started we had no idea we would have to build a product company; we thought we were just going to do great research. We also had no idea we would need such a crazy amount of capital. There are new things we have to go build now that we didn\u2019t understand a few years ago, and there will be new things in the future we can barely imagine now.\u00a0</p><p>We are proud of our track-record on research and deployment so far, and are committed to continuing to advance our thinking on safety and benefits sharing. We continue to believe that the best way to make an AI system safe is by iteratively and gradually releasing it into the world, giving society time to adapt and co-evolve with the technology, learning from experience, and continuing to make the technology safer. We believe in the importance of being world leaders on safety and alignment research, and in guiding that research with feedback from real world applications.</p><p>We are now confident we know how to build AGI as we have traditionally understood it. We believe that, in 2025, we may see the first AI agents \u201cjoin the workforce\u201d and materially change the output of companies. We continue to believe that iteratively putting great tools in the hands of people leads to great, broadly-distributed outcomes.</p><p>We are beginning to turn our aim beyond that, to superintelligence in the true sense of the word. We love our current products, but we are here for the glorious future. With superintelligence, we can do anything else. Superintelligent tools could massively accelerate scientific discovery and innovation well beyond what we are capable of doing on our own, and in turn massively increase abundance and prosperity.</p><p>This sounds like science fiction right now, and somewhat crazy to even talk about it. That\u2019s alright\u2014we\u2019ve been there before and we\u2019re OK with being there again. We\u2019re pretty confident that in the next few years, everyone will see what we see, and that the need to act with great care, while still maximizing broad benefit and empowerment, is so important. Given the possibilities of our work, OpenAI cannot be a normal company.</p><p>How lucky and humbling it is to be able to play a role in this work.</p><p>(Thanks to <a href=\"https://www.bloomberg.com/features/2025-sam-altman-interview/\" title=\"Link: https://www.bloomberg.com/features/2025-sam-altman-interview/\">Josh Tyrangiel for sort of prompting this</a>. I wish we had had a lot more time.)</p><p><br /></p><p><br /></p><p><br /></p><p>[1]</p><p>There were a lot of people who did incredible and gigantic amounts of work to help OpenAI, and me personally, during those few days, but two people stood out from all others.</p><p>Ron Conway and Brian Chesky went so far above and beyond the call of duty that I\u2019m not even sure how to describe it. I\u2019ve of course heard stories about Ron\u2019s ability and tenaciousness for years and I\u2019ve spent a lot of time with Brian over the past couple of years getting a huge amount of help and advice.</p><p>But there\u2019s nothing quite like being in the foxhole with people to see what they can really do. I am reasonably confident OpenAI would have fallen apart without their help; they worked around the clock for days until things were done.</p><p>Although they worked unbelievably hard, they stayed calm and had clear strategic thought and great advice throughout. They stopped me from making several mistakes and made none themselves. They used their vast networks for everything needed and were able to navigate many complex situations. And I\u2019m sure they did a lot of things I don\u2019t know about.</p><p>What I will remember most, though, is their care, compassion, and support.</p><p>I thought I knew what it looked like to support a founder and a company, and in some small sense I did. But I have never before seen, or even heard of, anything like what these guys did, and now I get more fully why they have the legendary status they do. They are different and both fully deserve their genuinely unique reputations, but they are similar in their remarkable ability to move mountains and help, and in their unwavering commitment in times of need. The tech industry is far better off for having both of them in it.</p><p>There are others like them; it is an amazingly special thing about our industry and does much more to make it all work than people realize. I look forward to paying it forward.</p><p>On a more personal note, thanks especially to Ollie for his support that weekend and always; he is incredible in every way and no one could ask for a better partner.</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>The second birthday of ChatGPT was only a little over a month ago, and now we have transitioned into the next paradigm of models that can do complex reasoning. New years get people in a reflective mood, and I wanted to share some personal thoughts about how it has gone so far, and some of the things I\u2019ve learned along the way.</p><p>As we get closer to AGI, it feels like an important time to look at the progress of our company. There is still so much to understand, still so much we don\u2019t know, and it\u2019s still so early. But we know a lot more than we did when we started.</p><p>We started OpenAI almost nine years ago because we believed that AGI was possible, and that it could be the most impactful technology in human history. We wanted to figure out how to build it and make it broadly beneficial; we were excited to try to make our mark on history. Our ambitions were extraordinarily high and so was our belief that the work might benefit society in an equally extraordinary way.</p><p>At the time, very few people cared, and if they did, it was mostly because they thought we had no chance of success.</p><p>In 2022, OpenAI was a quiet research lab working on something temporarily called \u201cChat With GPT-3.5\u201d. (We are much better at research than we are at naming things.) We had been watching people use the playground feature of our API and knew that developers were really enjoying talking to the model. We thought building a demo around that experience would show people something important about the future and help us make our models better and safer.</p><p>We ended up mercifully calling it ChatGPT instead, and launched it on November 30th of 2022.</p><p>We always knew, abstractly, that at some point we would hit a tipping point and the AI revolution would get kicked off. But we didn\u2019t know what the moment would be. To our surprise, it turned out to be this.</p><p>The launch of ChatGPT kicked off a growth curve like nothing we have ever seen\u2014in our company, our industry, and the world broadly. We are finally seeing some of the massive upside we have always hoped for from AI, and we can see how much more will come soon.</p><p></p><hr /><p></p><p>It hasn\u2019t been easy. The road hasn\u2019t been smooth and the right choices haven\u2019t been obvious.</p><p>In the last two years, we had to build an entire company, almost from scratch, around this new technology. There is no way to train people for this except by doing it, and when the technology category is completely new, there is no one at all who can tell you exactly how it should be done.</p><p>Building up a company at such high velocity with so little training is a messy process. It\u2019s often two steps forward, one step back (and sometimes, one step forward and two steps back). Mistakes get corrected as you go along, but there aren\u2019t really any handbooks or guideposts when you\u2019re doing original work. Moving at speed in uncharted waters is an incredible experience, but it is also immensely stressful for all the players. Conflicts and misunderstanding abound.</p><p>These years have been the most rewarding, fun, best, interesting, exhausting, stressful, and\u2014particularly for the last two\u2014unpleasant years of my life so far. The overwhelming feeling is gratitude; I know that someday I\u2019ll be retired at our ranch watching the plants grow, a little bored, and will think back at how cool it was that I got to do the work I dreamed of since I was a little kid. I try to remember that on any given Friday, when seven things go badly wrong by 1 pm.</p><p></p><hr /><p></p><p>A little over a year ago, on one particular Friday, the main thing that had gone wrong that day was that I got fired by surprise on a video call, and then right after we hung up the board published a blog post about it. I was in a hotel room in Las Vegas. It felt, to a degree that is almost impossible to explain, like a dream gone wrong.</p><p>Getting fired in public with no warning kicked off a really crazy few hours, and a pretty crazy few days. The \u201cfog of war\u201d was the strangest part. None of us were able to get satisfactory answers about what had happened, or why.\u00a0</p><p>The whole event was, in my opinion, a big failure of governance by well-meaning people, myself included. Looking back, I certainly wish I had done things differently, and I\u2019d like to believe I\u2019m a better, more thoughtful leader today than I was a year ago.</p><p>I also learned the importance of a board with diverse viewpoints and broad experience in managing a complex set of challenges. Good governance requires a lot of trust and credibility. I appreciate the way so many people worked together to build a stronger system of governance for OpenAI that enables us to pursue our mission of ensuring that AGI benefits all of humanity.</p><p>My biggest takeaway is how much I have to be thankful for and how many people I owe gratitude towards: to everyone who works at OpenAI and has chosen to spend their time and effort going after this dream, to friends who helped us get through the crisis moments, to our partners and customers who supported us and entrusted us to enable their success, and to the people in my life who showed me how much they cared. [1]</p><p>We all got back to the work in a more cohesive and positive way and I\u2019m very proud of our focus since then. We have done what is easily some of our best research ever. We grew from about 100 million weekly active users to more than 300 million. Most of all, we have continued to put technology out into the world that people genuinely seem to love and that solves real problems.</p><p></p><hr /><p></p><p>Nine years ago, we really had no idea what we were eventually going to become; even now, we only sort of know. AI development has taken many twists and turns and we expect more in the future.</p><p>Some of the twists have been joyful; some have been hard. It\u2019s been fun watching a steady stream of research miracles occur, and a lot of naysayers have become true believers. We\u2019ve also seen some colleagues split off and become competitors. Teams tend to turn over as they scale, and OpenAI scales really fast. I think some of this is unavoidable\u2014startups usually see a lot of turnover at each new major level of scale, and at OpenAI numbers go up by orders of magnitude every few months. The last two years have been like a decade at a normal company. When any company grows and evolves so fast, interests naturally diverge. And when any company in an important industry is in the lead, lots of people attack it for all sorts of reasons, especially when they are trying to compete with it.</p><p>Our vision won\u2019t change; our tactics will continue to evolve. For example, when we started we had no idea we would have to build a product company; we thought we were just going to do great research. We also had no idea we would need such a crazy amount of capital. There are new things we have to go build now that we didn\u2019t understand a few years ago, and there will be new things in the future we can barely imagine now.\u00a0</p><p>We are proud of our track-record on research and deployment so far, and are committed to continuing to advance our thinking on safety and benefits sharing. We continue to believe that the best way to make an AI system safe is by iteratively and gradually releasing it into the world, giving society time to adapt and co-evolve with the technology, learning from experience, and continuing to make the technology safer. We believe in the importance of being world leaders on safety and alignment research, and in guiding that research with feedback from real world applications.</p><p>We are now confident we know how to build AGI as we have traditionally understood it. We believe that, in 2025, we may see the first AI agents \u201cjoin the workforce\u201d and materially change the output of companies. We continue to believe that iteratively putting great tools in the hands of people leads to great, broadly-distributed outcomes.</p><p>We are beginning to turn our aim beyond that, to superintelligence in the true sense of the word. We love our current products, but we are here for the glorious future. With superintelligence, we can do anything else. Superintelligent tools could massively accelerate scientific discovery and innovation well beyond what we are capable of doing on our own, and in turn massively increase abundance and prosperity.</p><p>This sounds like science fiction right now, and somewhat crazy to even talk about it. That\u2019s alright\u2014we\u2019ve been there before and we\u2019re OK with being there again. We\u2019re pretty confident that in the next few years, everyone will see what we see, and that the need to act with great care, while still maximizing broad benefit and empowerment, is so important. Given the possibilities of our work, OpenAI cannot be a normal company.</p><p>How lucky and humbling it is to be able to play a role in this work.</p><p>(Thanks to <a href=\"https://www.bloomberg.com/features/2025-sam-altman-interview/\" title=\"Link: https://www.bloomberg.com/features/2025-sam-altman-interview/\">Josh Tyrangiel for sort of prompting this</a>. I wish we had had a lot more time.)</p><p><br /></p><p><br /></p><p><br /></p><p>[1]</p><p>There were a lot of people who did incredible and gigantic amounts of work to help OpenAI, and me personally, during those few days, but two people stood out from all others.</p><p>Ron Conway and Brian Chesky went so far above and beyond the call of duty that I\u2019m not even sure how to describe it. I\u2019ve of course heard stories about Ron\u2019s ability and tenaciousness for years and I\u2019ve spent a lot of time with Brian over the past couple of years getting a huge amount of help and advice.</p><p>But there\u2019s nothing quite like being in the foxhole with people to see what they can really do. I am reasonably confident OpenAI would have fallen apart without their help; they worked around the clock for days until things were done.</p><p>Although they worked unbelievably hard, they stayed calm and had clear strategic thought and great advice throughout. They stopped me from making several mistakes and made none themselves. They used their vast networks for everything needed and were able to navigate many complex situations. And I\u2019m sure they did a lot of things I don\u2019t know about.</p><p>What I will remember most, though, is their care, compassion, and support.</p><p>I thought I knew what it looked like to support a founder and a company, and in some small sense I did. But I have never before seen, or even heard of, anything like what these guys did, and now I get more fully why they have the legendary status they do. They are different and both fully deserve their genuinely unique reputations, but they are similar in their remarkable ability to move mountains and help, and in their unwavering commitment in times of need. The tech industry is far better off for having both of them in it.</p><p>There are others like them; it is an amazingly special thing about our industry and does much more to make it all work than people realize. I look forward to paying it forward.</p><p>On a more personal note, thanks especially to Ollie for his support that weekend and always; he is incredible in every way and no one could ask for a better partner.</p></div>", "image": null, "banner_image": null, "date_published": "2025-01-05T17:37:29Z", "date_modified": "2026-04-09T10:50:42Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2109392", "url": "https://blog.samaltman.com/gpt-4o", "external_url": null, "title": "GPT-4o", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>There are two things from our announcement today I wanted to highlight.</p><p>First, a key part of our mission is to put very capable AI tools in the hands of people for free (or at a great price). I am very proud that we\u2019ve made the best model in the world available for free in ChatGPT, without ads or anything like that.\u00a0</p><p>Our initial conception when we started OpenAI was that we\u2019d create AI and use it to create all sorts of benefits for the world. Instead, it now looks like we\u2019ll create AI and then other people will use it to create all sorts of amazing things that we all benefit from.\u00a0</p><p>We are a business and will find plenty of things to charge for, and that will help us provide free, outstanding AI service to (hopefully) billions of people.\u00a0</p><p>Second, the new voice (and video) mode is the best computer interface I\u2019ve ever used. It feels like AI from the movies; and it\u2019s still a bit surprising to me that it\u2019s real. Getting to human-level response times and expressiveness turns out to be a big change.</p><p>The original ChatGPT showed a hint of what was possible with language interfaces; this new thing feels viscerally different. It is fast, smart, fun, natural, and helpful.</p><p></p><p>Talking to a computer has never felt really natural for me; now it does. As we add (optional) personalization, access to your information, the ability to take actions on your behalf, and more, I can really see an exciting future where we are able to use computers to do much more than ever before.<br /></p><p>Finally, huge thanks to the team that poured so much work into making this happen!</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>There are two things from our announcement today I wanted to highlight.</p><p>First, a key part of our mission is to put very capable AI tools in the hands of people for free (or at a great price). I am very proud that we\u2019ve made the best model in the world available for free in ChatGPT, without ads or anything like that.\u00a0</p><p>Our initial conception when we started OpenAI was that we\u2019d create AI and use it to create all sorts of benefits for the world. Instead, it now looks like we\u2019ll create AI and then other people will use it to create all sorts of amazing things that we all benefit from.\u00a0</p><p>We are a business and will find plenty of things to charge for, and that will help us provide free, outstanding AI service to (hopefully) billions of people.\u00a0</p><p>Second, the new voice (and video) mode is the best computer interface I\u2019ve ever used. It feels like AI from the movies; and it\u2019s still a bit surprising to me that it\u2019s real. Getting to human-level response times and expressiveness turns out to be a big change.</p><p>The original ChatGPT showed a hint of what was possible with language interfaces; this new thing feels viscerally different. It is fast, smart, fun, natural, and helpful.</p><p></p><p>Talking to a computer has never felt really natural for me; now it does. As we add (optional) personalization, access to your information, the ability to take actions on your behalf, and more, I can really see an exciting future where we are able to use computers to do much more than ever before.<br /></p><p>Finally, huge thanks to the team that poured so much work into making this happen!</p></div>", "image": null, "banner_image": null, "date_published": "2024-05-13T09:39:38Z", "date_modified": "2026-04-10T21:25:25Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/2065118", "url": "https://blog.samaltman.com/what-i-wish-someone-had-told-me", "external_url": null, "title": "What I Wish Someone Had Told Me", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><ol>\n<li>Optimism, obsession, self-belief, raw horsepower and personal connections are how things get started.</li>\n<li>Cohesive teams, the right combination of calmness and urgency, and unreasonable commitment are how things get finished. Long-term orientation is in short supply; try not to worry about what people think in the short term, which will get easier over time.</li>\n<li>It is easier for a team to do a hard thing that really matters than to do an easy thing that doesn\u2019t really matter; audacious ideas motivate people.</li>\n<li>Incentives are superpowers; set them carefully.</li>\n<li>Concentrate your resources on a small number of high-conviction bets; this is easy to say but evidently hard to do. You can delete more stuff than you think.</li>\n<li>Communicate clearly and concisely.</li>\n<li>Fight bullshit and bureaucracy every time you see it and get other people to fight it too. Do not let the org chart get in the way of people working productively together.</li>\n<li>Outcomes are what count; don\u2019t let good process excuse bad results.</li>\n<li>Spend more time recruiting. Take risks on high-potential people with a fast rate of improvement. Look for evidence of getting stuff done in addition to intelligence.</li>\n<li>Superstars are even more valuable than they seem, but you have to evaluate people on their net impact on the performance of the organization.</li>\n<li>Fast iteration can make up for a lot; it\u2019s usually ok to be wrong if you iterate quickly. Plans should be measured in decades, execution should be measured in weeks.</li>\n<li>Don\u2019t fight the business equivalent of the laws of physics.</li>\n<li>Inspiration is perishable and life goes by fast. Inaction is a particularly insidious type of risk.</li>\n<li>Scale often has surprising emergent properties.</li>\n<li>Compounding exponentials are magic. In particular, you really want to build a business that gets a compounding advantage with scale.</li>\n<li>Get back up and keep going.</li>\n<li>Working with great people is one of the best parts of life.</li>\n</ol></div>", "summary": "<div class=\"posthaven-post-body\"><ol>\n<li>Optimism, obsession, self-belief, raw horsepower and personal connections are how things get started.</li>\n<li>Cohesive teams, the right combination of calmness and urgency, and unreasonable commitment are how things get finished. Long-term orientation is in short supply; try not to worry about what people think in the short term, which will get easier over time.</li>\n<li>It is easier for a team to do a hard thing that really matters than to do an easy thing that doesn\u2019t really matter; audacious ideas motivate people.</li>\n<li>Incentives are superpowers; set them carefully.</li>\n<li>Concentrate your resources on a small number of high-conviction bets; this is easy to say but evidently hard to do. You can delete more stuff than you think.</li>\n<li>Communicate clearly and concisely.</li>\n<li>Fight bullshit and bureaucracy every time you see it and get other people to fight it too. Do not let the org chart get in the way of people working productively together.</li>\n<li>Outcomes are what count; don\u2019t let good process excuse bad results.</li>\n<li>Spend more time recruiting. Take risks on high-potential people with a fast rate of improvement. Look for evidence of getting stuff done in addition to intelligence.</li>\n<li>Superstars are even more valuable than they seem, but you have to evaluate people on their net impact on the performance of the organization.</li>\n<li>Fast iteration can make up for a lot; it\u2019s usually ok to be wrong if you iterate quickly. Plans should be measured in decades, execution should be measured in weeks.</li>\n<li>Don\u2019t fight the business equivalent of the laws of physics.</li>\n<li>Inspiration is perishable and life goes by fast. Inaction is a particularly insidious type of risk.</li>\n<li>Scale often has surprising emergent properties.</li>\n<li>Compounding exponentials are magic. In particular, you really want to build a business that gets a compounding advantage with scale.</li>\n<li>Get back up and keep going.</li>\n<li>Working with great people is one of the best parts of life.</li>\n</ol></div>", "image": null, "banner_image": null, "date_published": "2023-12-21T14:44:25Z", "date_modified": "2026-04-11T01:59:41Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1854539", "url": "https://blog.samaltman.com/helion-needs-you", "external_url": null, "title": "Helion Needs You", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>Helion\u00a0has been progressing even faster than I expected and is on pace in 2024 to 1) demonstrate Q &gt; 1 fusion and 2) resolve all questions needed to design a mass-producible fusion generator.</p>        <div class=\"posthaven-gallery\" id=\"posthaven_gallery[1876625]\">\n                  <p class=\"posthaven-file posthaven-file-image posthaven-file-state-processed\">\n          <img class=\"posthaven-gallery-image\" src=\"https://phaven-prod.s3.amazonaws.com/files/image_part/asset/2831322/GkJUE4udQtYQW_H6ppnzsNo7cZ8/medium_FEP-edited-small.png\" />\n        </p>\n\n        </div>\n<p>The goals of the company are quite ambitious\u2014clean, continuous energy for 1 cent per kilowatt-hour, and the ability to manufacture enough power plants to satisfy the current electrical demand of earth in a ten year period.</p><p>If both things happen, it will transform the world. Abundant, clean, and radically inexpensive energy will elevate the quality of life for all of us\u2014think about how much the cost of energy factors into what we do and use. Also, electricity at this price will allow us to do things like efficiently capture carbon (so although we\u2019ll still rely on gasoline for awhile, it\u2019ll be ok).</p><p>Although\u00a0Helion\u2019s scientific progress of the past 8 years is phenomenal and necessary, it is not sufficient to rapidly get to this new energy economy.\u00a0Helion\u00a0now needs to figure out how to engineer machines that don\u2019t break, how to build a factory and supply chain capable of manufacturing a machine every day, how to work with power grids and governments around the world, and more.</p><p>The biggest input to the degree and speed of success at the company is now the talent of the people who join the team. Here are a few of the most critical jobs, but please don\u2019t let the lack of a perfect fit deter you from applying.</p><p>Electrical Engineer, Low Voltage:\u00a0<a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044506005\" target=\"_blank\">https://boards.greenhouse.io/helionenergy/jobs/4044506005</a><br />Electrical Engineer, Pulsed Power:\u00a0<a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044510005\" target=\"_blank\">https://boards.greenhouse.io/helionenergy/jobs/4044510005</a><br />Mechanical Engineer, Generator Systems:\u00a0<a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044522005\" target=\"_blank\">https://boards.greenhouse.io/helionenergy/jobs/4044522005</a><br />Manager of Mechanical Engineering:\u00a0<a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044521005\" target=\"_blank\">http</a><a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044521005\" target=\"_blank\">s://boards.greenhouse.io/helionenergy/jobs/4044521005</a></p><div>(All current jobs:\u00a0<a href=\"https://www.helionenergy.com/careers/\" target=\"_blank\">https://www.helionenergy.com/careers/</a>)<br />\n</div></div>", "summary": "<div class=\"posthaven-post-body\"><p>Helion\u00a0has been progressing even faster than I expected and is on pace in 2024 to 1) demonstrate Q &gt; 1 fusion and 2) resolve all questions needed to design a mass-producible fusion generator.</p>        <div class=\"posthaven-gallery\" id=\"posthaven_gallery[1876625]\">\n                  <p class=\"posthaven-file posthaven-file-image posthaven-file-state-processed\">\n          <img class=\"posthaven-gallery-image\" src=\"https://phaven-prod.s3.amazonaws.com/files/image_part/asset/2831322/GkJUE4udQtYQW_H6ppnzsNo7cZ8/medium_FEP-edited-small.png\" />\n        </p>\n\n        </div>\n<p>The goals of the company are quite ambitious\u2014clean, continuous energy for 1 cent per kilowatt-hour, and the ability to manufacture enough power plants to satisfy the current electrical demand of earth in a ten year period.</p><p>If both things happen, it will transform the world. Abundant, clean, and radically inexpensive energy will elevate the quality of life for all of us\u2014think about how much the cost of energy factors into what we do and use. Also, electricity at this price will allow us to do things like efficiently capture carbon (so although we\u2019ll still rely on gasoline for awhile, it\u2019ll be ok).</p><p>Although\u00a0Helion\u2019s scientific progress of the past 8 years is phenomenal and necessary, it is not sufficient to rapidly get to this new energy economy.\u00a0Helion\u00a0now needs to figure out how to engineer machines that don\u2019t break, how to build a factory and supply chain capable of manufacturing a machine every day, how to work with power grids and governments around the world, and more.</p><p>The biggest input to the degree and speed of success at the company is now the talent of the people who join the team. Here are a few of the most critical jobs, but please don\u2019t let the lack of a perfect fit deter you from applying.</p><p>Electrical Engineer, Low Voltage:\u00a0<a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044506005\" target=\"_blank\">https://boards.greenhouse.io/helionenergy/jobs/4044506005</a><br />Electrical Engineer, Pulsed Power:\u00a0<a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044510005\" target=\"_blank\">https://boards.greenhouse.io/helionenergy/jobs/4044510005</a><br />Mechanical Engineer, Generator Systems:\u00a0<a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044522005\" target=\"_blank\">https://boards.greenhouse.io/helionenergy/jobs/4044522005</a><br />Manager of Mechanical Engineering:\u00a0<a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044521005\" target=\"_blank\">http</a><a href=\"https://boards.greenhouse.io/helionenergy/jobs/4044521005\" target=\"_blank\">s://boards.greenhouse.io/helionenergy/jobs/4044521005</a></p><div>(All current jobs:\u00a0<a href=\"https://www.helionenergy.com/careers/\" target=\"_blank\">https://www.helionenergy.com/careers/</a>)<br />\n</div></div>", "image": null, "banner_image": null, "date_published": "2022-07-13T07:47:32Z", "date_modified": "2026-04-02T17:42:23Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1815658", "url": "https://blog.samaltman.com/dall-star-e-2", "external_url": null, "title": "DALL\u2022E 2", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>Today we did a research launch of DALL\u2022E 2, a new AI tool that can create and edit images from natural language instructions.\u00a0</p><p>Most importantly, we hope people love the tool and find it useful. For me, it\u2019s the most delightful thing to play with we\u2019ve created so far. I find it to be creativity-enhancing, helpful for many different situations, and fun in a way I haven\u2019t felt from technology in a while.</p><p>But I also think it\u2019s noteworthy for a few reasons:</p><p>1)\u00a0This is another example of what I think is going to be a new computer interface trend: you say what you want in natural language or with contextual clues, and the computer does it. We offer this for code and now image generation; both of these will get a lot better. But the same trend will happen in new ways until eventually it works for complex tasks\u2014we can imagine an \u201cAI office worker\u201d that takes requests in natural language like a human does.</p><p>2)\u00a0It sure does seem to \u201cunderstand\u201d concepts at many levels and how they relate to each other in sophisticated ways.</p><p>3)\u00a0Copilot is a tool that helps coders be more productive, but still is very far from being able to create a full program. DALL\u2022E 2 is a tool that will help artists and illustrators be more creative, but it can also create a \u201ccomplete work\u201d. This may be an early example of the impact AI on labor markets. Although I firmly believe AI will create lots of new jobs, and make many existing jobs much better by doing the boring bits well, I think it\u2019s important to be honest that it\u2019s increasingly going to make some jobs not very relevant (like technology frequently does).</p><p>4)\u00a0It\u2019s a reminder that predictions about AI are very difficult to make. A decade ago, the conventional wisdom was that AI would first impact physical labor, and then cognitive labor, and then maybe someday it could do creative work. It now looks like it\u2019s going to go in the opposite order.</p><p>5)\u00a0It\u2019s an example of a world in which good ideas are the limit for what we can do, not specific skills.</p><p>6) Although the upsides are great, the model is powerful enough that it's easy to imagine the downsides.</p><p>Hopefully this summer, we\u2019ll do a product launch and people will be able to use it for all sorts of things. We wanted to start with a research launch to figure out how to minimize the downsides in collaboration with a larger group of researchers and artists, and to give people some time to adapt to the change\u2014in general, we are believers in incremental deployment strategies. (Obviously the world already has Photoshop and we already know that images can be manipulated, for good and bad.)</p><p>        <div class=\"posthaven-gallery\" id=\"posthaven_gallery[1836338]\">\n                  <p class=\"posthaven-file posthaven-file-image posthaven-file-state-processed\">\n          <img class=\"posthaven-gallery-image\" src=\"https://phaven-prod.s3.amazonaws.com/files/image_part/asset/2779845/2_Q4jvs8eBye40c9jring6AP1eM/medium_Screen_Shot_2022-04-06_at_11.12.20_AM.png\" />\n        </p>\n\n        </div>\n\u00a0(A robot hand drawing, by DALL\u2022E)</p><p><br /></p></div>", "summary": "<div class=\"posthaven-post-body\"><p>Today we did a research launch of DALL\u2022E 2, a new AI tool that can create and edit images from natural language instructions.\u00a0</p><p>Most importantly, we hope people love the tool and find it useful. For me, it\u2019s the most delightful thing to play with we\u2019ve created so far. I find it to be creativity-enhancing, helpful for many different situations, and fun in a way I haven\u2019t felt from technology in a while.</p><p>But I also think it\u2019s noteworthy for a few reasons:</p><p>1)\u00a0This is another example of what I think is going to be a new computer interface trend: you say what you want in natural language or with contextual clues, and the computer does it. We offer this for code and now image generation; both of these will get a lot better. But the same trend will happen in new ways until eventually it works for complex tasks\u2014we can imagine an \u201cAI office worker\u201d that takes requests in natural language like a human does.</p><p>2)\u00a0It sure does seem to \u201cunderstand\u201d concepts at many levels and how they relate to each other in sophisticated ways.</p><p>3)\u00a0Copilot is a tool that helps coders be more productive, but still is very far from being able to create a full program. DALL\u2022E 2 is a tool that will help artists and illustrators be more creative, but it can also create a \u201ccomplete work\u201d. This may be an early example of the impact AI on labor markets. Although I firmly believe AI will create lots of new jobs, and make many existing jobs much better by doing the boring bits well, I think it\u2019s important to be honest that it\u2019s increasingly going to make some jobs not very relevant (like technology frequently does).</p><p>4)\u00a0It\u2019s a reminder that predictions about AI are very difficult to make. A decade ago, the conventional wisdom was that AI would first impact physical labor, and then cognitive labor, and then maybe someday it could do creative work. It now looks like it\u2019s going to go in the opposite order.</p><p>5)\u00a0It\u2019s an example of a world in which good ideas are the limit for what we can do, not specific skills.</p><p>6) Although the upsides are great, the model is powerful enough that it's easy to imagine the downsides.</p><p>Hopefully this summer, we\u2019ll do a product launch and people will be able to use it for all sorts of things. We wanted to start with a research launch to figure out how to minimize the downsides in collaboration with a larger group of researchers and artists, and to give people some time to adapt to the change\u2014in general, we are believers in incremental deployment strategies. (Obviously the world already has Photoshop and we already know that images can be manipulated, for good and bad.)</p><p>        <div class=\"posthaven-gallery\" id=\"posthaven_gallery[1836338]\">\n                  <p class=\"posthaven-file posthaven-file-image posthaven-file-state-processed\">\n          <img class=\"posthaven-gallery-image\" src=\"https://phaven-prod.s3.amazonaws.com/files/image_part/asset/2779845/2_Q4jvs8eBye40c9jring6AP1eM/medium_Screen_Shot_2022-04-06_at_11.12.20_AM.png\" />\n        </p>\n\n        </div>\n\u00a0(A robot hand drawing, by DALL\u2022E)</p><p><br /></p></div>", "image": null, "banner_image": null, "date_published": "2022-04-06T10:15:13Z", "date_modified": "2026-02-27T07:56:44Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1755859", "url": "https://blog.samaltman.com/helion", "external_url": null, "title": "Helion", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>I\u2019m delighted to be investing more in <a href=\"https://www.helionenergy.com/articles/announcing-500-million-fundraise/\">Helion</a>. Helion is by far the most promising approach to fusion I\u2019ve seen.<br /></p><p>David and Chris are two of the most impressive founders and builders (in the sense of building fusion machines, in addition to building companies!) I have ever met, and they have done something remarkable. When I first invested in them back in 2014, I was struck by the thoughtfulness of their plans about the scientific approach, the system design, cost optimizations, and the fuel cycle.</p><p>And now, with a tiny fraction of the money spent on other fusion efforts but the culture of a startup, they and their team have built a generator that produces electricity. <b>Helion has a clear path to net electricity by 2024, and has a long-term goal of delivering electricity for 1 cent per kilowatt-hour.</b> (!)</p><p>If this all works as we hope, we may have a path out of the climate crisis. Even though there are a lot of emissions that don\u2019t come from electrical generation, we\u2019d be able to use abundant energy to capture carbon and other greenhouses gases.</p><p>And if we have much cheaper energy than ever before, we can do things that are difficult to imagine today. The cost of energy is one of the fundamental inputs in the costs of so much else; dramatically cheaper energy will lead to dramatically better quality of life for many people.</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>I\u2019m delighted to be investing more in <a href=\"https://www.helionenergy.com/articles/announcing-500-million-fundraise/\">Helion</a>. Helion is by far the most promising approach to fusion I\u2019ve seen.<br /></p><p>David and Chris are two of the most impressive founders and builders (in the sense of building fusion machines, in addition to building companies!) I have ever met, and they have done something remarkable. When I first invested in them back in 2014, I was struck by the thoughtfulness of their plans about the scientific approach, the system design, cost optimizations, and the fuel cycle.</p><p>And now, with a tiny fraction of the money spent on other fusion efforts but the culture of a startup, they and their team have built a generator that produces electricity. <b>Helion has a clear path to net electricity by 2024, and has a long-term goal of delivering electricity for 1 cent per kilowatt-hour.</b> (!)</p><p>If this all works as we hope, we may have a path out of the climate crisis. Even though there are a lot of emissions that don\u2019t come from electrical generation, we\u2019d be able to use abundant energy to capture carbon and other greenhouses gases.</p><p>And if we have much cheaper energy than ever before, we can do things that are difficult to imagine today. The cost of energy is one of the fundamental inputs in the costs of so much else; dramatically cheaper energy will lead to dramatically better quality of life for many people.</p></div>", "image": null, "banner_image": null, "date_published": "2021-11-05T05:39:16Z", "date_modified": "2026-02-21T06:47:53Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1623396", "url": "https://blog.samaltman.com/the-strength-of-being-misunderstood", "external_url": null, "title": "The Strength of Being Misunderstood", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>A founder recently asked me how to stop caring what other people think. I didn\u2019t have an answer, and after reflecting on it more, I think it's the wrong question.</p><p>Almost everyone cares what someone thinks (though caring what everyone thinks is definitely a mistake), and it's probably important. Caring too much makes you a sheep. But you need to be at least a little in tune with others to do something useful for them.</p><p>It seems like there are two degrees of freedom: you can choose the people whose opinions you care about (and on what subjects), and you can choose the timescale you care about them on. Most people figure out the former [1] but the latter doesn\u2019t seem to get much attention.</p><p>The most impressive people I know care a lot about what people think, even people whose opinions they really shouldn\u2019t value (a surprising numbers of them do something like keeping a folder of screenshots of tweets from haters). But what makes them unusual is that they generally care about other people\u2019s opinions on a very long time horizon\u2014as long as the history books get it right, they take some pride in letting the newspapers get it wrong.\u00a0</p><p>You should trade being short-term low-status for being long-term high-status, which most people seem unwilling to do. A common way this happens is by eventually being right about an important but deeply non-consensus bet. But there are lots of other ways\u2013the key observation is that as long as you are right, being misunderstood by most people is a strength not a weakness. You and a small group of rebels get the space to solve an important problem that might otherwise not get solved.</p><p><br /></p><p>\u00a0</p><p>[1] In the memorable words of Coco Chanel, \u201cI don\u2019t care what you think about me. I don\u2019t think about you at all.\u201d</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>A founder recently asked me how to stop caring what other people think. I didn\u2019t have an answer, and after reflecting on it more, I think it's the wrong question.</p><p>Almost everyone cares what someone thinks (though caring what everyone thinks is definitely a mistake), and it's probably important. Caring too much makes you a sheep. But you need to be at least a little in tune with others to do something useful for them.</p><p>It seems like there are two degrees of freedom: you can choose the people whose opinions you care about (and on what subjects), and you can choose the timescale you care about them on. Most people figure out the former [1] but the latter doesn\u2019t seem to get much attention.</p><p>The most impressive people I know care a lot about what people think, even people whose opinions they really shouldn\u2019t value (a surprising numbers of them do something like keeping a folder of screenshots of tweets from haters). But what makes them unusual is that they generally care about other people\u2019s opinions on a very long time horizon\u2014as long as the history books get it right, they take some pride in letting the newspapers get it wrong.\u00a0</p><p>You should trade being short-term low-status for being long-term high-status, which most people seem unwilling to do. A common way this happens is by eventually being right about an important but deeply non-consensus bet. But there are lots of other ways\u2013the key observation is that as long as you are right, being misunderstood by most people is a strength not a weakness. You and a small group of rebels get the space to solve an important problem that might otherwise not get solved.</p><p><br /></p><p>\u00a0</p><p>[1] In the memorable words of Coco Chanel, \u201cI don\u2019t care what you think about me. I don\u2019t think about you at all.\u201d</p></div>", "image": null, "banner_image": null, "date_published": "2020-12-01T10:56:39Z", "date_modified": "2026-03-29T09:27:35Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1597376", "url": "https://blog.samaltman.com/pg-and-jessica", "external_url": null, "title": "PG and Jessica", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>A lot of people want to replicate YC in some other industry or some other place or with some other strategy. In general, people seem to assume that: 1) although there was some degree of mystery or luck about how YC got going, it can\u2019t be that hard, and 2) if you can get it off the ground, the network effects are self-sustaining.<br /></p><p></p><p>More YC-like things are good for the world; I generally try to be helpful. But almost none of them work. People are right about the self-sustaining part, but they can\u2019t figure out how to get something going.</p><p>The entire secret to YC getting going was PG and Jessica\u2014there was no other magic trick. A few times a year, I end up in a conversation at a party where someone tells a story about how much PG changed their life\u2014people speak with more gratitude than they do towards pretty much anyone else. Then everyone else agrees, YC founders and otherwise (non-YC founders might talk about an impactful essay or getting hired at a YC company). Jessica still sadly doesn\u2019t get nearly the same degree of public credit, but the people who were around the early days of YC know the real story.</p><p>What did they do? They took bets on unknown people and believed in them more than anyone had before. They set strong norms and fought back hard against bad behavior towards YC founders. They trusted their own convictions, were willing to do things their way, and were willing to be disliked by the existing power structures. They focused on the most important things, they worked hard, and they spent a huge amount of time 1:1 with people. They understood the value of community and long-term orientation. When YC was very small, it felt like a family.</p><p>Perhaps most importantly, they built an ecosystem (thanks to Joe Gebbia for pointing this out). This is easy to talk about but hard to do, because it requires not being greedy. YC has left a lot of money on the table; other people have made more money from the ecosystem than YC has itself. This has cemented YC\u2019s place\u2014the benefits to the partners, alumni, current batch founders, Hacker News readers, Demo Day investors, and everyone else around YC is a huge part of what makes it work.</p><p>I am not sure if any of this is particularly useful advice\u2014none of it sounds that hard, and yet in the 15 years since, it hasn\u2019t been close to replicated.</p><p>But it seems worth trying. I am pretty sure no one has had a bigger total impact on the careers of people in the startup industry over that time period than the two of them.</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>A lot of people want to replicate YC in some other industry or some other place or with some other strategy. In general, people seem to assume that: 1) although there was some degree of mystery or luck about how YC got going, it can\u2019t be that hard, and 2) if you can get it off the ground, the network effects are self-sustaining.<br /></p><p></p><p>More YC-like things are good for the world; I generally try to be helpful. But almost none of them work. People are right about the self-sustaining part, but they can\u2019t figure out how to get something going.</p><p>The entire secret to YC getting going was PG and Jessica\u2014there was no other magic trick. A few times a year, I end up in a conversation at a party where someone tells a story about how much PG changed their life\u2014people speak with more gratitude than they do towards pretty much anyone else. Then everyone else agrees, YC founders and otherwise (non-YC founders might talk about an impactful essay or getting hired at a YC company). Jessica still sadly doesn\u2019t get nearly the same degree of public credit, but the people who were around the early days of YC know the real story.</p><p>What did they do? They took bets on unknown people and believed in them more than anyone had before. They set strong norms and fought back hard against bad behavior towards YC founders. They trusted their own convictions, were willing to do things their way, and were willing to be disliked by the existing power structures. They focused on the most important things, they worked hard, and they spent a huge amount of time 1:1 with people. They understood the value of community and long-term orientation. When YC was very small, it felt like a family.</p><p>Perhaps most importantly, they built an ecosystem (thanks to Joe Gebbia for pointing this out). This is easy to talk about but hard to do, because it requires not being greedy. YC has left a lot of money on the table; other people have made more money from the ecosystem than YC has itself. This has cemented YC\u2019s place\u2014the benefits to the partners, alumni, current batch founders, Hacker News readers, Demo Day investors, and everyone else around YC is a huge part of what makes it work.</p><p>I am not sure if any of this is particularly useful advice\u2014none of it sounds that hard, and yet in the 15 years since, it hasn\u2019t been close to replicated.</p><p>But it seems worth trying. I am pretty sure no one has had a bigger total impact on the careers of people in the startup industry over that time period than the two of them.</p></div>", "image": null, "banner_image": null, "date_published": "2020-09-25T06:45:50Z", "date_modified": "2025-12-07T02:21:17Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1561774", "url": "https://blog.samaltman.com/researchers-and-founders", "external_url": null, "title": "Researchers and Founders", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>I spent many years working with founders and now I work with researchers.</p><p>Although there are always individual exceptions, on average it\u2019s surprising to me how different the best people in these groups are (including in some qualities that I had assumed were present in great people everywhere, like very high levels of self-belief).</p><p>So I\u2019ve been thinking about the ways they\u2019re the same, because maybe there is something to learn about qualities of really effective people in general.</p><p>The best people in both groups spend a lot of time reflecting on some version of the Hamming question\u2014\"what are the most important problems in your field, and why aren\u2019t you working on them?\u201d In general, no one reflects on this question enough, but the best people do it the most, and have the best \u2018problem taste\u2019, which is some combination of learning to think independently, reason about the future, and identify attack vectors. (This from John Schulman is worth reading:\u00a0<a href=\"http://joschu.net/blog/opinionated-guide-ml-research.html\" title=\"Link: http://joschu.net/blog/opinionated-guide-ml-research.html\">http://joschu.net/blog/opinionated-guide-ml-research.html</a>).</p><p>They have a laser focus on the next step in front of them combined with long-term vision. Most people only have one or the other.</p><p>They are extremely persistent and willing to work hard. As far as I can tell, there is no high-probability way to be very successful without this, and you should be suspicious of people who tell you otherwise unless you\u2019d be happy having their career (and be especially suspicious if they worked hard themselves).</p><p>They have a bias towards action and trying things, and they\u2019re clear-eyed and honest about what is working and what isn\u2019t (importantly, this goes both ways\u2014I\u2019m amazed by how many people will see something working and then not pursue it).\u00a0</p><p>They are creative idea-generators\u2014a lot of the ideas may be terrible, but there is never a shortage.</p><p>They really value autonomy and have a hard time with rules that they don\u2019t think make sense. They are definitely not lemmings.</p><p>Their motivations are often more complex than they seem\u2014specifically, they are frequently very driven by genuine curiosity.</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>I spent many years working with founders and now I work with researchers.</p><p>Although there are always individual exceptions, on average it\u2019s surprising to me how different the best people in these groups are (including in some qualities that I had assumed were present in great people everywhere, like very high levels of self-belief).</p><p>So I\u2019ve been thinking about the ways they\u2019re the same, because maybe there is something to learn about qualities of really effective people in general.</p><p>The best people in both groups spend a lot of time reflecting on some version of the Hamming question\u2014\"what are the most important problems in your field, and why aren\u2019t you working on them?\u201d In general, no one reflects on this question enough, but the best people do it the most, and have the best \u2018problem taste\u2019, which is some combination of learning to think independently, reason about the future, and identify attack vectors. (This from John Schulman is worth reading:\u00a0<a href=\"http://joschu.net/blog/opinionated-guide-ml-research.html\" title=\"Link: http://joschu.net/blog/opinionated-guide-ml-research.html\">http://joschu.net/blog/opinionated-guide-ml-research.html</a>).</p><p>They have a laser focus on the next step in front of them combined with long-term vision. Most people only have one or the other.</p><p>They are extremely persistent and willing to work hard. As far as I can tell, there is no high-probability way to be very successful without this, and you should be suspicious of people who tell you otherwise unless you\u2019d be happy having their career (and be especially suspicious if they worked hard themselves).</p><p>They have a bias towards action and trying things, and they\u2019re clear-eyed and honest about what is working and what isn\u2019t (importantly, this goes both ways\u2014I\u2019m amazed by how many people will see something working and then not pursue it).\u00a0</p><p>They are creative idea-generators\u2014a lot of the ideas may be terrible, but there is never a shortage.</p><p>They really value autonomy and have a hard time with rules that they don\u2019t think make sense. They are definitely not lemmings.</p><p>Their motivations are often more complex than they seem\u2014specifically, they are frequently very driven by genuine curiosity.</p></div>", "image": null, "banner_image": null, "date_published": "2020-06-19T09:39:12Z", "date_modified": "2026-03-12T00:35:43Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1560229", "url": "https://blog.samaltman.com/project-covalence", "external_url": null, "title": "Project Covalence", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>Almost every company and non-profit working on COVID-19 that I offered to help asked for support with clinical trials\u2014for companies focusing on developing novel drugs, vaccines, and diagnostics, rapidly spinning up trials is one of their biggest bottlenecks.\u00a0</p><p>Science remains the only way out of the COVID-19 crisis. Dramatically improving clinical trials, which are usually time-consuming and cost tens to hundreds of millions of dollars, is one of the highest-leverage ways to get out of it faster. \u00a0</p><p>The goal of this project, in collaboration with TrialSpark and Dr. Mark Fishman, is to offer much better clinical trial support to COVID-19 projects than anything that currently exists.</p><p>Project\u00a0Covalence\u2019s platform, powered by TrialSpark, is uniquely optimized to support COVID-19 trials, which are ideally run in community settings or at the patient\u2019s home to reduce the burden placed on hospitals and health systems. Project\u00a0Covalence\u00a0is well-positioned to tackle the operational and logistical challenges involved in launching such trials, and supports trial execution, 21 CFR Part 11 compliant remote data collection, telemedicine, biostatistics, sample kits for at-home specimen collection, and protocol writing.\u00a0</p><p>Researchers across academia and industry can leverage this shared infrastructure to rapidly launch their clinical trials. To facilitate coordination between studies, we will also be creating master protocols for platform studies to enable shared control arms and adaptive trial designs.</p><p>If you\u2019re interested in getting involved or have a trial that needs support, please get in touch at\u00a0<a href=\"mailto:ProjectCovalence@trialspark.com\" target=\"_blank\">ProjectCovalence@trialspark.com</a>\u00a0or visit\u00a0<a href=\"http://www.projectcovalence.com/\" target=\"_blank\">www.projectcovalence.com</a>.</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>Almost every company and non-profit working on COVID-19 that I offered to help asked for support with clinical trials\u2014for companies focusing on developing novel drugs, vaccines, and diagnostics, rapidly spinning up trials is one of their biggest bottlenecks.\u00a0</p><p>Science remains the only way out of the COVID-19 crisis. Dramatically improving clinical trials, which are usually time-consuming and cost tens to hundreds of millions of dollars, is one of the highest-leverage ways to get out of it faster. \u00a0</p><p>The goal of this project, in collaboration with TrialSpark and Dr. Mark Fishman, is to offer much better clinical trial support to COVID-19 projects than anything that currently exists.</p><p>Project\u00a0Covalence\u2019s platform, powered by TrialSpark, is uniquely optimized to support COVID-19 trials, which are ideally run in community settings or at the patient\u2019s home to reduce the burden placed on hospitals and health systems. Project\u00a0Covalence\u00a0is well-positioned to tackle the operational and logistical challenges involved in launching such trials, and supports trial execution, 21 CFR Part 11 compliant remote data collection, telemedicine, biostatistics, sample kits for at-home specimen collection, and protocol writing.\u00a0</p><p>Researchers across academia and industry can leverage this shared infrastructure to rapidly launch their clinical trials. To facilitate coordination between studies, we will also be creating master protocols for platform studies to enable shared control arms and adaptive trial designs.</p><p>If you\u2019re interested in getting involved or have a trial that needs support, please get in touch at\u00a0<a href=\"mailto:ProjectCovalence@trialspark.com\" target=\"_blank\">ProjectCovalence@trialspark.com</a>\u00a0or visit\u00a0<a href=\"http://www.projectcovalence.com/\" target=\"_blank\">www.projectcovalence.com</a>.</p></div>", "image": null, "banner_image": null, "date_published": "2020-06-16T09:08:44Z", "date_modified": "2026-02-22T22:43:33Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1551062", "url": "https://blog.samaltman.com/idea-generation", "external_url": null, "title": "Idea Generation", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>The most common question prospective startup founders ask is how to get ideas for startups. The second most common question is if you have any ideas for their startup.<br /></p><p>But giving founders an idea almost always doesn\u2019t work. Having ideas is among the most important qualities for a startup founder to have\u2014you will need to generate lots of new ideas in the course of running a startup.</p><p>YC once tried an experiment of funding seemingly good founders with no ideas. I think every company in this no-idea track failed. It turns out that good founders have lots of ideas about everything, so if you want to be a founder and can\u2019t get an idea for a company, you should probably work on getting good at idea generation first.</p><p>How do you do that?</p><p>It\u2019s important to be in the right kind of environment, and around the right kind of people. You want to be around people who have a good feel for the future, will entertain improbable plans, are optimistic, are smart in a creative way, and have a very high idea flux. These sorts of people tend to think without the constraints most people have, not have a lot of filters, and not care too much what other people think.\u00a0</p><p>The best ideas are fragile; most people don\u2019t even start talking about them at all because they sound silly. Perhaps most of all, you want to be around people who don\u2019t make you feel stupid for mentioning a bad idea, and who certainly never feel stupid for doing so themselves.</p><p>Stay away from people who are world-weary and belittle your ambitions. Unfortunately, this is most of the world. But they hold on to the past, and you want to live in the future.</p><p>You want to be able to project yourself 20 years into the future, and then think backwards from there. Trust yourself\u201420 years is a long time; it\u2019s ok if your ideas about it seem pretty radical.\u00a0</p><p>Another way to do this is to think about the most important tectonic shifts happening right now. How is the world changing in fundamental ways? Can you identify a leading edge of change and an opportunity that it unlocks? The mobile phone explosion from 2008-2012 is the most recent significant example of this\u2014we are overdue for another!</p><p>In such a tectonic shift, the world changes so fast that the big incumbents usually get beaten by fast-moving and focused startups. (By the way, it\u2019s useful to get good at differentiating between real trends and fake trends. A key differentiator is if the new platform is used a lot by a small number of people, or used a little by a lot of people.)</p><p>Any time you can think of something that is possible this year and wasn\u2019t possible last year, you should pay attention. You may have the seed of a great startup idea. This is especially true if next year will be too late.</p><p>When you can say \u201cI am sure this is going to happen, I\u2019m just not sure if we\u2019ll be the ones to do it\u201d, that\u2019s a good sign. Uber was like this for me\u2014after the first time I used it, it was clear we weren\u2019t going to be calling cabs for that much longer, but I wasn\u2019t sure that Uber was going to win the space.</p><p>A good question to ask yourself early in the process of thinking about an idea is \u201ccould this be huge if it worked?\u201d There are many good ideas in the world, but few of them have the inherent advantages that can make a startup massively successful. Most businesses don\u2019t generate a valuable accumulating advantage as they scale. Think early about why an idea might have that property. It\u2019s obvious for Facebook or Airbnb, but it often exists in more subtle ways.</p><p>It\u2019s also important to think about what you\u2019re well-suited for. This is hard to do with pure introspection; ideally you can ask a mentor or some people you\u2019ve worked with what you\u2019re particularly good at. I\u2019ve come to believe that founder/company fit is as important as product/market fit.</p><p>Finally, a good test for an idea is if you can articulate why most people think it\u2019s a bad idea, but you understand what makes it good.</p><p><br /></p><p><i>This is from my notes for a talk I gave at a YC event in China in 2018. Thanks to Eric Migicovsky for encouraging me to post it!</i><br /></p><p><i>I wrote it when I thought mostly about startups; now I think mostly about AI development. I am struck by how much of it applies, particularly paragraphs 5-9.</i></p></div>", "summary": "<div class=\"posthaven-post-body\"><p>The most common question prospective startup founders ask is how to get ideas for startups. The second most common question is if you have any ideas for their startup.<br /></p><p>But giving founders an idea almost always doesn\u2019t work. Having ideas is among the most important qualities for a startup founder to have\u2014you will need to generate lots of new ideas in the course of running a startup.</p><p>YC once tried an experiment of funding seemingly good founders with no ideas. I think every company in this no-idea track failed. It turns out that good founders have lots of ideas about everything, so if you want to be a founder and can\u2019t get an idea for a company, you should probably work on getting good at idea generation first.</p><p>How do you do that?</p><p>It\u2019s important to be in the right kind of environment, and around the right kind of people. You want to be around people who have a good feel for the future, will entertain improbable plans, are optimistic, are smart in a creative way, and have a very high idea flux. These sorts of people tend to think without the constraints most people have, not have a lot of filters, and not care too much what other people think.\u00a0</p><p>The best ideas are fragile; most people don\u2019t even start talking about them at all because they sound silly. Perhaps most of all, you want to be around people who don\u2019t make you feel stupid for mentioning a bad idea, and who certainly never feel stupid for doing so themselves.</p><p>Stay away from people who are world-weary and belittle your ambitions. Unfortunately, this is most of the world. But they hold on to the past, and you want to live in the future.</p><p>You want to be able to project yourself 20 years into the future, and then think backwards from there. Trust yourself\u201420 years is a long time; it\u2019s ok if your ideas about it seem pretty radical.\u00a0</p><p>Another way to do this is to think about the most important tectonic shifts happening right now. How is the world changing in fundamental ways? Can you identify a leading edge of change and an opportunity that it unlocks? The mobile phone explosion from 2008-2012 is the most recent significant example of this\u2014we are overdue for another!</p><p>In such a tectonic shift, the world changes so fast that the big incumbents usually get beaten by fast-moving and focused startups. (By the way, it\u2019s useful to get good at differentiating between real trends and fake trends. A key differentiator is if the new platform is used a lot by a small number of people, or used a little by a lot of people.)</p><p>Any time you can think of something that is possible this year and wasn\u2019t possible last year, you should pay attention. You may have the seed of a great startup idea. This is especially true if next year will be too late.</p><p>When you can say \u201cI am sure this is going to happen, I\u2019m just not sure if we\u2019ll be the ones to do it\u201d, that\u2019s a good sign. Uber was like this for me\u2014after the first time I used it, it was clear we weren\u2019t going to be calling cabs for that much longer, but I wasn\u2019t sure that Uber was going to win the space.</p><p>A good question to ask yourself early in the process of thinking about an idea is \u201ccould this be huge if it worked?\u201d There are many good ideas in the world, but few of them have the inherent advantages that can make a startup massively successful. Most businesses don\u2019t generate a valuable accumulating advantage as they scale. Think early about why an idea might have that property. It\u2019s obvious for Facebook or Airbnb, but it often exists in more subtle ways.</p><p>It\u2019s also important to think about what you\u2019re well-suited for. This is hard to do with pure introspection; ideally you can ask a mentor or some people you\u2019ve worked with what you\u2019re particularly good at. I\u2019ve come to believe that founder/company fit is as important as product/market fit.</p><p>Finally, a good test for an idea is if you can articulate why most people think it\u2019s a bad idea, but you understand what makes it good.</p><p><br /></p><p><i>This is from my notes for a talk I gave at a YC event in China in 2018. Thanks to Eric Migicovsky for encouraging me to post it!</i><br /></p><p><i>I wrote it when I thought mostly about startups; now I think mostly about AI development. I am struck by how much of it applies, particularly paragraphs 5-9.</i></p></div>", "image": null, "banner_image": null, "date_published": "2020-05-28T11:12:40Z", "date_modified": "2026-03-16T19:02:50Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1525506", "url": "https://blog.samaltman.com/please-fund-more-science", "external_url": null, "title": "Please Fund More Science", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>Experts on the COVID-19 pandemic seem to think there are three ways out\u2014that is, for life, health, and the economy to return roughly to normal.\u00a0</p><p>Either we get a vaccine good enough that R0 for the world goes below 1, a good enough treatment that people no longer need to be afraid, or we develop a great culture of testing, contract tracing, masks, and isolation.</p><p>I wish that the federal government were doing much more\u2014it would be great to see even a few percent of the recent stimulus bill go to funding R+D.\u00a0 But they don\u2019t seem to be funding enough science, and although I think concerns about the private sector and philanthropy doing what the government is supposed to be doing are somewhat valid, there isn\u2019t a great alternative right now.</p><p>On the positive side, I have never seen a field focused on one problem with such ferocity before.\u00a0 The response of biotech companies and research labs is amazing, and the speed they are operating at seems to have increased by more than 10x.\u00a0 It\u2019s the best of the spirit of innovation, and it\u2019s inspiring to see what these companies and research labs are doing.</p><p>Scientists can get us out of this.\u00a0 What they need are money and connections.</p><p>Investors and donors\u2014this is where we can help.\u00a0 Please consider shifting some of your focus and capital to scientific efforts addressing the pandemic.\u00a0 (And future pandemics too\u2014I think this will be a before-and-after moment in the world, and until we can defend against new viruses quickly, things are going to be different.)</p><p>The learning curve is quick, and there are a lot of experts willing to help you with diligence.\u00a0 It feels good to do something that might be useful, it\u2019s interesting to do something totally new, and it will make you more optimistic.</p><p>If you make it known to your network that you want to fund efforts working on COVID-19, you\u2019ll get flooded with opportunities.\u00a0 And it\u2019s always good to invest where the best founders are congregating.</p></div>", "summary": "<div class=\"posthaven-post-body\"><p>Experts on the COVID-19 pandemic seem to think there are three ways out\u2014that is, for life, health, and the economy to return roughly to normal.\u00a0</p><p>Either we get a vaccine good enough that R0 for the world goes below 1, a good enough treatment that people no longer need to be afraid, or we develop a great culture of testing, contract tracing, masks, and isolation.</p><p>I wish that the federal government were doing much more\u2014it would be great to see even a few percent of the recent stimulus bill go to funding R+D.\u00a0 But they don\u2019t seem to be funding enough science, and although I think concerns about the private sector and philanthropy doing what the government is supposed to be doing are somewhat valid, there isn\u2019t a great alternative right now.</p><p>On the positive side, I have never seen a field focused on one problem with such ferocity before.\u00a0 The response of biotech companies and research labs is amazing, and the speed they are operating at seems to have increased by more than 10x.\u00a0 It\u2019s the best of the spirit of innovation, and it\u2019s inspiring to see what these companies and research labs are doing.</p><p>Scientists can get us out of this.\u00a0 What they need are money and connections.</p><p>Investors and donors\u2014this is where we can help.\u00a0 Please consider shifting some of your focus and capital to scientific efforts addressing the pandemic.\u00a0 (And future pandemics too\u2014I think this will be a before-and-after moment in the world, and until we can defend against new viruses quickly, things are going to be different.)</p><p>The learning curve is quick, and there are a lot of experts willing to help you with diligence.\u00a0 It feels good to do something that might be useful, it\u2019s interesting to do something totally new, and it will make you more optimistic.</p><p>If you make it known to your network that you want to fund efforts working on COVID-19, you\u2019ll get flooded with opportunities.\u00a0 And it\u2019s always good to invest where the best founders are congregating.</p></div>", "image": null, "banner_image": null, "date_published": "2020-03-30T09:46:36Z", "date_modified": "2025-11-10T07:38:24Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}, {"id": "tag:blog.samaltman.com,2013:Post/1520123", "url": "https://blog.samaltman.com/funding-for-covid-19-projects", "external_url": null, "title": "Funding for COVID-19 Projects", "content_text": "", "content_html": "<div class=\"posthaven-post-body\"><p>I\u2019m trying to fund startups/projects helping with COVID-19, because it\u2019s basically the one thing I know how to do that can help. \u00a0I think we will soon have enough testing capacity, so now I\u2019d like to start funding more startups working on:</p><ol>\n<li>Producing a lot of ventilators or masks/gowns very quickly. \u00a0This will require a lot of repurposing and creativity but thankfully is an engineering problem not a scientific ones.</li>\n<li>Screening existing drugs for effectiveness.</li>\n<li>Novel approaches to vaccines (i.e., not doing what the big pharma companies are already doing).</li>\n<li>Novel therapeutics that the big pharma companies are unlikely to work on.</li>\n</ol><p>We tried this public <a href=\"https://docs.google.com/spreadsheets/d/1iH7r9crb3btOuO1y_NUPa1xvXAY6XlQ8Qc1hxhnSmQ8/edit?usp=sharing\" title=\"Link: https://docs.google.com/spreadsheets/d/1iH7r9crb3btOuO1y_NUPa1xvXAY6XlQ8Qc1hxhnSmQ8/edit?usp=sharing\">spreadsheet</a>\u00a0but it didn't work that well; please email me instead.</p><p>Also, if anyone knows of a contract research company that can run a viral challenge against SARS-CoV-2 in a humanized ACE2 animal model, that would help a startup I\u2019m working with. \u00a0Please reach out!</p><div>And of course, I think the best thing to do is still to get people to stay home.</div></div>", "summary": "<div class=\"posthaven-post-body\"><p>I\u2019m trying to fund startups/projects helping with COVID-19, because it\u2019s basically the one thing I know how to do that can help. \u00a0I think we will soon have enough testing capacity, so now I\u2019d like to start funding more startups working on:</p><ol>\n<li>Producing a lot of ventilators or masks/gowns very quickly. \u00a0This will require a lot of repurposing and creativity but thankfully is an engineering problem not a scientific ones.</li>\n<li>Screening existing drugs for effectiveness.</li>\n<li>Novel approaches to vaccines (i.e., not doing what the big pharma companies are already doing).</li>\n<li>Novel therapeutics that the big pharma companies are unlikely to work on.</li>\n</ol><p>We tried this public <a href=\"https://docs.google.com/spreadsheets/d/1iH7r9crb3btOuO1y_NUPa1xvXAY6XlQ8Qc1hxhnSmQ8/edit?usp=sharing\" title=\"Link: https://docs.google.com/spreadsheets/d/1iH7r9crb3btOuO1y_NUPa1xvXAY6XlQ8Qc1hxhnSmQ8/edit?usp=sharing\">spreadsheet</a>\u00a0but it didn't work that well; please email me instead.</p><p>Also, if anyone knows of a contract research company that can run a viral challenge against SARS-CoV-2 in a humanized ACE2 animal model, that would help a startup I\u2019m working with. \u00a0Please reach out!</p><div>And of course, I think the best thing to do is still to get people to stay home.</div></div>", "image": null, "banner_image": null, "date_published": "2020-03-15T08:31:23Z", "date_modified": "2025-11-10T07:38:27Z", "authors": ["Sam Altman"], "tags": [], "language": null, "attachments": ["\n                            {\"url\": string, \n                            'mime_type': string, \n                            'title': strinrg,\n                            'size_in_bytes': int,\n                            'duration_in_seconds': int\n                            "]}]}