New blog post from StackOverflow’s CEO Prashanth: Community is the future of AI

Please vote:

Throughout history, great thinkers have made predictions about how new technology would reshape the way in which humans work and live. With every paradigm shift, some jobs grow, some change, and some are lost. John Maynard Keynes wrote in 1930 that new technology meant humans would be working 30 hours a week or less, and that the main challenge would be what to do with all our free time. So far, predictions of this nature haven’t exactly come true. As new technology empowers us, we push ourselves to new heights and reach for previously unattainable goals.

Can we get a citation for that “30 hours a week or less”? NPR and The Guardian both say 15 hours, not 30. Additionally, assuming you are referring to this document written by Keynes, it says “Three-hour shifts or a fifteen-hour week may put off the
problem for a great while.” (page five).

Over nearly 15 years, Stack Overflow has built the largest online community for coders to exchange knowledge, a place where anyone with an internet connection can ask or answer questions, free of charge, and learn from their peers. Stack Overflow for Teams, our enterprise SaaS product, is trusted by over 15,000 organizations to serve as their internal knowledge bases. With the recent advent of dramatically improved artificial intelligence, many industries are wondering how technologies like ChatGPT will change their business. For software development, the answer seems more immediate than most. Even before the latest wave of AI, a third of the code being written on popular code repositories was authored by an AI assistant.

Again, according to what? I haven’t looked, but what specifically does “popular code repositories” actually mean? And how is this (un-named) source determining how much of the code is AI-assisted? Is it some statistic from a product like GitHub Copilot?

Today, sophisticated chatbots, built on top of cutting edge large language models (LLM), can write functional code for a website based on nothing more than a photo of a rough sketch drawn on a napkin. They can answer complex queries about how to build apps, help users to debug errors, and translate between different languages and frameworks in minutes. At Stack Overflow, we’ve had to sit down and ask ourselves some hard questions. What role do we have in the software community when users can ask a chatbot for help as easily as they can another person? How can our business adapt so that we continue to empower technologists to learn, share, and grow?

That’s more fair. I’ve seen the project about generating the website – I agree, it’s pretty interesting. At the same time, though, AI-generated code from chatbots has many limitations, perhaps most notably that it is often flat-out wrong, even if it sounds confident/correct. This is a large part of why SO and other SE sites have banned ChatGPT-generated answers.

It’s worth reflecting on an important property of technological progress. The Jevons Paradox shows us that, as innovation allows us to do more, we settle on a new normal, moving the goal posts for what we expect of people and organizations, then competing to see who can find new ways to pull ahead of the pack. For knowledge work, as the cost of an action diminishes, we often do more of it. Abstracting away repetitive or tedious tasks frees technologists up to make new discoveries or progress innovation.

Fair enough, that’s reasonable.

If new AI systems make it possible to create software simply by chatting with a computer, my prediction is that, far from the job of programmer disappearing, we’ll end up with millions of new software developers, as workers from fields like finance, education, and art begin making use of AI-powered tools that were previously inaccessible to them. We are enthusiastic about welcoming this next generation of developers and technologists, providing them with a community and with solutions, just as we have for the last 15 years. We’ve got a dedicated team working on adding GenAI to Stack Overflow and Stack Overflow for Teams and will have some exciting news to share this summer.

At risk of being harsh/rude, WHY?! Please no. A very large amount of work and effort has been put into finding and removing AI-content on SO, and the rest of the network (when it is banned on a site). ChatGPT answers are often wrong and not wanted. I cannot speak for others, but I certainly do not welcome that, nor am I excited for that in any way.

[image caption] Community members and AI must work together to share knowledge and solve problems

No, we certainly don’t. SO and SE has been solving problems and sharing knowledge for many years without ChatGPT. It is certainly not a “must”.

I’m not alone in thinking AI might lead to an explosion of new developers. I’ve heard similar sentiments expressed recently by Microsoft founder Bill Gates, by Geoff Hinton, the godfather of the neural network approach that produced today’s AI revolution, and by Stephen Wolfram, a pioneer across computer science and mathematics. Each sees in today’s AI the potential for the loss of certain jobs, yes, but also, if history is a guide, a future in which a great variety of more highly skilled work becomes available to an even larger group of people. Just as tractors made farmers more productive, we believe these new generative AI tools are something all developers will need to use if they want to remain competitive. Given that, we want to help democratize knowledge about these new AI technologies, ensuring that they are accessible to all, so that no developers are left behind.

Sure, some developers are excited about it. I also find it interesting. But when you say “we want to help democratize knowledge about these new AI technologies”, does that mean you want to share information about those technologies? I don’t have an issue with that, and doing so is good (assuming it is on-topic, of course)… but actually adding AI into SE is different. Can you clarify which you’re referring to?

I talk to developers of varying experience levels all of the time, and I’ve been hearing anecdotes of novice programmers building simple web apps with the help of AI. Most of these stories, however, don’t begin and end with an AI prompt. Rather, the AI provides a starting point and some initial momentum, and the human does additional research and learning to finish the job. The AI can debug some errors, but is stymied by others. It can suggest a good backend service, but often can’t solve all the points of friction that arise when integrating different services. And of course, when a problem is the result not of instructions from a machine, but human error, the best answers come from other people who have experienced the same issues.

Yes. Exactly. Sometimes, it does it correctly. Sometimes, it can’t. The best answers do come from other people, not chatbots.

For more experienced programmers, AI will be an amplifier of their existing skill, making them more ambitious in their projects. The result, as Jevons would predict, is that they spend more time with AI, but also more time creating new ideas, researching new topics, and asking new questions that had not occurred to them before. They feel empowered to reach farther beyond their traditional skillset and to push the boundaries in terms of the kind of work they want to take on.

I’m not convinced of your first sentence, but that’s just my personal viewpoint, though.

We are excited about what we can bring to the fast moving arena of generative AI. One problem with modern LLM systems is that they will provide incorrect answers with the same confidence as correct ones, and will “hallucinate” facts and figures if they feel it fits the pattern of the answer a user seeks. Grounding our responses in the knowledge base of over 50 million asked and answered questions on Stack Overflow (and proprietary knowledge within Stack Overflow for Teams) helps users to understand the provenance of the code they hope to use. We want to help coders stay in the flow state, allowing them to create with the latest tools with the confidence that they will be able to document and understand the provenance, source, and context of the code being generated.

That’s a good goal… but, how exactly will you “ground … responses in the knowledge base of 50 million asked and answered questions”? Specifically, how will you ensure only correct answers?

Community and reputation will also continue to be core to our efforts. If AI models are powerful because they were trained on open source or publicly available code, we want to craft models that reward the users who contribute and keep the knowledge base we all rely on open and growing, ensuring we remain the top destination for knowledge on new technologies in the future.

Yep, the community is very important. But could you elaborate on this a bit? Like, community members will get rewarded when a chatbot answer based on theirs is marked helpful? What exactly does this mean?

AI systems are, at their core, built upon the vast wealth of human knowledge and experiences. They learn by training on data – for example open-source code and Stack Overflow Q&A. It is precisely this symbiotic relationship between humans and AI that ensures the ongoing relevance of community-driven platforms like Stack Overflow. Allowing AI models to train on the data developers have created over the years, but not sharing the data and learnings from those models with the public in return, would lead to a tragedy of the commons. It might be in the self-interest of each developer to simply turn to the AI for a quick answer, but unless we all continue contributing knowledge back to a shared, public platform, we risk a world in which knowledge is centralized inside the black box of AI models that require users to pay in order to access their services.

I do somewhat agree with this, although I’m still not convinced of any benefit from adding AI content into SO, even if it is free.

As the AI landscape continues to evolve, the need for communities that can nurture, inform, and challenge these technologies becomes paramount. These platforms will not only offer the necessary guidance to refine AI algorithms and models but also serve as a space for healthy debate and exchange of ideas, fostering the spirit of innovation and pushing the boundaries of what AI can accomplish.

Yes. Innovation is good, as is progress. Productive discussion of those technologies is also good.

Our thesis on community as the center of a safe, productive, and open future for AI also offers some exciting prospects for our business. Stack Overflow for Teams, our enterprise, private version of Stack Overflow, helps to power a community-driven knowledge base inside of 15K+ organizations like Box, Microsoft, and Liberty Mutual. Decades of institutional knowledge, shaped and curated by subject matter experts and experienced teams, allows the employees at these organizations to more easily collaborate, improving productivity and trust.

Seems more like an ad for SO for Teams, so I’ll move onto the next paragraph.

Incorporating generative AI technologies into the organizations using Stack Overflow for Teams will allow us to layer a conversational interface on top of this wealth of information. We believe this could lead to tremendous productivity gains: from new hires being able to onboard more quickly, to speed up developer workflows, as users are able to quickly ask questions and retrieve answers tapping into the company’s history, documentation and Q&A.

Would it be limited for SO for Teams? Or would it go to actual SO? Not trying to be overly harsh, but can you elaborate on what information a chatbot would be able to give that wouldn’t be accessible by searching?

The example above is just one of many possible applications of GenAI to our Stack Overflow public platform and Stack Overflow for Teams, and they have energized everyone at our company. We’ll be working closely with our customers and community to find the right approach to this burgeoning new field and I’ve tasked a dedicated team to work full time on such GenAI applications. I’ll continue to share updates through channels such as my quarterly CEO blog, but I’ll be back in touch soon to announce something big on this topic. In the meantime, thank you to our community and customers for continuing to help us on our mission to empower the world to develop technology through collective knowledge.

NO please.

I do NOT want generative AI in the public SO. Also, is everyone at the company actually excited about this? I suppose maybe, but it seems like an exaggeration. Not going to name any names, but what about the previous statement (from a staff member, on MSE), that SE would “work internally progresses on identifying these posts and making our systems more resilient to issues like this in the future” (referring to ChatGPT-answers). That answer made me think SE was going to help communities deal with AI-answers. Can you clarify if that answer is still true?

Also, my answers on SE are intended for humans, not for training a machine learning model with.

As someone that has personally put in a large amount of effort and time into dealing with AI-generated answers on SE, I’m extremely disappointed to hear that you may incorporate generative AI models into SO.




Get The Latest Updates

Subscribe To Our Weekly Newsletter

No spam, notifications only about new products, updates.
On Key

Related Posts

Anduril’s New Drone Killer Is Locked on to AI-Powered Warfare

Anduril’s New Drone Killer Is Locked on to AI-Powered Warfare

After Palmer Luckey founded Anduril in 2017, he promised it would be a new kind of defense contractor, inspired by hacker ingenuity and Silicon Valley speed. The company’s latest product, a jet-powered, AI-controlled combat drone called Roadrunner, is inspired by the grim reality of modern conflict, especially in Ukraine, where large numbers of cheap, agile

Read More »