Timeline of AI in programming

From Timelines
Jump to navigation Jump to search

This is a timeline of AI in programming.

Sample questions

The following are some interesting questions that can be answered by reading this timeline:

Big picture

Years Period Main AI Paradigm Influence on Programming
1950s–1980s Symbolic AI Logic, rule-based systems, expert systems, formal semantics, automated theorem proving This period establishes many foundations of programming theory. Logic-based languages (like Lisp and Prolog) influenced functional and declarative programming. Automated reasoning contributed to early program verification and compiler correctness. Expert systems demonstrated that knowledge encoding could guide code-generation templates and domain-specific automation. Symbolic approaches shaped thinking about abstraction, recursion, and problem decomposition that still defines modern programming practice.
1990s–2010s Statistical AI Machine learning, probabilistic models, Bayesian networks, early neural nets Programming tools shift from hand-crafted rules to pattern-recognition systems. Enabled probabilistic bug detection, anomaly detection in large-scale systems, and early statistical autocomplete (n-gram models). Introduced ML-based static analysis and refactoring suggestions. Helped shape data-driven software engineering practices and influenced compiler heuristics, program optimization, and predictive modeling of developer behavior. Created the first bridge between code as formal logic and code as statistical signal.
2014–2020 Deep Learning for Code Deep neural networks (RNNs, CNNs), transformer-based code models, code embeddings, graph neural networks This period marks the first major leap in AI systems that understand code structure. Embeddings capture semantic relationships between identifiers, types, and functions. Tools like Code2Vec[1], CodeBERT[2], and sequence-to-sequence models enable code summarization, docstring generation, neural code search, API recommendation, and clone detection. Deep learning begins outperforming traditional symbolic/static analysis in several tasks. Neural program synthesis moves from theoretical curiosity to practical utility.
2021–present LLM Era Large language models, instruction-tuned transformers, retrieval-augmented generation, multimodal AI AI becomes a real-time programming assistant capable of generating, debugging, refactoring, explaining, and documenting code at scale. Natural language becomes a valid interface for software creation. LLM-driven tools reshape the entire development workflow—automated test generation, design reasoning, code review, dependency management, and system exploration. Integrated into IDEs, CI/CD, and documentation pipelines. Creates new paradigms such as AI pair-programming, AI agents executing coding tasks, and semi-autonomous codebases. Drives productivity boosts and raises new concerns around reliability, security, licensing, and software engineering norms.

Full timeline

Year AI subfield Area affected Event type Event description
1950 Theoretical Foundations General Programming Theoretical Development English mathematician Alan Turing publishes Computing Machinery and Intelligence, a foundational paper on artificial intelligence. Turing reframes the vague question “Can machines think?” by proposing the Imitation Game, later known as the Turing test, which evaluates whether a machine can converse indistinguishably from a human. In this setup, a human judge communicates with both a human and a computer; if the judge cannot reliably tell them apart, the machine is said to succeed. Turing shifts the debate from defining “thinking” to assessing observable performance. The paper would deeply influence AI philosophy, provoking extensive discussion and criticism.[3][4]
1951 Early AI / Machine Learning Programming concepts Milestone British computer scientist Christopher Strachey develops a Checkers (Draughts) program, demonstrating that machines can implement rule-based logic for game-playing. One of the earliest video games and the first written for a general-purpose computer, it runs on the Ferranti Mark 1 at the University of Manchester and may have been the first to display visuals on an electronic screen. The game allows a player to face a simple AI.[5][6][7]
1955 Symbolic AI Software development methods Milestone The term “artificial intelligence” is introduced in the proposal A Proposal for the Dartmouth Summer Research Project on Artificial Intelligence, written by John McCarthy, Marvin Minsky, Nathaniel Rochester, and Claude Shannon. Their document outlines a plan to investigate how machines can simulate human intelligence and formally names the emerging area of study. The proposed workshop is held in the summer of 1956 at Dartmouth College, during July and August. This meeting is widely regarded as the official birth of the AI field.[8]
1958 Symbolic AI Programming languages Innovation American John McCarthy develops LISP, the first programming language explicitly designed for artificial intelligence. Building on the earlier Information Processing Language (IPL), he simplifies its complexity by incorporating ideas from lambda calculus, enabling clearer symbolic computation. LISP introduces major innovations such as recursion, garbage collection, dynamic typing, and homoiconicity, allowing code and data to share the same structure. Its interpreter evaluates expressions through symbols, associations, and functions, supporting flexible definition of new operations. LISP quickly becomes the dominant AI language in the United States and would shape decades of symbolic AI research.[9]
1965 Expert Systems Software problem-solving Milestone The Dendral project launches at Stanford University as one of the first landmark efforts in artificial intelligence. Developed by Edward Feigenbaum, Bruce G. Buchanan, Joshua Lederberg, and Carl Djerassi, it aims to model scientific hypothesis formation by helping chemists identify unknown organic molecules from mass spectrometry data. By encoding expert chemical rules, Dendral generates plausible molecular structures and becomes the first true expert system, proving that computers can mimic specialized human reasoning. Its success heavily influences later rule-based systems such as MYCIN and shapes early approaches to AI-driven problem solving across multiple scientific fields.[10][11][12][13]
1966 Robotics Control Systems Robot Development Research at the Stanford Research Institute produces Shakey, a groundbreaking mobile robot that can reason about its actions. Initially a boxy machine on wheels with bump sensors, a TV camera, and a range finder, Shakey communicates first by cable and later by wireless link to larger computers. It combines vision, reasoning, planning, and action, even accepting simple English commands. Shakey’s major achievement is executing high-level, non-step-by-step instructions by generating its own plans and adapting to obstacles. Demonstrations show it moving blocks, navigating rooms, and adjusting to surprises. Shakey becomes a landmark in AI and robotics history.[14]
1966 Natural Language Processing (NLP) Human-Computer Interaction Program Creation German-American computer scientist Joseph Weizenbaum creates ELIZA, a pioneering chatbot that uses simple pattern-matching rules to simulate conversation, famously mimicking a psychotherapist. Its convincing interactions show how readily humans attribute understanding to machines. Disturbed by this reaction, Weizenbaum warns that computational responses should not be mistaken for genuine thought. He would later criticize AI for its potential to dehumanize and reinforce social inequalities. Written in Lisp, ELIZA becomes a foundational milestone in natural language processing and human–computer interaction, illustrating how scripted dialogue systems could appear intelligent despite lacking real comprehension.[15]
1970 Knowledge Representation Programming paradigms Innovation Development of production systems (IF-THEN rules) changes how programs encode knowledge for decision-making.
1972 Symbolic AI Software engineering Language creation Prolog emerges as a new logic-programming language developed by Alain Colmerauer and Philippe Roussel in Marseille, building on earlier natural-language processing research. Working with Robert Kowalski, they combine theorem-proving concepts with linguistic goals to create a language based on formal logic. Roussel builds the first interpreter, and David Warren later creates the first compiler, shaping the influential Edinburgh syntax. Prolog quickly spreads through Europe and Japan, powering major AI initiatives such as the Fifth Generation Computer Systems project. Its logical paradigm, rooted in automated reasoning and symbolic inference, makes Prolog a foundational language in AI and computational linguistics.[16][17][18][19]
1972 Expert Systems Code Generation Tool The MYCIN project begins at Stanford University as one of the earliest and most influential medical expert systems. Designed to diagnose bacterial blood infections, MYCIN analyzes patient symptoms and laboratory results, asks follow-up questions, and recommends appropriate antibiotic treatments. The system uses around 500 production rules and can explain the reasoning behind its conclusions, a key innovation in explainable AI. MYCIN performs at a level comparable to medical specialists and often better than general practitioners. Its rule-based structure becomes a foundational model for later expert systems and significantly shapes early AI and software engineering methodologies.[20]
1978 Expert Systems Enterprise software Milestone John P. McDermott of Carnegie Mellon University develops XCON, also known as R1, a production-rule-based expert system written in OPS5 to automate the configuration of Digital Equipment Corporation’s VAX computer systems. Faced with millions of possible system configurations, DEC had relied on skilled technical editors for manual validation, a slow process. XCON automates component selection, cabling, and technical checks, assisting sales, manufacturing, and field service. By integrating order validation and assembly guidance, it reduces shipping times from weeks to days, improved accuracy, lowered costs, and increased customer satisfaction. XCON exemplifies the commercial potential of AI in automating complex, rule-based business and engineering tasks.[21][22]
1982 Neural Networks Programming tools Algorithm development American social scientist and machine learning pioneer Paul Werbos applies the backpropagation algorithm to multilayer perceptrons (MLPs), establishing the standard approach for training neural networks. This method, originally introduced by Seppo Linnainmaa in 1970[23] as the "reverse mode of automatic differentiation," enables efficient calculation of gradients for complex networks. Werbos’s adaptation revitalizes neural network programming, allowing models to learn from data effectively. The success of backpropagation leads to its widespread adoption and the development of early neural network toolkits, initially in Lisp and later in C and Fortran, laying the groundwork for modern machine learning and deep learning research.[24][25]
1984 Expert Systems Development Tools Tool Programmer's Apprentice project at MIT explored using AI to assist in software development and debugging.
1980s Expert Systems Software development practices Tooling Expert-system shells (e.g., CLIPS, OPS5) become common; programmers start building rule-based systems using forward/backward chaining engines instead of traditional procedural code.
1990 Machine Learning Bug Detection Research Early research into using ML for automatic bug detection and program analysis begins.
1991 Machine Learning Programming languages Language creation Python 0.9.0 released by Guido van Rossum – designed for readability and rapid prototyping; quickly becomes the dominant language for AI/ML research.
1990s Machine Learning Code generation & optimization Development AI techniques begin to optimize compilers and automate code refactoring using learned heuristics.
1995 NLP Documentation Tool First automated documentation generation tools using natural language processing emerge.
1997 Reinforcement Learning Algorithm design Milestone Deep Blue defeats Garry Kasparov, showcasing AI’s ability to program and optimize strategies in complex environments.
1997 Search & Game AI Algorithmic programming Milestone Deep Blue defeats Garry Kasparov; the program (written in C/C++) uses massive parallel alpha-beta search, influencing high-performance and concurrent programming techniques.
1997 Machine Learning General Programming Milestone Achievement IBM's Deep Blue defeats world chess champion Garry Kasparov, showcasing AI's strategic capabilities.
1997 Search Algorithms Problem Solving Milestone IBM's Deep Blue supercomputer defeats world chess champion Garry Kasparov, showcasing the power of advanced search algorithms and brute-force computation in solving complex problems.
2001 Machine Learning Code Completion Tool IntelliSense and similar intelligent code completion tools gain widespread adoption in IDEs.
2005 Data Mining Testing Research Research on mining software repositories to predict bugs and improve testing strategies.
2006 Deep Learning Numerical computing Library Release of Torch (Lua-based); early deep-learning framework that influenced later frameworks.
2000s Data Mining & ML Software development Application AI-driven tools assist programmers in code analysis, bug detection, and automated testing.
2000s Big Data Data Processing Technological Advancement AI techniques are increasingly used for processing and analyzing large datasets.
2009–2012 Deep Learning GPU programming Paradigm shift CUDA + early neural-net libraries (e.g., Theano 2009, Caffe 2013) make GPU-accelerated deep learning practical; programmers shift from CPU-only to massive parallel computing.
2010 Machine Learning Code Search Tool Google's Code Search and similar tools use ML algorithms to improve code discovery and navigation.
2012 Deep Learning Research Foundation Foundation AlexNet breakthrough in deep learning creates foundation for future AI programming tools.
2010s Deep Learning Parallel Computing/Tools Hardware/Software Integration The widespread adoption of Deep Learning was made possible by the use of GPUs for parallel processing, accelerating the training of large neural networks and demanding specialized programming libraries (e.g., TensorFlow, PyTorch).
2010s Deep Learning Image/Voice Recognition Breakthrough Deep learning revolutionizes image and voice recognition, enabling advancements in AI applications.
2011 Natural Language Processing Programming assistance Innovation IBM Watson demonstrates NLP-powered reasoning, inspiring AI code assistants that understand human-like queries.
2011 Deep Learning Machine-learning tooling Library scikit-learn released – brings accessible classical ML to Python programmers.
2014 Generative Models Content Creation Model Development Generative Adversarial Networks (GANs) are introduced, enabling AI to generate realistic images and content.
2014 NLP Code Generation Research Sequence-to-sequence models show promise for translating natural language to code.
2015 Deep Learning Framework revolution Library TensorFlow (Google) and PyTorch prototype (Facebook) appear; dynamic computation graphs and autograd radically simplify writing and debugging neural networks.
2016 Deep Learning Bug Detection Tool DeepBugs and similar neural approaches to bug detection published, using deep learning on code.
2016 Deep Learning Code generation Development AI models like DeepCoder begin generating code snippets automatically from problem descriptions.
2016 Reinforcement Learning Game Development Milestone Achievement AlphaGo defeats world champion Lee Sedol in Go, demonstrating advanced AI decision-making.
2017 Large Language Models (LLMs) Code Generation/Refactoring Architecture Innovation The introduction of the Transformer architecture and its subsequent use in Large Language Models (LLMs) revolutionized NLP and led to the ability to generate, summarize, and correct complex code.
2017 Deep Learning Differentiation & training Core technique Autodifferentiation becomes ubiquitous (PyTorch, TensorFlow eager, JAX); programmers no longer hand-write gradients – transformative for research speed.
2018 NLP Code Completion Tool TabNine introduces GPT-2 based code completion, marking shift toward transformer-based tools.
2018 (March 26) Code2Vec is introduced as a neural framework designed to generate fixed-length distributed vector representations of code snippets for semantic prediction tasks. The approach decomposes each snippet into a set of abstract-syntax-tree paths and jointly learns representations for individual paths and their aggregation. Trained on a corpus of 14 million methods, the model demonstrates the ability to infer method names from previously unseen files and produces vector embeddings that reflect semantic similarity and analogical structure. Evaluated against prior techniques on the same dataset, it achieves a relative performance improvement exceeding 75%.[1]
2018 Machine Learning / NLP Programming productivity Innovation GitHub Copilot precursor models show AI can suggest code completions and assist developers in writing software faster.
2018 Reinforcement Learning Code generation & automation Tool OpenAI’s early work on AI that writes simple code (using RL-trained models) begins.
2019 Natural Language Processing Programming assistance Tool GitHub releases Copilot prototype (based on OpenAI Codex, a GPT-3 descendant); first widely used AI pair programmer.
2019 Deep Learning Code Understanding Research Microsoft releases CodeBERT, pre-trained model for programming and natural language understanding.
2020 (February 19) CodeBERT is introduced as a bimodal pre-trained model designed to learn joint representations of programming languages (PL) and natural language (NL) to support downstream tasks such as code search and code documentation generation. Built on a Transformer architecture, it uses a hybrid objective combining masked language modeling with replaced token detection, enabling effective use of both NL–PL paired data and unimodal code data. When fine-tuned, CodeBERT achieves state-of-the-art results on NL-based code search and documentation generation. Zero-shot probing further shows that CodeBERT captures NL-PL relationships better than earlier pre-trained models.[2]
2020 Large Language Models Development environments Commercial tool GitHub Copilot (powered by Codex) officially launched – mainstream adoption of AI-assisted coding.
2020 NLP Code Search Research GitHub's semantic code search uses neural networks to understand code meaning, not just syntax.
2020s AI in Software Development Code Generation Tool Development AI-powered tools like GitHub Copilot assist developers by suggesting code snippets and automating repetitive tasks.
2021 Large Language Models Programming productivity Milestone OpenAI Codex launches, enabling AI-assisted coding in multiple languages and transforming programming workflows.
2021 Large Language Models Multi-language Research OpenAI Codex demonstrates strong performance across multiple programming languages and tasks.
2021 Large Language Models Code understanding & generation Model OpenAI releases Codex as API; AlphaCode (DeepMind) shows competitive-programming performance.
2021 Large Language Models Code Generation Product GitHub Copilot launches, powered by OpenAI Codex, providing AI pair programming at scale.
2022 Large Language Models Natural Language Processing Model Release Large language models like GPT-3 demonstrate advanced text generation and understanding capabilities.
2022 Large Language Models Full-stack development Tool Release of ChatGPT (Nov 2022) and Copilot Chat; developers start using conversational AI for debugging, documentation, and entire feature implementation.
2022 Large Language Models Code Explanation Product ChatGPT released, widely adopted for code explanation, debugging, and learning programming.
2022 Large Language Models Code Generation Research AlphaCode by DeepMind achieves competitive programming performance at Codeforces competitions.
2023 Generative AI Software engineering Development Advanced AI systems can generate, debug, and refactor entire programs, bridging natural language instructions with executable code.
2023 Large Language Models IDE integration Ecosystem Copilot X, CodeWhisperer, Tabnine, Cursor, and dozens of IDE plugins; AI becomes a standard part of most programmers’ workflow.
2023 (May 21) Efficiency Rodney Brooks, a robotics researcher and AI expert, argues that large language models like OpenAI's ChatGPT are not as intelligent as people believe and are far from being able to compete with humans on an intellectual level. Brooks highlights that these models lack an underlying understanding of the world and merely exhibit correlations in language. Current language models can sound like they understand, but they lack the ability to logically infer meaning, leading to potential misinterpretations. Brooks emphasizes that these models are good at generating answers that sound right but may not be accurate. He shares his experience of relying on large language models for coding tasks and finding that they often provide confidently wrong answers. Brooks concludes that while future iterations of AI may bring interesting advancements, they are unlikely to achieve artificial general intelligence (AGI).[26]
2023 (October 17) Natural language processing (NLP); educational AI Programming education (Java) Empirical study A study presents preliminary findings on how students interact with AI tools like ChatGPT and GitHub Copilot in introductory Java programming courses. Using a mixed-method design—including quizzes, programming tasks under different support conditions, and interviews—the study highlights the diverse attitudes and behaviors students display toward AI assistance. While tools like ChatGPT offer flexibility and reduce hesitation in seeking help, concerns remain about their impact on developing core programming skills. The findings offer valuable insights for integrating AI in education responsibly.[27]
2023 AI Ethics Regulation and Ethics Regulatory Development Increased focus on ethical AI, leading to regulations and guidelines for responsible AI development and use.
2023 LLMs Testing Tool AI test generation tools become mainstream, automatically creating unit tests from code.
2023 (December 31) Machine learning; deep learning; natural language processing (NLP); expert systems Software engineering lifecycle Systematic literature review An article systematically reviews 110 studies to assess how AI has been integrated into software engineering over the past decade. It highlights the widespread application of AI techniques—especially machine learning, deep learning, natural language processing, optimization algorithms, and expert systems—across all phases of the software development life cycle. Key benefits include improved defect prediction, code recommendation, automated requirement analysis, and maintenance precision. The review emphasizes the need for interpretable and ethical AI tools to ensure responsible advancement in software engineering.[28]
2023 LLMs Security Tool AI-powered security scanning tools emerge, using LLMs to detect vulnerabilities in code.
2023 Large Language Models Full Development Product GPT-4 demonstrates advanced coding capabilities including architecture design and complex debugging.
2024 (February 6) Machine learning AI systems development, applied software engineering Research publication A paper examines the rapid progress and societal implications of AI and machine learning (ML). It outlines AI’s core capabilities—such as learning, problem-solving, and decision-making—and ML’s role in enabling systems to improve through data analysis. The paper explores real-world applications including natural language processing, image and speech recognition, and autonomous vehicles. It also addresses potential risks, such as job displacement and misuse of technology. Emphasizing the importance of ethics, the study advocates for responsible AI development to balance innovation with minimizing harm to individuals and society.[29]
2024 (March 22) Generative artificial intelligence The article explores whether artificial intelligence will replace programmers, concluding that AI will augment rather than eliminate programming roles. Instructors Norman McEntire and James Gappy from UC San Diego Extended Studies explain that generative AI, despite its power to automate coding, debugging, and optimization, still relies on human oversight, creativity, and technical understanding. They emphasize the importance of mastering fundamentals, using AI as a collaborator, and maintaining continuous learning to stay relevant. Programmers who effectively integrate AI tools into their workflow will be more productive, adaptable, and valuable. Ultimately, AI is framed as an assistant—not a replacement—for coders.[30]
2024 (May 9) Software engineering with AI assistance; machine learning AI-assisted programming and code collaboration Research publication An article examines the use of AI-pair programming—collaborative coding between human developers and AI assistants—at TiMi Studio, a prominent game development company. Analyzing data from code repositories, reviews, surveys, and interviews, the study finds that AI-pair programming enhances code quality and developer satisfaction. Benefits include time-saving, error reduction, skill development, and better feedback. However, challenges such as trust issues, lack of explainability, and reduced autonomy also emerge. The paper offers practical insights for optimizing AI-pair programming in real-world software development environments.[31]
2024 Multi-modal AI UI Development Tool AI tools that convert designs and screenshots directly to code become commercially viable.
2024 LLMs Code Review Tool AI-powered code review tools integrated into CI/CD pipelines, providing automated feedback on PRs.
2024 (June 16) Large language models (LLMs); natural language processing; code generation Experimental study An article examines how large language models (LLMs) like GPT and Codex affect programmer productivity and behavior. In a study with 24 participants completing Python tasks, researchers compare three setups: GitHub Copilot (auto-complete), GPT-3 (conversational), and traditional tools (web browser). Results show that AI-assisted coding significantly boosts productivity and alters coding strategies. The study highlights how interaction design (autocomplete vs. conversational) influences user engagement and problem-solving approaches. Overall, the research underscores the transformative impact of LLMs on programming and the need to optimize their integration in development workflows.[32]
2024 (September 12) A study by economists from MIT, Princeton, and the University of Pennsylvania find that AI coding assistants like GitHub Copilot boost developer productivity by 26% in enterprise environments. Analyzing data from 4,800 developers at Microsoft, Accenture, and another Fortune 100 firm, the research shows a 13.5% rise in code commits and a 38.4% increase in compilation frequency, with no decline in code quality. Junior developers benefit most, improving output by up to 40%. The study emphasizes gradual adoption, training, and governance as key to maximizing AI’s benefits while avoiding overreliance and integration challenges.[33]
2024 (October 5) Large language models (LLMs); code generation A study investigates the impact of AI coding tools on novice programming education in a first-semester course with 73 engineering student teams over 12 weeks. Using surveys and qualitative reports, it finds that AI tool familiarity rose from 28% to 100%, with increasing student satisfaction. Students primarily used AI for writing code comments (91.7%), debugging (80.2%), and information seeking (68.5%). The tools enhanced learning and improved the perceived real-world relevance of programming. However, concerns emerged regarding potential cheating, over-reliance on AI, and weaker grasp of core programming concepts, highlighting the need for balanced and guided AI integration in education.[34]
2024 Agentic AI Full Development Product AI coding agents like Devin, Cursor, and others emerge, capable of autonomous software development tasks.
2024 (November 25) Applied artificial intelligence; machine learning; software engineering automation Software development life cycle Academic publication An article examines how AI is transforming the software development life cycle. It highlights AI’s applications in areas such as design, coding, testing, project management, and maintenance, emphasizing its role in automating tasks, improving efficiency, and enhancing code quality. The paper also discusses key challenges, including over-reliance on AI tools, ethical dilemmas, and security issues. Looking ahead, it explores emerging trends like adaptive systems, AI-enhanced team collaboration, and fully automated software development. Overall, the study underscores AI’s profound and growing influence on the future of software engineering.[35]
2024 (December 3) Generative artificial intelligence Programming education and AI-assisted learning Academic publication A study evaluates the impact of the GenAI Gemini tool on programming education in a polytechnic university in Guayaquil, Ecuador. Using a quantitative, quasi-experimental design, it finds that AI integration significantly enhances student motivation, interest, and satisfaction. Notably, 91% of students report increased enthusiasm for programming, and 90% feel their learning expectations were met or exceeded. The research highlights GenAI's potential to transform teaching but stresses the importance of proper educator training, ethical guidance for students, sustained engagement, and curriculum alignment to harness its full benefits.[36]
2024 (December 8) Educational artificial intelligence; intelligent tutoring systems; learning analytics Programming education policy and strategy Academic publication A study reviews the role of AI in transforming education. It highlights AI’s growing application in areas like intelligent tutoring, automated grading, and learning analytics, driven by the need for personalized learning. While acknowledging various challenges and limitations, the study emphasizes AI’s potential to create more efficient and intelligent education systems. Programming education is identified as especially crucial, fostering students’ logical thinking, creativity, and social engagement. The paper proposes strategic guidance for integrating AI in education and underscores its relevance for shaping future talent and educational policy.[37]
2024 (December 23) Generative artificial intelligence Software development, AI-assisted programming Academic publication An article envisions how AI will reshape software engineering by the end of the decade. It contrasts current AI-assisted tools like GitHub Copilot and ChatGPT with projected advancements, forecasting a shift in developers’ roles—from manual coders to coordinators of AI-driven ecosystems. The study introduces the concept of HyperAssistant, a future AI tool designed to enhance coding, debugging, collaboration, and even mental health support. Rather than replacing developers, AI is seen as a powerful partner, enhancing software quality, efficiency, and creativity in a transformed development landscape.[38]
2025 (February 20) Commentary A New York Times article argues that generative AI is transforming, rather than replacing, software developers. Tools like GitHub Copilot now assist with debugging, documentation, and translation, improving productivity by up to 30%. While entry-level hiring has weakened, demand for experienced developers and AI literacy is rising. Experts predict AI will automate most code writing, shifting programmers’ roles toward design, oversight, and creative problem-solving. Training programs are adapting at the time, emphasizing core computer science, critical thinking, and the ability to guide AI-driven development.[39]
2025 (March 24) Commentary An article by Adlene Sifi explores how generative AI, particularly tools like GitHub Copilot, enhances developer experience (DevEx)—the overall satisfaction, productivity, and well-being of software developers. It explains that DevEx depends on company culture, processes, collaboration, and tools, and can be improved through faster feedback loops, lower cognitive load, and better flow states.[40]
2025 (May) A study examines how 231 students in an “Object-Oriented Programming” course use AI chatbots like ChatGPT and how this relates to their academic performance. The study concludes that most students use AI for debugging and code comprehension, but few rely on it weekly, indicating limited dependency. Students value AI’s speed but criticize its errors and inconsistencies. The study finds a negative correlation between frequent AI use and grades, suggesting weaker students depend more on AI tools. Researchers conclude that unstructured AI use may hinder learning and urge educators to guide critical, reflective integration of AI into coursework.[41]
2025 (June 5) A Coursera article concludes that AI will not replace programmers in the near future, though it is reshaping their work. Generative AI tools can automate repetitive coding, assist with debugging, documentation, and forecasting, but still lack creativity, critical thinking, and reliability. These limitations—such as hallucinated code, security, and copyright risks—mean human oversight remains essential. According to the article, AI may reduce entry-level positions but create new roles in AI development and supervision. Long-term replacement is constrained by trust and societal acceptance. Programmers can future-proof their careers by mastering AI, ML, prompt engineering, and related technologies.[42]
2025 (July 10) A study by the AI research nonprofit METR finds that advanced AI coding assistants can slow down experienced software developers rather than accelerate their work. In experiments using the tool Cursor on familiar open-source projects, seasoned programmers complete tasks 19% slower when aided by AI. Participants had expected a 24% speedup and still believe they worked faster, despite results showing otherwise. Researchers express surprise, noting they had predicted a “2x speed up.” The findings question assumptions that AI consistently boosts productivity and highlight challenges in human–AI collaboration in software development.[43]
2025 (August 9) A Reuters investigation finds that artificial intelligence is accelerating the decline of coding bootcamps, once a key entry point into software engineering. As AI tools automate programming tasks and eliminate many entry-level developer roles, job prospects for recent graduates have sharply diminished. Placement rates at bootcamps like Codesmith fell from 83% in 2021 to 37% in 2023. Venture investors and educators cite market saturation and shifting employer needs, but AI is now seen as the “final blow.” The industry’s collapse reflects a broader trend: shrinking demand for junior coders and rising pay for elite AI researchers.[44]
2025 Agentic AI Full Stack Development Product Integrated AI development environments offering end-to-end assistance from design to deployment.

Meta information on the timeline

How the timeline was built

The initial version of the timeline was written by Sebastian Sanchez.

Funding information for this timeline is available.

Feedback and comments

Feedback for the timeline can be provided at the following places:

  • FIXME

What the timeline is still missing

Timeline update strategy

See also

References

  1. 1.0 1.1 Template:Cite arxiv
  2. 2.0 2.1 Template:Cite arxiv
  3. Swiechowski, Maciej (2020). "Game AI Competitions: Motivation for the Imitation Game-Playing Competition" (PDF). Proceedings of the 2020 Federated Conference on Computer Science and Information Systems. IEEE Publishing. pp. 155–160. doi:10.15439/2020F126. ISBN 978-83-955416-7-4. S2CID 222296354. Archived (PDF) from the original on 26 January 2021. Retrieved 8 September 2020.
  4. Withers, Steven (11 December 2007), "Flirty Bot Passes for Human", iTWire, archived from the original on 4 October 2017, retrieved 10 February 2010
  5. "NOT the world's first video game". Fake History Hunter. 15 May 2021. Retrieved 28 November 2025.
  6. Template:Cite video
  7. Experimental Games. MIT Press. 2021. p. 303. Retrieved 28 November 2025.
  8. "History of artificial intelligence". IBM. Retrieved 28 November 2025.
  9. "Programming Languages". EBSCO Research Starters. Retrieved 28 November 2025.
  10. Lindsay et al., 1980
  11. "DENDRAL". Britannica. Retrieved 28 November 2025.
  12. "Computers and Thought: A Conference on Artificial Intelligence". NASA Technical Reports Server. Retrieved 28 November 2025.
  13. Lindsay, Robert K.; Buchanan, Bruce G.; Feigenbaum, Edward A.; Lederberg, Joshua (1993). "DENDRAL: A Case Study of the First Expert System for Scientific Hypothesis Formation". Artificial Intelligence. 61 (2): 209–261. doi:10.1016/0004-3702(93)90068-M. {{cite journal}}: |access-date= requires |url= (help)
  14. "SHAKEY". Robot Hall of Fame. Retrieved 28 November 2025.
  15. Tarnoff, Ben (25 July 2023). "Weizenbaum's nightmares: how the inventor of the first chatbot turned against AI". The Guardian. Retrieved 28 November 2025.
  16. Colmerauer, Alain; Roussel, Philippe (19 November 1992). "The Birth of Prolog" (PDF). Retrieved 28 November 2025.
  17. "SWI‑Prolog Online Interactive Demo". SWI‑Prolog. Retrieved 28 November 2025.
  18. Colmerauer, Alain; Roussel, Philippe (1996). "The Birth of Prolog". In Bergin Jr., Thomas J.; Gibson Jr., Richard G. (eds.). The birth of Prolog. ACM. pp. 331–367. doi:10.1145/234286.1057820. {{cite book}}: |access-date= requires |url= (help); Unknown parameter |book= ignored (help)
  19. "History of Prolog". Mount Allison University. Retrieved 28 November 2025.
  20. "MYCIN". Britannica. Retrieved 28 November 2025.
  21. "XCON". Semantic Scholar. Retrieved 28 November 2025.
  22. "XCON expert system overview". AI the Future (blog). Retrieved 28 November 2025.
  23. Template:Cite thesis
  24. Werbos, Paul (1982). "Applications of advances in nonlinear sensitivity analysis" (PDF). System modeling and optimization. Springer. pp. 762–770. Archived (PDF) from the original on 14 April 2016. Retrieved 2 July 2017.
  25. Werbos, Paul J. (1994). The Roots of Backpropagation: From Ordered Derivatives to Neural Networks and Political Forecasting. New York: John Wiley & Sons. ISBN 0-471-59897-6.
  26. "AI Expert Says ChatGPT Is Way Stupider Than People Realize". Futurism. Retrieved 24 May 2023.
  27. Maher, Mary Lou; Tadimalla, Sharvani Y.; Dhamani, Dhruva (17 October 2023). An Exploratory Study on the Impact of AI Tools on the Student Experience in Programming Courses: An Intersectional Analysis Approach. pp. 1–5. doi:10.1109/fie58773.2023.10343037. Retrieved 4 June 2025.
  28. Durrani, Usman; Akpınar, Mehmet; Adak, Mustafa Furkan; Kabakuş, Ahmet Talha; Öztürk, Mustafa Murat; Saleh, Mohammed (31 December 2023). "A Decade of Progress: A Systematic Literature Review on the Integration of AI in Software Engineering Phases and Activities (2013–2023)". IEEE Access. 1. doi:10.1109/access.2024.3488904. Retrieved 4 June 2025.
  29. Rana, Sohel (6 February 2024). "Exploring the Advancements and Ramifications of Artificial Intelligence". Deleted Journal. 2 (1): 30–35. doi:10.60087/jaigs.v2i1.p35. Retrieved 4 June 2025.
  30. “Extended Studies Blog” (10 October 2024). "Will AI Replace Programmers? Navigating the Future of Coding". UC San Diego Extended Studies. Retrieved 4 November 2025.
  31. Chen, Tianyi (9 May 2024). "The Impact of AI-Pair Programmers on Code Quality and Developer Satisfaction: Evidence from TiMi Studio". doi:10.1145/3665348.3665383. Retrieved 4 June 2025. {{cite journal}}: Cite journal requires |journal= (help)
  32. Weber, Thomas; Brandmaier, Maximilian; Schmidt, Albrecht; Mayer, Sven (16 June 2024). "Significant Productivity Gains through Programming with Large Language Models". Proceedings of the ACM on Human-Computer Interaction. 8 (EICS): 1–29. doi:10.1145/3661145. Retrieved 4 June 2025.
  33. Brown, Leah (12 September 2024). "New Research Reveals AI Coding Assistants Boost Developer Productivity by 26%: What IT Leaders Need to Know". IT Revolution. Retrieved 6 November 2025.
  34. Zviel-Girshin, Rina (2024). "The Good and Bad of AI Tools in Novice Programming Education". Education Sciences. 14 (10): 1089. doi:10.3390/educsci14101089. Retrieved 4 June 2025.{{cite journal}}: CS1 maint: unflagged free DOI (link)
  35. Zhang, Q. (25 November 2024). "The Role of Artificial Intelligence in Modern Software Engineering". Applied and Computational Engineering. 97 (1): 18–23. doi:10.54254/2755-2721/97/20241339. Retrieved 4 June 2025.
  36. Llerena-Izquierdo, Joe; Méndez Reyes, Johan; Ayala Carabajo, Raquel; Andrade Martínez, César Miguel (3 December 2024). "Innovations in Introductory Programming Education: The Role of AI with Google Colab and Gemini". Education Sciences. 14 (12). Multidisciplinary Digital Publishing Institute. doi:10.3390/educsci14121330. Retrieved 4 June 2025.{{cite journal}}: CS1 maint: unflagged free DOI (link)
  37. Wang, Xing-lian (8 December 2024). "Application and Impact of Artificial Intelligence in Education: A Case Study of Programming Education". Lecture Notes in Education Psychology and Public Media. 74 (1). EWA Publishing: 182–187. doi:10.54254/2753-7048/2024.bo17948. Retrieved 4 June 2025.
  38. Qiu, Ketai; Puccinelli, Niccolò; Ciniselli, Matteo; Di Grazia, Luca (23 December 2024). "From Today's Code to Tomorrow's Symphony: The AI Transformation of Developer's Routine by 2030". ACM Transactions on Software Engineering and Methodology. doi:10.1145/3709353. Retrieved 4 June 2025.
  39. Lohr, Steve (20 February 2025). "A.I. Is Prompting an Evolution, Not Extinction, for Coders". The New York Times. Retrieved 6 November 2025.
  40. Sifi, Adlene (24 March 2025). "How does generative AI impact Developer Experience?". Microsoft Developer Blog. Retrieved 6 November 2025.
  41. Lepp, Marina; Kaimre, Joosep (2025). "Does generative AI help in learning programming: Students' perceptions, reported use and relation to performance". Computers in Human Behavior Reports. 18: 100642. doi:10.1016/j.chbr.2025.100642. {{cite journal}}: |access-date= requires |url= (help)
  42. "Will AI Replace Programmers and Software Engineers?". Coursera. 5 June 2025. Retrieved 5 November 2025.
  43. Tong, Anna (10 July 2025). "AI slows down some experienced software developers, study finds". Reuters. Retrieved 4 November 2025.
  44. Tong, Anna (9 August 2025). "From bootcamp to bust: How AI is upending the software development industry". Reuters. Retrieved 6 November 2025.