-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Week 3: Memos - Measurement and the Nature of Innovation #9
Comments
I think the measure of novelty within the article titled "Surprise! Measuring Novelty as Expectation Violation" was an interesting way of trying to quantify and measure a phenomena that no two people will ever define the same way. There are two components to this definition that are interesting to parse, 1. the use of patents as a instrument of quantity and 2. Mathematical approaches to bring structure to novelty as a concept |
Innovation can come from both field-specific research and cross-disciplinary collaboration. Field-specific research provides depth and domain expertise, building a strong foundation for incremental advancements. In contrast, interdisciplinary research introduces breadth, enabling novel combinations of knowledge that lead to transformative breakthroughs. The study "Surprising combinations of research contents and contexts are related to impact and emerge with scientific outsiders from distant disciplines" highlights the critical role of cross-disciplinary research in producing high-impact discoveries. This memo uses a Cobb-Douglas economic framework to evaluate the effects of incentivizing either form of research on total innovation in an academic environment that does not currently facilitate interdisciplinary exploration very much, making it much costlier to undertake than intradisciplinary research even though it has a higher marginal return in terms of innovation. The mathematical model assumes that innovation ( where:
The total budget is ( where:
Using these relationships, I calculate optimal allocations of Case 1: Baseline Scenario
Innovation output is: This scenario highlights how the high cost of cross-disciplinary research limits its contribution. Case 2: Reducing Field-Specific Research Costs
Innovation output rises to: While this intervention increases innovation, it is not the most effective way of doing so. Case 3: Reducing Cross-Disciplinary Research Costs
Innovation output improves significantly: This scenario demonstrates the highest innovation output, highlighting the importance of reducing the disproportionately high cost of cross-disciplinary research since proportional reductions in cost to the two types of research resulted in a greater gain in innovation when applied to cross-disciplinary research. The Shi and Evans article underscores the value of “surprising combinations” that arise from interdisciplinary research. They found that impactful breakthroughs often occur when researchers from distant disciplines collaborate, integrating diverse methodologies and knowledge domains. This aligns with the results of the model, which shows that reducing the cost of cross-disciplinary research enables greater resource allocation to these transformative combinations, amplifying their contribution to innovation at a greater rate than field-specific research. Policy Recommendations
By targeting the cost of interdisciplinary research, policymakers can foster the so-called “surprising combinations,” reducing the barriers to cross-disciplinary work and turbocharging innovation. |
Uniformity in Simulating Subjects with Different LLMsIn Simulating Subjects: The Promise and Peril of AI Stand-ins for Social Agents and Interactions, Professor Evans and Austin Kozlowski cover the issues with trying to simulate human subjects using large language models (LLMs). One such issue they raise is uniformity, in that a group of people you are trying to simulate will have diversity and high variance in their responses, which an LLM fails to replicate. For example, although abortion is popular in the United States, it's still around 60% who are pro-choice. Simulating these subjects with GPT-4 gets a pro-choice rate of over 95%. I wanted to explore how LLM sizes may impact whether or not the uniformity increases or decreases, with the expectation that a larger model may have less uniformity in its responses since it "understands" more nuances in the population we're trying to mimic. To do so, I chose a policy proposal that occurred after the October 2023 cutoff for OpenAI's GPT-4 models: the current TikTok ban by the US government. According to Pew Research Center, 42% of Republicans support a ban, while only 24% of Democrats support a ban. In theory, then, the perfect simulation with an LLM would result in similar estimates. I tested these two prompts on OpenAI's GPT-4o and GPT-4o-Mini. While OpenAI hasn't published the exact size of each of these parameters, estimates suggest GPT-4o is made of hundreds of billions of parameters, while GPT-4o-Mini is around 10 billion parameters. I decided to ask the same prompt 4 times: asking each model to role play as a Republican or Democrat and getting its thoughts on the TikTok ban. Each of these 4 scenarios was prompted 8 times. I chose to only do 8 prompts as the API calls are fairly expensive and I just wanted a quick sense of what the differences might be. I also set the temperature of the responses to be 1.8. This parameter controls the diversity and creativity of the response, and this temperature ensures the highest level of diversity. For each message, the model was given a system message: "You are a {party_affiliation}. Only answer YES or NO to the following questions." The question asked was "The short-form video-hosting service TikTok has been under a de jure nationwide ban in the United States since January 19, 2025. Do you support the TikTok ban?" I got the context sentence from Wikipedia and chose to minimize political bias. The results are summarized in the graphical figure below: ![]() These results suggest that larger models may have slightly less uniformity, based at least on the samples from the LLM simulating a Democrat. However, it is still very far from the true estimates, and the only variability came in 2 responses. All this suggests that the uniformity issue with LLMs cannot be overcome by scaling LLMs to be larger or just increasing the temperature parameter. One potential answer, which I didn't explore but may explore more in future memos, is to add granularity to the system instruction. Instead of specifying that the LLM is a Democrat, we can specify it is a 25-year-old Democrat, or that it is a Democratic white male to get more diverse answers. |
This week’s readings explore how modular recombination and assembly pathways influence innovation and growth. Modular recombination emphasizes the reuse of existing components to create novel outputs, while Assembly Theory highlights how historical constraints shape the complexity and diversity of innovations. This memo investigates how different innovation patterns—whether incremental or transformative—correlate with sectoral growth rates across industries. Question: Hypothesis: To analyze this relationship, I propose using: By linking innovation patterns to economic growth, this memo provides a framework for understanding how modular recombination and assembly pathways shape industries differently. This dual lens offers critical insights into optimizing innovation-driven growth strategies across sectors. |
Innovation and Word Embeddings: Apple vs. BlackBerryIn “Modularity, Higher-Order Recombination, and New Venture Success,” Cao, Chen, and Evans build a semantic space through word embedding to measure modularity and resulting successes of new companies. The article explores temporal changes and semantic shifts; for example, Amazon has become more closely associated with ecommerce than the rainforest. I asked in my question this week if word embeddings could capture more pragmatic notions. For example, in the public mind, Apple tends to be associated with innovation and technological progress. Can and, if so, do word embeddings capture such sentiments about innovation? MethodologyI found a random selection of 151,573 Wikipedia articles, thanks to the Wikipedia dumps. Of these pages, both Apple, Inc. and BlackBerry were present in the corpus, so I settled on comparing those two. Also, since BlackBerry has stopped making phones (though they are now an IoT company), it felt like a good example of a less innovative enterprise. I filtered across all articles for “Apple” and “BlackBerry” (case insensitive). I then used the Python package I then define a list of “innovation” words. These include “innovation”, “creativity”, “design”, and “progress”. Though not exhaustive, I use this as a proxy for relating these companies with the general notion of innovation. I then calculated the similarity scores, which is the cosine similarity between the two vectors Results and DiscussionI first output the most similar words to both “apple” and “blackberry”. Both of these are technology companies and fruits, so they both have the same issues. “Apple” is heavily associated with “orchard”, but also “iigs”, “macintosh”, and “wozniak”. “Blackberry” is associated with the fig “banyan”, but also Android “oreo” and “cisco”. Future iterations would ideally control for this. The following chart compares the similarity scores. Across the board, there are stronger associations with “apple” and innovation words than with “blackberry”, though not by very large amounts, so statistical significance tests would be necessary. I also would try similar analyses on other company pairs (maybe Netflix/Blockbuster and Tesla/Mitsubishi). For Apple/BlackBerry, it is possible that semantically, both of these started out near the same point, but as Apple innovated and BlackBerry did not, Apple’s score increased. Temporal analysis, then, would be interesting in further examining this topic. I also noticed that the amount by which Apple is greater is similar for all four innovation words. Ultimately, though this is not necessarily how word embeddings should be used, it seems it is possibly effective to capture differing levels of innovation through this method. For reference, my code can be accessed at the |
Relationship between Founders' Industry Background and Venture SuccessIn the first paper by Feng SHi and James Evans relating surprise in research content to impact, one key finding was that surprising advances in innovation typically emerge across different fields rather than within a single field, often when entrepreneurs / scientists from one discipline push ideas across to another. The beneficial impact of cross-disciplinary collaboration on innovation prompted to think about how venture success may be correlated to the diversity of industry backgrounds that founders have. My hypothesis is that since ventures often have multiple founders, if they have professional experience across different industrys and/or if they studied different majors in college, then this may be positively correlated to higher venture success and exit price. Methdology To test the hypothesis, I gathered the following data:
Results:
This has several important implications for fostering entrepreneurial activity. First, cross-industry collaborations appear to be associated with greater commercial success in ventures. Second, diverse industry backgrounds among founders are linked to higher exit prices, suggesting that such diversity may enhance both the ideation process and the operational execution that follows. From a policy perspective, this highlights the need for governments and research institutions to actively promote cross-disciplinary dialogue by hosting events, courses, and programs that encourage collaboration between different fields. |
The shrinking value stock premium and why we might not need policy interventions: Consider an inverted U shaped relationship between transdisciplinarity and impact. I believe that this relationship could be feasible since it is easy to see how practitioners that come from other disciplines can bring with them new perspectives and methods beneficial to knowledge creation (Consider Gary Becker and his application of economic tools to what were considered “social issues”). Beyond a certain point, however, synergies between disciplines or applicability of ideas might diminish. For example, I would not expect the methods of a poet to be of much use in physics (although of course I could be wrong). Given this U shape, attempts to subsidize novelty through induced transdisciplinarity has the potential to negatively impact the value of the research being produced. To picture an extreme case, individuals might prioritize transdisciplinarity over research impact. It is unclear what a good incentive structure might look like to push transdisciplinary research towards optimal levels. I feel however, that it might be the case that the systematic undersupply of trans-disciplinary ventures that the paper identifies could be due to academics simply not realising the value behind them which this paper changes. Mclean and Pontiff in their 2016 paper identifies how academia destroys stock market predictability post publishing as investors move to close inefficiencies. Some attribute the recent shrinking of the value stock premium to this. I feel that academia and their past disdain for novel methods could resolve itself through papers like this, provided academics are principally motivated to do “good” work, and publishers and award-givers wish to stay relevant, which I feel they are. I would argue that the value of this paper lies not only in its influence on policy but also in its ability to inform academics/inventors and incentivize them to pursue novelty as a means to impact not as an end in itself. I would also like to consider how the model might behave if we were to see this shift among academics. Novelty is measured by predicting what is the most likely combination of contents and contexts that are likely to occur in future research or patents and seeing by how much actual journal/patent entries subvert this expectation. Those that closely adhere to the prediction are deemed un-novel while those that deviate are deemed novel. The model’s predictions about what might be most feasible is dependent on the configurations it encounters in the training data. Assuming that there is a common thread to innovation (a big if I know) more encounters with novel types of ideas the model might demand even more outrageous and unthought of combinations to consider ideas similarly novel. The model is trained on a dataset that transverses a pretty large time span meaning that increases in novelties in more modern data might constitute a small enough portion of the data such that this inflation might not amount to much. However, I would assume that as the market for ideas approaches the optimal level of novelty, we might see the sensitivity of impact to novelty increase as ideas on this efficient frontier of novelty are perceived less and less novel by the model. |
In my project, I will investigate whether modular AI architectures, such as transformer-based models and established cloud platforms, help AI-oriented start-ups move more quickly from seed funding to follow-on rounds. Preliminary evidence suggests that these ventures can attract broad coalitions of investors, each with expertise in different modules—e.g., data engineering, software-as-a-service (SaaS), or consumer-facing applications. This dynamic mirrors the phenomenon described by Cao and colleagues, wherein higher-order invention not only lowers technological risk but also harnesses network effects that amplify growth. An empirical angle I am exploring is the relationship between “local” and “global” novelty in AI ventures. “Local novelty” captures how distant each AI module’s subcomponents are from one another (e.g., software infrastructure and language processing). “Global novelty,” by contrast, indicates how far apart these AI modules lie from other sets of established technologies in the broader innovation space (e.g., combining generative AI with biotech workflows). The hypothesis, informed by higher-order recombination theory, is that small “local distance” paired with large “global distance” will yield faster scale-up and better funding outcomes—because the base modules are well-tested, yet their combination is unexpected and can open new markets. The R demonstration below visualizes how “local” and “global” distances can interact to explain hypothetical “success” scores among AI start-ups. Although this is only a simplified simulation, it sets the stage for a more in-depth empirical analysis of real-world AI firms. By collecting actual data on AI startups’ technology stacks, pitch descriptions, and funding histories, I plan to test whether ventures focused on assembling stable modules truly outperform those pursuing radical, “ground-up” innovation. |
Justifying My Second Major in English: Applying Assembly Theory to Poetry In the article about Assembly Theory, we are introduced to the assembly equation, which attempts to quantify the “total amount of selection necessary to produce an ensemble of observed object”. While the article uses biology and chemistry as its field of study, as an Econ and English major, I try to justify completing my second degree to myself by seeing if we could apply AT to the study of poetry as well. In essence, can we use AT to quantify how “new” and “complex” a poem is. Of course, this would take many artistic liberties to define what a complex and new poem is, which would be more suited for a thesis than a memo on github. For simplicity’s sake, let’s take “complex” as just: how many big words and new phrases are used in the poem. We start with the given assembly equation: ![]() In this case, we simply replace “objects” with words. The assembly index a_i becomes the amount of unique letters in that word, n_i is the amount of times the word appears in that poem. N is the total number of unique words. e is still euler’s number. N_t is the total number of words in the poem. We use this formula but now call this variable “word complexity” or W. What’s interesting is trying to quantify the amount of “new phrases” the poem has. For this, I created a variable, P, that we will use to “deflate” W. (i.e. it could have many complex long words but if they have all been said before then it’s not “new” according to our definition). First, we set a variable P, which will represent the total similarity of our phrases in the poem to other famous poems. Next we manually extract phrases from the poem. Then. assuming we have access to the 1000 most famous poems ever written, we compare each phrase to each of the 1000 Poems. We use the TF-IDF method to get a number between 0 and 1 for each poem it’s compared to. Then we sum this up over all 1000 poems and then again over all phrases in our poem. (To measure similarity, we use the TF-IDF method for measuring semantic similarity. (https://medium.com/@igniobydigitate/a-beginners-guide-to-measuring-sentence-similarity-f3c78b9da0bc) As an example, if my current poem has the phrase "sweet like chocolate", and we compare it to another poem that has "tasty like candy" and "crunchy like peanuts" and say the former has a TF-IDF value of 0.68 and the latter has 0.37, because we take the maximum, the µ max value of that i-th phrase (sweet like chocolate) compared with the j-th poem (the one with the candy and peanut's phrase) will be 0.68. Then we sum it up over all other 999 poems to get an overall score for that phrase, and then we sum up for all phrases in my poem. We will now deflate W by a factor of P, to get our final value, A, that shows the “complexity” of a poem Of course, this is a practically messy and subjective measurement that needs much refinement before it is somewhat useful. For example, we would need to craft an algorithm to automate the process of comparing each phrase to the list of poems, and we still need to choose what kind of semantic similarity measure we can use. Even then, simply arithmetically summing up the µ is also very dubious, and our choice of deflating W so simply even more so. Accuracy aside, it’s still interesting to see how we tailor the measures of novelty to the subject matter, much like what was said in class about using subjective measurements for different fields. |
The Relationship Between Novelty and Probability of InnovationThis week, I wanted to explore the relationship between the perceived novelty of an invention and the probability of that invention occurring, based on prior knowledge and the inventive process. The central argument, as presented in Foster et al. 2020, is that novelty is not an objective quality but rather a context-dependent assessment based on how surprising an invention is, given what is already known and how inventions typically come about. This can be formalized through an equation that captures how novelty is a decreasing function of the probability of a discovery. The core relationship can be expressed as: Where:
In this equation, as the probability P(D|K,S) increases, the value of N decreases, demonstrating the inverse relationship between the probability of an invention and its perceived novelty. When P(D|K,S) is very low (the invention is highly improbable), the novelty N approaches a maximum value determined by the constant m. Alternatively, when P(D|K,S) is high (the invention is highly probable), the novelty N approaches zero. Foster et al. emphasizes that novelty is a function of surprise, which is directly related to the improbability of an event. The inverse function represents this relationship, where improbable events (low P(D|K,S)) are surprising, and therefore have high novelty. The constant m allows for the context-sensitivity of novelty assessment. In fields where highly novel inventions are common, m can be larger. In contexts where most inventions are incremental improvements, m can be smaller. As a real-world example, consider the development of the first PC. At the time, the probability of creating such a device given the existing knowledge and technology was very low, meaning P(D|K,S) was small. As a result, the perceived novelty (N) was very high, since the inverse function m / P(D|K,S) would give a high output. Now, consider the release of a new model of an iPhone: given the existing knowledge and technology, the probability of producing this device P(D|K,S) is high, and the perceived novelty is low (N). The constant, m, can be thought of as a baseline to compare how novel two inventions are relative to each other, regardless of their context, and can be adjusted in relation to the inventive domain. In conclusion, the relationship between novelty and probability is inverse: higher probability of invention results in lower novelty, and vice-versa. This relationship is mediated by prior knowledge and beliefs about the inventive process, which are subjective, vary by context, and can be measured. |
The readings emphasize that “unexpected” combinations of ideas, contexts, and disciplines frequently drive impactful innovation. Building on this, I set out to examine how these concepts play out in U.S.-based AI startups that have achieved substantial early-stage funding, particularly Series A rounds exceeding $15 million. This level of funding serves as a strong indicator of both investor confidence and the perceived potential for groundbreaking innovation. |
The paper on Modularity, Higher-Order Recombination, and New Venture Success examines how start-ups combine business “modules”— proven technologies and applications—to speed time to market and reduce the risk of failure. The distinction is between lower-order inventions, which integrate largely untested components, and higher-order inventions, which assemble already-functional modules into novel configurations. This modular assembly confers many advantages because the underlying technologies are already proven. Furthermore, such process helps founders attract more diverse investors, accelerating scaling and boosting IPO or acquisition odds. Building on this idea, I propose a simplified economic model from the perspective of a venture capital firm looking to invest in or combine different modular companies. The model evaluates key factors such the following:
Acquirer’s Value=p×[α⋅R]+(1−p)×[β⋅R]−C. Start-ups with higher-order modules enable synergy (𝛼) to be realized more quickly and increase the likelihood of successful integration (𝑝). By contrast, lower-order components, which consist of untested or novel inventions, might offer significant upside if they complement the acquirer’s existing modules(R) but come with lower probabilities of success and salvageable value(β) if the technology fails. Real-life example:Google’s Acquisition of YouTube This example and model highlights how investors and corporations decide to acquire companies based on modularity and synergy. By focusing on stable, higher-order modules instead of entirely novel technologies, acquirers can significantly reduce time-to-market and the risk of failure. This strategic approach demonstrates the practical advantages of higher-order invention, as outlined in the modularity framework |
Memo: The Dynamics of Scientific Impact Through Knowledge ExpeditionsTheoretical FrameworkDrawing from recent empirical studies of scientific discovery and technological innovation, I propose a dynamic impact equation that captures how breakthroughs emerge through the interaction of knowledge expeditions and the adjacent possible: Where:
This equation attempts to identify the scientific impact by integrating three key mechanisms identified in recent empirical work. The correlation function Empirical ApplicationThe empirical evidence strongly supports this theoretical framework. In biomedical research, papers representing knowledge expeditions across semantic domains achieve 3.5 times higher citation rates than typical papers, according to Shi & Evans' analysis of millions of MEDLINE articles. The equation explains this through the interplay between semantic distance ( The framework also accounts for observed differences between scientific domains. In physics, where Tria et al. found more hierarchical knowledge structures, the distance decay parameter This theoretical framework also identifies the pattern of breakthrough discoveries. When examining Nobel Prize-winning papers, we find they typically display high The equation thus provides not only a quantitative model for understanding past scientific breakthroughs but also suggests strategies for fostering future innovations. It implies that the most impactful discoveries emerge from well-calculated knowledge expeditions that balance semantic distance with cognitive accessibility, while generating sufficient surprise relative to the current adjacent possible space. |
Modularity and Higher-Order Recombination in Taiwan’s Semiconductor IndustryTaiwan's success in the semiconductor industry is a clear example of how modularity and higher-order recombination can drive innovation and growth (as talked about in this week's reading). Modularity, the division of labor into specialized components, has allowed Taiwanese firms to excel in specific segments of the semiconductor chain. Higher-order recombination (integrating these components in innovative ways) has helped Taiwan's large and small firms meet the demands of the industry. Modularity: Foundation for Success in TSMCA great example of modularity is Taiwan Semiconductor Manufacturing Company (TSMC), which has become the world’s largest foundry by focusing solely on wafer fabrication. This specialization allows TSMC to invest heavily in cutting-edge technologies, such as 3nm and 5nm process nodes. Specialization leads to efficiency. Its focus on manufacturing lets global design firms like AMD and NVIDIA concentrate on IC design while outsourcing production to TSMC. This collaboration, illustrated in the figure below, shows how modularity allows companies to specialize and innovate in a very specific factor. ![]() Smaller firms like MediaTek, which specializes in IC design for consumer electronics, also benefits from TSMC's modularity. Modularity enables MediaTek to focus on its core strengths in designing chips for smartphones and IoT devices, while it can leverage the TSMC foundry’s expertise for production. Higher-Order Recombination Drives InnovationWhat sets Taiwan apart is its ability to combine specialized modules with global partners. An example is TSMC’s collaboration with Apple to manufacture the A-series chips used in iPhones. Apple supplies its proprietary designs, and TSMC uses its efficient manufacturing processes to produce these high-performance chips. This has led to significant revenue growths for both companies. The back-end manufacturing segment also plays a key role in Taiwan’s success. Companies like ASE Technology excel in packaging and testing, critical steps that protect chips and ensure their reliability. ASE combines advanced materials with TSMC-fabricated wafers, and this transitions effectively from manufacturing to end-user-ready chips. As seen in the figure below, packaging and testing accounted for 13.4% of Taiwan's semiconductor output in 2023. Every process in the supply chain is essential - disruptions in any single aspect may lead to significant delays in semiconductor production. The industry’s value distribution also shows how different modules—design, manufacturing, and back-end processes—contribute to Taiwan’s dominance. Foundry services alone accounted for 57.4% of the sector’s output in 2023, and without specialization, this dominance would have been impossible. ![]() Taiwan's use of modularity and higher-order recombination also ensures long-term resilience. As shown below, the semiconductor industry weathered global downturns and maintain strong revenue growth. During the 2021 global chip shortage, Taiwan’s firms rapidly scaled production by using their modular processes and collaborating within while most of the world is in pandemic lockdown. This ensures supply remains optimal. ![]() Taiwan’s semiconductor success is rooted in modularity and higher-order recombination. Focusing on specific strengths and integrating expertise from other global firms, companies like TSMC and ASE have seen huge growth and innovation. Aiming for modularity has made Taiwan a global semiconductor power, and more growth is yet to come. Source: Pwc Taiwan |
The Evolution of Technological Development in Image Capturing In Combination and Structure and Mechanisms of Evolution, Arthur presents technology as evolving within a recursive system, where innovations build upon existing technologies, creating sequences and branches of advancements. This recursive structure follows a tree-like model, with fundamental principles forming the trunk, major technological breakthroughs as main branches, and subsequent refinements extending as sub-branches. Using this framework, we can analyze how image-capturing technology evolved into film and digital cameras today within this recursive structure. At the core of image-capturing technology is the fundamental principle of light projection and recording. The earliest known technology, the camera obscura, established foundational knowledge in optics which helped us understand how light could be projected and manipulated. This led to the development of daguerreotype photography in the 19th century, which refined the concept to allow permanent image recording using light-sensitive materials, fulfilling the opportunity niche of personal and commercial portraiture capturing. The invention of film photography plays an important role in the evolution of image-capturing technology. The shift from plate-based methods to flexible roll film created an opportunity niche by making photography more accessible and convenient for the average consumer. This development not only enhanced image storage potential but also introduced modular components like interchangeable lenses, leading to the diversification of camera types. Digital cameras replaced film with electronic sensors, enabling instant image capture without chemical processing. This advancement opened new opportunities, such as digital storage solutions with SD cards, digital image processing software, and customizable settings like ISO and shutter speed. To visualize this evolution, I created a simplified tree structure illustrating the transition from film to digital cameras. At the core of the tree is the camera itself. The camera's primary function—capturing light to produce visual representations—remains constant, but the underlying mechanisms have transformed as we move from film to digital. The branches are divided into four main assemblies: Processing Assembly, Lens System, Power & Control, and Image Capture. These assemblies further break down into subassemblies supporting the main systems. Following the pathways highlighted in purple, we see how electronic image processors replace chemical film processing, which in turn leads to the opportunity niche and development such as digital storage replacing physical film rolls. As digital cameras evolved, their increased ability to adjust image settings digitally created a demand for a more sophisticated user interface—an entirely new branch that emerged only because of the advancements digital technology introduced. As visualized in the diagram, components not highlighted in purple represent technologies that have become inactive in digital cameras. This demonstrates the sequence technology goes through in evolving, through opportunity niches and replacement of sequences of components. |
Sentiment Differences in AI- and Human-Produced TextIn Simulating Subjects, Evans and Kozlowksi discuss how LLMs face the problem of uniformity: a collection of responses from a LLM in an attempt to model a human sample of responses offer less variance than a true human sample. When asked the question, “Should the government do more to help the poor?” GPT-4 overwhelmingly responds with the moderate answer on a scale of 1-5, whereas the human GSS sample was far more diversified. When looking toward Human-AI systems––where longer responses are key to simulating interaction instead of a binary yes/no or scale-based answer––I was curious if this uniformity extended to the style of LLM responses. Sentiment analysis is a way of assessing the degree and direction of emotion in text. Google’s Natural Language API is trained on millions of contextual sentence fragments and can provide two sentiment metrics for text: a sentiment score from -1 to 1 and a magnitude score > 0. Sentiment scores closer to 1 indicate more positive sentiment in text (e.g. compliment), whereas scores closer to -1 indicate negative sentiment (e.g. criticism); a higher magnitude score indicates greater conveyed emotion, which is often dependent on text length. For example, a heated argument between two people on an online forum would likely have scores close to -1 and high magnitudes. In order to properly simulate interaction, we want diversity in sentiment scores and magnitudes for LLM-produced text, akin to human-produced text. The stigma around base GPT models is that they sound “robotic” or “awkward” without proper prompting, so we might expect sentiment scores closer to 0 (neutral) with low magnitudes (non-expressive). I tried to compare the sentiment metrics for a random sample of human-produced responses to an open-ended, bias-inducing question with those from GPT-4 responses. I took a random sample of 20 comments on Reddit to the question, “What is your unpopular opinion?”, collected sentiment metrics on each of these responses using Google’s Natural Language API, and recorded the length of the comment to normalize the magnitude. I then initialized 20 separate calls (no saved cache) to GPT-4 with the same question and recorded the same metrics. The results show that the Reddit (human) response sentiment scores on average are more extreme than ChatGPT sentiment scores and show higher variance. The data also suggests that the question lends itself to more negative responses (which is expected, since contrarian opinions often criticize the dominant opinion), as most responses score below 0. In other words, ChatGPT was generally more neutral in tone than humans. However, humans voluntarily answered the Reddit question, whereas ChatGPT was prompted––there is selection bias for humans to answer if they strongly feel about something, either positive or negative. Human responses also had a higher mean magnitude score (0.044 compared to 0.038), which implies that humans were also more expressive in their responses than ChatGPT. However, it’s interesting to note that ChatGPT provided more responses with magnitudes above 0.04 than Reddit based on a single prompt. Because the question intended to generate personal bias, ChatGPT’s predictive capacity likely depended on previous expressive text. With a proper web-scraping tool and a subscription to Google’s API, we could obtain a much more accurate sentiment distribution of human-produced text beyond a sample size of 20. However, I think this case more largely highlights the obstacles in creating out-of-the-box LLMs that are prepared to simulate humans. I expect that with significant prompting and characterization of an LLM, these distributions could look far more similar. |
Studying Surprise and Interdisciplinary Work in the Social Sciences“Surprising combinations of research contents and contexts are related to impact and emerge with scientific outsiders from distant disciplines.” discusses novelty as related to interdisciplinary studies in the physical sciences. This suggests analogous study that could be performed in the social sciences that may have interesting implications. Due to the overlapping subjects of study in the social sciences, we would like to encourage interdisciplinary work. This would both further understanding of the concepts they seek to understand (e.g. power, democracy, wealth, happiness), and prevent wasted efforts in pursuing research in avenues that are already well documented within other disciplines. A paper titled “The Superiority of Economists” collects data on citations in the social sciences referencing works in fields other than their own. In the attached chart, economics stands out as being the most cross referenced work out of economics, political sciences, and sociology. A sanguine reading suggests that economics has great efficacy as a social science, causing the other fields to look up to their research. However, an alternative hypothesis, especially knowing the value of interdisciplinary work, suggests that economics as a field could use more insight from its fellow social sciences. In fact, while a lot of the techniques of economic research have diffused into the greater social sciences, the reverse has not been seen as much. There may be limitations with the current methodology in performing this study in the social sciences. Firstly, there is a consideration we have to make about keywords. While terminology in the physical sciences is more well defined, there isn’t such rigidity in the social sciences. While this flexibility around terminology is necessary to make the papers themselves coherent, it causes difficulty in using them as keywords to assess how predictable one paper is. For example, the word “institution” has a different definition from field to field, and may even be defined differently from paper to paper within a field. This might make it difficult to construct models around the word “institution”, since its presence can imply many different constructs. This would confound our ability to measure surprise off of words. Studying interdisciplinary work in the social sciences would let us understand if surprise plays a factor in the significance of research in the social sciences. We would expect that social scientists aware of the state of research in fields other than their own would have a better perspective of the grand scheme of things and be able to produce better research. Interdisciplinary work would also combine approaches found in multiple fields, which can yield novel techniques for interesting results. |
How Expanding Open Access Will Drive Novelty“Surprise! Measuring Novelty by Simulating Discovery,” describes the following relationship between novelty (N) and innovation (I) where A is prior knowledge or art, and S is the creative / search process As I was brainstorming ideas for this memo, I was met with many paywalls. This led to me looking into how paywalls limit innovation by restricting the availability of research papers –ex. Nature, Science, etc. - and news articles - ex. The New York Times, The Economist, etc. I observed the following: With this, I created a simple model for the probability of innovation, Secondly, I modeled the relationship between Plugging this into the function for the probability of innovation: Thus, we get the following relationships: The intuition behind this is that “open-access research outputs receive more diverse citations.” As “Surprising combinations of research contents and contexts are related to impact and emerge with scientific outsiders from distant disciplines” showed, diverse careers, teams, and methods are positively correlated with hit probability. The The implication of this is that we should increase open access. More access to knowledge will lead to more novelty, which would boost our economy. Currently, only about 38% of all scholarly articles are open-access and this is harming our economic potential. |
Biotech/pharma is characterized by lengthy clinical timelines and high attrition, so Cao's emphasis on modular recombination is particularly salient here. Two qualitative considerations stand out. Validated “modules” like CRISPR-Cas9 for gene editing or lipid nanoparticles for drug delivery permit quicker iteration in drug discovery than entirely de novo approaches. Because these modules typically come with established safety or efficacy data, startups can attract more diverse capital by proving they can integrate existing technology stacks rather than taking on all discovery risks alone. Presented here is a model that captures this dynamic and employs a multi-stage hazard framework. Let where Financing dynamics are also a critical variable in biotech, so to incorporate them it's necessary to extend the model to a system of simultaneous hazards: where On the measurement front, capturing |
Analysis of Factors Affecting Startup Success Using Random Forest Model I found a dataset on Kaggle containing information about 190 startups, including their funding round numbers, total funding amount, geographic location, time to first funding, average number of VC fundraising participants, industry, and their status (either "closed" or "acquired"). After cleaning the dataset, including filling in missing values for numerical columns with the median, I selected 8 features that I believe might influence a startup’s status. These features include: funding_total_usd, funding_rounds, avg_participants, is_software, is_ecommerce, age_first_funding_year, milestones, and relationships. I then used a Random Forest machine learning model to predict startup status. The dataset was split into 80% training data and 20% testing data. The classification report (attached below) for the model indicates an overall accuracy of 77.3%, which is a solid result. The model performs well in predicting acquired startups, with a recall rate of 91%. However, the recall for closed startups is lower, suggesting that the model struggles more with predicting this class. ![]() In terms of feature importance, the variable funding_total_usd has the highest weight (0.2366), which aligns with our expectations: investor funding is crucial for a startup's survival and success. The second most important feature is relationships (0.2272). While the exact nature of this "relationship" score is unclear, it likely reflects factors such as investor networks, customer relationships, and possibly political connections, all of which are known to impact business success. Interestingly, funding_rounds has a relatively low impact on predicting startup success. This could be because well-funded startups with strong cash flows may not need to rely on multiple funding rounds to grow, allowing them to skip rounds and still thrive. Lastly, industry or sector has the lowest feature importance, which may seem surprising. Despite the common belief that fast-growing industries like software and e-commerce drive acquisition, this dataset suggests that industry type doesn't strongly predict whether a startup will be acquired successfully. This model was trained on a relatively small dataset of startups, and there is room for improvement, particularly in feature engineering and model selection. Nonetheless, the insights gained offer interesting perspectives into factors influencing startup success, particularly regarding funding, relationships, and age at first funding. |
From this week's readings, it's clear that modeling business complexity in comparison to various performance metrics (returns, survival, market capture, patent portfolios, etc.) in a way that meaningfully and adequately accounts for different organizational structures, technological developments, and industry environments, is a difficult problem. Readings from this week and external research wrestle with this problem, and explore unique ways to evaluate firm complexity in order to model the relationship between complexity and performance. "Modularity, Higher-Order Recombination, and New Venture Success" by Cao, Chen, and Evans discusses how successful firms incorporate modular recombination, ideally high-order recombinations, to experiment with innovative ideas and business models. Another study, Hur and Kim (2023), also emphasizes the role of organizational innovation and modular recombination to create competitive edge and uses standardized metrics on these characteristics within firms in comparison to financial performance metrics. This type of innovation in firm structure can leverage new technology to boost profitability, which is important in fields where a technological edge isn't enough to sustain profitability. Therefore, new ventures that leverage modularity for organization growth and design will accelerate innovative capabilities and get access to capital quicker. If we are to believe that modularity benefits early ventures, we can assume this is related to the paper, "Assembly theory explains and quantifies selection and evolution," where modular design allows firms to adapt quickly and efficiently to changes in market environments and technologies. My main issue is with how we quantify firms' technological complexity and the effect of modularity in various industries. Modularity will have higher impacts on early-stage, more versatile business models, and particularly within technologically complex industries. However, many practitioners still use patent portfolios as a proxy for technological complexity, which we seen can be used non-productively by larger incumbent firms to stifle innovation. My proposed solution represent technological complexity of firms is to include a term to represent modularity within firms, and penalize that term with increasing firm size where the effects are less pronounced. With this modified function, sheer patent portfolio size presents a more meaningful innovative advantage to smaller firms with modularity and higher-order recombination, which acts as a sort of multiplier for smaller, but highly innovative, patent portfolios. |
Context and Argument One contemporary application of this insight is seen in the development of quantum computing algorithms for chemistry. This field brings together quantum physicists, computer scientists, and chemists—distinct groups traditionally working in separate silos. For example, the Variational Quantum Eigensolver (VQE), a key algorithm for solving complex molecular problems, emerged from the combined efforts of quantum computing researchers and chemists. The success of VQE underscores Shi and Evans’ argument: high-impact research is born at the intersections of fields, often driven by interdisciplinary collaborations. Empirical Case: Quantum Computing for Climate Solutions This case highlights the importance of disciplinary distance in driving innovation: quantum computing researchers bring fresh perspectives and tools, while materials scientists contribute domain-specific expertise. The resulting breakthroughs align with Shi and Evans' finding that impactful research arises from outsiders navigating between distant contexts. Analytical Demonstration: Novelty as an Intersection of Disciplinary Distance and Cognitive Integration 𝑁 = 𝛾 ⋅ 𝛿 ⋅ 𝐼 Where: 𝛾 is a scaling factor reflecting the broader research environment (e.g., funding, institutional support). This equation captures how novelty emerges when distant disciplines are bridged effectively. In the quantum-carbon case, 𝛿 is high due to the distinct methodologies of quantum physics and chemistry, while 𝐼 depends on the extent of collaboration and knowledge exchange among researchers. Shi and Evans’ findings are deeply relevant to today’s most pressing challenges. In areas like climate change, fostering interdisciplinary collaborations can unlock transformative solutions. |
A very surface-level review of modularity and imitation between China and the US Modularity, Higher-Order Recombination, and New Venture Success seems to claim that combinations of higher / lower modules, particularly those that have large distances between them, will lead to greater probabilities of successful innovation -- since the data they collected seems to be US based, we can interpret the result in the context of the US. However, can we draw similar conclusions in other countries? In the past, China was a place where copycats of US startups would emerge, some of which would become successful. It would seem intellectually dishonest to claim that all comparable Chinese companies copied US counterparts; even if we looked at a metric like year founded, a company can go through many pivots. It also might be the case that humans in different countries encounter similar problems, and thus local founders independently discovered and resolved those issues. However, it also seems intellectually dishonest to not recognize that copying is possible. This point might be strengthened when founders have a history of copying US startups in Chinese markets, and even admit their source of inspiration publicly. Perhaps the most notable example is Wang Xing, who tried imitating Facebook and Twitter before finally becoming successful with a Groupon clone (which has since evolved). Examples of similar companies (not saying they're copycats) include Didi and Lark. Thus, one might conclude that this imitation is not the same as the modularity presented in Modularity, Higher-Order Recombination, and New Venture Success; that copycat venture success is formed by other factors like cultural adaptation and local go to market strategies. I think I'd agree with this, just look at businesses like Careem, or even Meituan (Groupon copy). But I can also see an argument that cloning a business is the same as the modularity that the paper speaks of. Modularity takes existing modules (which I thought were broadly defined in the paper) and reorganizes them. Why can a startup idea that has found success in the US, and cultural adaptation not be two different modules? The distance the paper argues is necessary for innovation exists between the idea of the solution and the problem in a local market. That is, novelty is subjective, and repurposing existing modules is similar to copying a startup in a new country. To be honest though, I don't really like this argument because I don't like the modularity argument that much -- aren't all things based on things that have been done in the past? What even is a module? Why do we have to abstract this so much? However, this class has made it obvious that competition (in healthy amounts) can lead to greater innovation, so copycats are probably good in that sense (maybe not in the sense of increasing US-companies' profits), but only if these US companies trying to take on new countries have the motivation and ability to iterate well (unlike Uber in China). Survey of Chinese Espionage in the United States Since 2000 |
Impact of Market Conditions of Startup Success Understanding the drivers of startup success is a foundational question in entrepreneurship research. The readings from this week on modular recombination and assembly theory suggest that success hinges not only on innovation but also on the strategic combination of existing components. In this memo, I want to explore the question of how market timing can amplify venture outcomes, particularly in the context of startups leveraging modular recombination. The concept of modularity highlights how startups combine pre-existing, proven components to mitigate risks and accelerate development. For example, a startup might integrate established technologies (such as cloud computing and mobile applications) to create a novel product or service. This modular strategy reduces uncertainty by drawing on previously validated innovations. However, modular recombination alone does not guarantee success; I think its impact is also significantly amplified or constrained by prevailing market conditions. Favorable market conditions (such as increasing demand, technological maturity, or supportive regulatory environments) can create windows of opportunity where modular recombination yields outsized returns. Conversely, unfavorable conditions may stifle even the most innovative startups, as markets fail to adopt their solutions. Assembly theory provides a useful lens here, emphasizing how historical pathways and selection pressures influence the viability of new ventures. To produce a theoretical analytical element that captures the interaction between modular recombination and market timing, I propose the following outline. To create the graph using existing data, I would extract startup descriptions from datasets like Crunchbase or VentureXpert and compute a modularity index using NLP techniques, such as clustering terms in company descriptions to identify the diversity of modules combined. A market favorability score would be derived from external metrics like GDP growth rates, venture capital cash-on-market levels, and industry funding trends during the startup’s launch. I chose these specific metrics because I found a research article titled “Startups, Growth, and the Macroeconomic Environment: Valuation and Business Cycles”, which found that these were the key macroeconomic factors that influence startup success. Success outcomes (like IPO likelihood or acquisition value) could then be mapped onto a heatmap or contour plot to visualize how modularity and market favorability interact to predict success. Understanding the interplay between modular recombination and market timing is crucial for guiding entrepreneurs to optimize strategies and launch timing. It also helps investors and policymakers foster environments that support innovation. This analysis could reveal how these factors combine to drive startup success, offering actionable insights for research and practice. |
In “Modularity, Higher-Order Recombination, and New Venture Success”, Likun Cao, Ziwen Chen, and James Evans explore the concept of modularity and higher-order recombination in the emergence of successful new ventures and complex systems. Successful start-ups do not simply combine random components, but instead strategically assemble novel combinations of successful modular components. This approach reduces the risk of entrepreneurial failure and enables rapid experimentation. Data Collection: Gather data on ventures that have gone through IPOs or acquisitions, including: Time Frame: Track these metrics for 5-10 years (or more depending on availability of data) post-IPO or acquisition. Grouping: Categorize ventures into: Performance Metrics: We'll use a composite score of multiple metrics to measure long-term success: Model: We can use a panel data regression model to analyze the long-term performance of these ventures. The model would look like this: Comparative Analysis: to compare the performance of different groups over time This comprehensive model allows us to track long-term performance using multiple metrics, compare different groups of ventures based on their initial success, and account for industry and market factors. By implementing this analysis, we can gain insights into whether the initial success of a venture (as measured by IPO price or acquisition value) is a reliable predictor of long-term performance and sustainability. |
Virtual reality (VR) goggles enable users to interact with a fully-customizable, visual form of reality. The invention allows for a gaming experience that is deeply immersive and makes the body itself a controller. Several of the readings present the notion that novelty is an important measure when determining an invention’s future success. Virtual reality was presented as a groundbreaking revelation that would soon take over the gaming world. This reality has not yet panned out and several devices have failed to achieve broad market adoption. From a technological standpoint, the device is a feat and extremely novel, so why has it failed to catch on? In lecture, we learned that “market success often depends more on consumer convenience, price, and content availability than on technical superiority”. Reflecting on my own use of VR, my most played games are Golf+ and Eleven: Table Tennis, both sports simulators that closely mimic real-life sports. Indeed, most of the highest-selling games are very simple and emulate real sports or activities. This has a straightforward reason: as a game gets more complex, it becomes more expensive and more likely to cause motion sickness. So, while Meta made an engineering beauty, the limitations of its product mean it fails to offer novelty to the consumer. This mistake was noticed by Apple who took a new route, mixed reality. Mixed reality goggles offer VR plus Augmented reality (AR), the ability to integrate apps into real life visuals. This invention is not nearly as impressive of a technological leap but offers much more novelty to the consumer since it can be used in situations outside of gaming. For example, you could use the device to seamlessly watch a Zoom presentation while walking to class. Further, it has the potential to bring an innovation from the entertainment industry to several new industries (AR could be used as a work assistant or training tool). Interestingly, Apple has also failed to capture a significant market, though the high price is usually blamed. This development demonstrates the importance of creating affordable novelty for the consumer when innovating. Inspired by the slow adoption of VR, The equation below attempts to capture the factors that influence adoption rates of cutting-edge technology. Adoption= Functionality + Quality + Novelty + Consumer Convenience + 1/Price Functionality is the same value for all inventions of a similar type (e.g. all AI models would have the same functionality value) and reflects how useful that invention is to society (e.g. functionality of AI>>>VR). Quality will mean different things for different industries. For example, in VR quality might reflect the number of apps available and picture resolution while in AI quality might reflect model speed and response accuracy. Applying this equation to the VR space, Apple realized that the novelty of current VR headsets was too low and tried to push a new, mixed-reality paradigm. While they succeeded in raising novelty, they also raised price which removed any adoption growth possibility. I predict that as Apple lowers price of the VisionPro (if possible), they could supplant Meta as the primary device maker. |
How Technology’s Modular Structure Propelled AI to New Heights Exponentially Faster than Other Technologies Chapter 2 in Arthur’s The Nature of Technology describes the concept of modularity as one of the key components for the structural integrity within any given technology. We learn that technologies are composed of different assemblies and subassemblies, and such a structure permits the technology at hand to thus execute and perform its function. Artificial Intelligence, by definition of being a technology, thus contains this modular structure. But why is it that AI seems to have suddenly boomed at such an exponential rate compared to other, perhaps more “physical” technologies? At the general level, Arthur would argue that AI and any other technology ultimately have the same structure, so what, if anything, is unique about AI’s modular structure that is acting as a catalyst for the technology’s recent success? The answer lies in the fact that reconfiguring AI’s modular components in order to experiment with different versions of the same model and find the optimal one, actually turns out to be an incredibly efficient process that continues to experience decreasing costs as computers become more and more advanced in their ability to process large amounts of data. To make this very clear, let’s take two technological examples: the transformer architecture of large language models like OpenAI’s GPT, and the hierarchical structure of the F35C aircraft Arthur describes. As computers become even more powerful, not only can these AI models accept new, and likely improved, data as initial inputs at the lowest level of the structure of the model, but it can also experiment with dozens, hundreds, thousands, even millions of different parameters as part of the next layers within the modular structure. This, along with increased computational power, provides for rapid prototyping and experimentation of hundreds of different AI “models” that can all be performed within a matter of a few days, hours, or even minutes depending on the complexity and size of the models. Now imagine trying to even experiment with just two or three different aircraft models. Simply from a cost standpoint of having to either build two or three different aircrafts, or build one, test it, make changes, test it again, make changes, test it again, and so forth, is ineffective both timewise and economically speaking. The ability for AI models to seamlessly experiment with so many versions of the same technology is what allows for innovation to really take place within the AI world, and thus proves how its modular structure differentiates itself from that of other technologies. The three graphs below illustrate the exponential growth seen in recent years regarding key components of AI’s modular structure: data points, parameters, and computational ability, all of which have contributed to AI’s rapid emergence and progression as the technology of the future. Citations: ![]() ![]() ![]() |
Many of the readings this week focus on the concept of how to identify degrees of novelty of an innovation. Whether it be the distinction between context vs. content, point vs. application vs. system solution, or higher vs. lower order recombination, all of these categorizations demonstrate that the least predictable (and thus the most innovative) idea is one that combines modular knowledge from a variety of disciplines to create new systems of understanding or experiencing the world. In essence, true creativity is unpredictable, or predicted to be improbable. Thus, in the age of AI, we should truly think about how we can best incentivize and educate people to think creatively. Going back to the topic of my first memo, there are many ways in which AI can completely revolutionize the education system. A lower order recombination of AI is already in use in some classrooms today. They include point solutions like customized lesson plans and/or assessments, graders, etc. These applications are meant to improve teaching efficiency and increase personalization in learning, which should lead to better quality of education for the student. However, with the existence of ChatGPT and other LLMs, most pre-university level homework and assessments have been rendered utterly useless. There should be fundamental change to the education system, from the foundational curriculum to the way in which students are assessed. How can we use AI as a tool to encourage students to think critically and creatively, with an emphasis on their ability to synthesize knowledge in different fields? We can take inspiration from the model in “Modularity, Higher-Order Recombination, and New Venture Success” to assess the degree to which school curriculums and textbooks are facilitating creativity and pushing students to engage in higher order recombination, in the same way that stocks can be assessed on innovative value based on semantic analysis. For example, we can create dynamic word embeddings of textbooks to analyse the semantic space. How are math textbooks connecting to other topics like science, history, or music? Or we can create embeddings of searches made by young students to actively incorporate topics they are interested in. The figure below is a mapping of how people’s interests may change over time, and how our education system should be evolving to capture these forefront topics. In this sense, AI can help facilitate the creative destruction of “unhelpful” instructional materials—ones that have limited semantic space and do not help students make connections between theoretical subjects and the world around them. Similarly, students can be assessed on their ability to form higher order recombinations in course projects. Rather than a focus on facts or knowledge, the appearance of AI should shift education towards what really matters: thinking logically and creatively. Empirically, we could design two types of courses for students, varying in content novelty but similar in technical difficulty, then assess their creativity (probability of surprise) and ability to make associations and recombinations. |
Connecting VC Funding with Prior Discussions of InnovationIn "Modularity, Higher-Order Recombination, and New Venture Success", Cao, Chen and Evans, argue that venture capital structurally rewards higher order innovation. Startups without sufficiently novel products tended to struggle to in obtaining further funding rounds in the high-intensity venture capital ecosystem. However, venture capital is not an ubiquitous asset class on the global scale. In Europe, high income economies notably do not put as much weight as the US on the asset class. According to the Ivey Business Journal, Germany and Switzerland have relatively low VC capital per capita and other countries like France have unique regulations on venture capital which operate counterproductive to fundraising levels. As discussed in past weeks, these countries are often on par with the US in terms of innovation quantity. In this memo post, I analyze the relationship between VC funding and innovation in terms of patent applications. This graphic describes the per capita VC funding for 24 countries (selected by Crunchbase to represent regions with highest VC potential in coming years). These figures match the observations made by the Ivey Business Journal. Apart from Estonia, Sweden, and the UK, European countries have significantly lower VC spending than the US and other top performers such as Singapore and Israel. However, adding per capita patent data (from the World Bank) into the graph reveals that patent application per capita, which we've used in previous weeks as a proxy for a country's innovativeness, is largely uncorrelated with venture capital. A potential explanation is the influence of incumbents on innovation. In the case of Germany, which tracks the US closely, the majority of patents were filed by incumbents rather than venture-backed startups. Preliminary research indicates that the same is true for a lot of the other European countries without strong startup ecosystems. Another interesting example to examine further is Estonia. Despite having an extremely strong VC ecosystem in terms of funding, per capita patents is relatively low compared to many of the other countries analyzed. This could be a result of Estonia's competitive tax system relative to the rest of the EU and other OECD countries, which would also explain Ireland's position in the graph. Due to lower taxes, foreign based startups (which countribute patents to other countries) may opt to raise funds in these countries. On the other end of the spectrum, China and South Korea have low VC funding per capita but exceptionally high patent applications. The incumbent explanation doesn't seem to make much sense especially in the case of China, which does not have the long-standing institution base found in Europe. In coming weeks, these would serve as potential case studies to look deeper into. Furthermore, patents are not a perfect metric to capture the quality of innovation. Even though venture-based higher order innovation is the source of much of US patents, further examination is needed to determine the character of patent innovation in other environments. Is incumbent based innovation in Europe likely to be higher order due to already existing infrastructure which innovators can build off of? Do higher patent application volumes in Asia reflect higher levels of creative destruction, or do they result from government incentives to reach certain research goals? |
This week’s reading on “Novelty” as a violation of expectation, especially the idea of assigning novelty scores to various innovations across fields based on combinatorial characteristics of innovations’ peculiarity, atypicality, and obscurity prompted me to think of other characteristics or models that could be indicative/ predictive of impact. I completely agree that when thinking about aspects of novelty in present day innovation, the interpretations of “novelty” are very subjective to the context of the field that the novel innovation is introduced in. Most of the models in the paper seem to account for the subjectivity and variability of various aspects potentially affecting the novelty scores in each distinct field. However, I also believe that there is a general, categorical element of novelty, that can be applied across all fields, which I would like to propose: Historically, when I think about the timeline of inventions, innovations, moments of growth that I have both lived through and learned about throughout my primary and higher education, I see the timeline as the following categorical progression, listed from the furthest distance in time, to the nearest distance in time to present day: Practical Discoveries and innovations for the foundations of human survival (more like tools, not related to “modern technology” in our most contemporary definition of the term). One example of this is the discovery of fire, and the fact that we can create it ourselves, and use it to keep us warm and alive, or cook our food, or sterilize and heal our wounds. Some similar such innovations that followed are the invention of the wheel, which completely revolutionized human transportation. I believe that present day innovations in category 4 have the lowest novelty factors. I believe that it is relatively least shocking to predict ahead chronologically, as for years people have had the ideas (but just not the means) to create the things that they imagine, such as flying driving cars/ various types of robots, etc. It is most natural for us to think ahead from the most recent discoveries which have been built on centuries and centuries of previous fundamental discoveries. But what would be most unexpected is to predict a type of invention or breakthrough that can be thought of as going “backwards” in the above timeline. No one would be able to predict a new invention such as the discovery of a new natural element equivalent to fire, because we believe that we have already mostly discovered what there is to discover in that area— such an invention would have the highest novelty score of the 4 categories. Discovering a new fundamental truth about our world that changes our perception and way of life altogether, such as (for example, what happens after humans die/ empirical evidence of an afterlife), would have the second highest novelty score in this spectrum. I would propose the following equation which acts as a novelty score multiplier, that can be incorporated into each of the model and field specific novelty equations proposed by the paper. Novelty Score Multiplier = C x P Where C is a numeric weight based on the above 4 categories: (1)Practical Discoveries for foundations of human survival: C = 8 , and P is an estimated percentage of the global population which the innovation will directly affect and be easily accessible. Ex: a category 1 innovation would be between 0.9 and 1. *The C values assigned to the 4 categories are relatively arbitrary at this point and can be adjusted so that they are non linear. *The exact method for the calculation of P would have to be refined and clearly defined to include various aspects of global reach across the environments in different regions of the world. |
Innovation in Parallel and Preemptive Combinatorial Evolution I found the demand-side story of innovation in Brian Arthur’s book The Nature of Technology: What it is and How it Evolves ripe for further exploration. Arthur argues that technologies emerge to fill opportunity niches that are created by human needs, and more prominently, technology itself. In particular, he elaborates that “[firstly,] every technology by its very existence sets up an opportunity for fulfilling its purpose more cheaply or efficiently,…[secondly,] every technology requires supporting technologies to manufacture it, organize for its production and distribution, maintain it, and enhance its performance,...[and thirdly,] technologies often cause problems–indirectly–and this generates needs, or opportunities, for solutions” (175-176). This recursive nature of technology demand prompted me to consider the repercussions of “combinatorial evolution” happening in parallel. For example, what happens when innovation actors begin to assume that a recursive chain of technologies will constantly improve? One case study lies in semiconductor design and manufacturing. Back in 1965, semiconductor designer Gordon Moore observed that the number of components per integrated circuit had been doubling every year, spawning the now infamous “Moore’s Law” that projects the number of transistors in an IC doubling every two years. Moore’s Law has held true up until now, and some argue that there has been a “self-fulfilling prophecy” dynamic at play. As chips became more complex, technologists at each stage of the chip manufacturing process set their long-term R&D planning according to this coordinated target. I argue that this behavior demonstrates the existence of preemptive combinations, which I sketched in the graphic below in the spirit of “Assembly theory explains and quantifies selection and evolution”. The innovations across lithography, ion implantation, deposition, etching, packaging, and myriad other stages of the process were not solely combinations of innovations in the past; they also relied heavily on innovations being designed in lockstep. Thus, innovation C (say, EUV lithography) depended not just on prior innovations A and B, but also on concurrent innovations D and E (e.g. new masking and photoresist innovations). While skeptics could argue that each new “node” or semiconductor manufacturing process like 3nm is one cohesive innovation, the reality is that the innovations at each stage of the process are applied to other use cases. In this way, coordination or preemption of adjacent innovations at the same number of assembly steps jumpstarts distinct innovations in parallel. Another relevant example of this innovation in parallel, or preemptive combinatorial evolution, is the current AI scaling law. At every level of the AI stack, from semiconductors to foundation models to AI applications, technologists are innovating with the assumption that the other pieces will innovate and come into place as well. Take foundation model developers, who are improving their data and algorithms with the assumption that more effective chips will be harnessable in the near future. This model of innovation demonstrates how technology’s recursive loop can become a formidable flywheel if technologists coordinate to innovate in parallel. ![]() |
Do Industry Characteristics Shape the Success of Novel Ideas?This week, we explored the relationship between novelty and impact in the context of innovation. For example, a study titled "Surprising combinations of research contents and contexts are related to impact and emerge with scientific outsiders from distant disciplines" finds that two dimensions of novelty—context novelty and content novelty—strongly predict the impact of both patents and scientific research. However, during class, Professor Evans raised an intriguing caveat: while the most novel innovations are, on average, more successful, their success is also marked by greater variance. Specifically, he noted that highly novel innovations are more likely to face significant challenges in implementation or paradigm transition. As a result, despite being objectively superior or even game-changing, they might fail entirely. This was a fascinating observation, but it made me wonder whether it applies equally across all industries. Some industries are characterized by more entrenched systems, often due to the complexity and scale of the physical capital required. In such cases, I could imagine that even lower levels of innovation might result in significant variance, as incremental changes are more likely to be selectively adopted or ignored, while radical changes face uniform resistance. For instance, Brian Potter argues that innovation in construction is particularly challenging because project costs are right-skewed and fat-tailed, driven by the compounding effects of delays and other factors. Novel systems that could dramatically alter construction processes often incur steep risk penalties that outweigh their expected benefits. As a result, innovations in construction tend to be more "incremental and evolutionary." Validating this theory fully would require more time and data than I had for this memo, but I conducted a preliminary investigation into some relevant evidence. As mentioned in class, venture capitalists tend to pursue the most novel innovations, accepting that most of their investments will fail in exchange for a few outliers that generate outsized returns. Given this, the industries that VCs favor might offer insight into how tangibility (as a proxy for capital requirements) relates to novelty and variance. Plotting the data, I found that nearly five times as much VC investment goes toward intangible goods and services—industries with lower capital requirements—compared to tangible goods. While this is an interesting result, I recognize its limitations, particularly the influence of confounding factors such as the higher profit margins associated with intangible goods, which don’t necessarily speak to variance in innovation success within those categories. Given more time, it would be interesting to use more sophisticated techniques, such as analysis of patent keywords, as well as higher quality and more comprehensive data, to see if the relationship between novelty and variance holds consistently, or whether it differs based on the structural characteristics of an industry. |
The reading “Surprise! Measuring Novelty as Expectation Violation” provides an overview of various methods for measuring novelty and introduces a framework based on the concept of surprise associated with invention. Novelty is significant because it positively correlates with the variance in success outcomes. Within the surprise framework, I wondered: when medical drugs that have not been used before are introduced, are they still considered innovations if they lack the element of surprise? For example, the COVID-19 vaccine dominated its domain, changed the landscape, and was widely adopted, yet it seemed to be an expected development rather than a surprising one. Another point mentioned in class was how drug prices typically drop once a patent expires, as heightened competition diminishes inventor’s market dominance. So I became interested in how quickly prices respond to patent expirations. Specifically, I wanted to analyze the statistics on the average prices of drugs whose patents expired in recent years. This is interesting because the number of patents increased significantly between 1996 and 2003 (as shown in Figure 1), meaning many of these patents were supposed to expire within the past eight years. This phenomenon of the abrupt drop in sales and market share loss that often follows patent expiration is known as the "patent cliff". For this analysis, I used data from Medicaid and the Children's Health Insurance Program on the weekly average prices pharmacies pay for prescription drugs from 2013 to 2025 (available at https://data.medicaid.gov/dataset/a217613c-12bc-5137-8b3a-ada0e4dad1ff#data-dictionary). I focused on the average price per milligram or per dose, as some drugs come in different dosages. I aimed to select drugs with patents that expired in recent years, but it was difficult to search for patent expiration dates. To overcome this, I used several websites listing expired patents and cross-referenced the information. In the end, I analyzed four drugs (see Figure 2), with red lines indicating the approximate dates of patent expiration or the dates of allowed realize of the generic.
Overall, this highlights the significant impact of patent expirations on drug prices, showing how the introduction of generics right after the patent expiration often leads to price drops. A major limitation of this analysis is that, since most drugs are protected by multiple patents, they do not all expire at once, making it difficult to pinpoint a single expiration date. Additionally, finding the exact patent expiration dates was surprisingly challenging, and for some drugs, the available price data was too limited to include in the analysis. ![]() Figure 1: Data from [5] ![]() Figure 2 [1] Biopharma Dive. (2024, January 23). Bristol Myers’ Sprycel faces generic competition as patent cliff approaches. Biopharma Dive. https://www.biopharmadive.com/news/bristol-myers-sprycel-generic-patent-cliff-xspray/721197/ |
Memo: Transformer-Based AI, Novelty, and Cross-Disciplinary Innovation Recent scholarship (Shi & Evans 2023; Foster, Shi & Evans 2020) underscores how “surprising” combinations of ideas from distant disciplines are prime engines of high-impact innovation. Building on Brian Arthur’s (2009) view that novel technologies evolve by combining existing building blocks, modern large language models (LLMs)—rooted in Transformer architectures (Evans & Kozlowski 2025)—offer a powerful technological catalyst for precisely these unexpected, high-value recombinations. At a high level, Transformers represent text (or other data) as vectors in a high-dimensional space, allowing them to capture links across seemingly unrelated domains. In principle, this creates fertile ground for “emergent” insights, since architectural mechanisms like attention can integrate tokens from previously unconnected topics. This capacity dovetails with Assembly Theory (Sharma et al. 2023), wherein the recombination of simpler “parts” can open doors to novel, complex outcomes. However, empirical evidence for LLM-driven innovation remains scattered, prompting a need to measure the extent to which LLMs facilitate cross-domain connections that surprise even human experts. One way to do this is by adapting the “novelty-as-expectation-violation” framework (Foster et al. 2020) to compare an LLM’s predicted connections to real-world disciplinary overlaps. Concretely, define a Surprise Score (S) for an idea ( x ) as: S(x) = | LLM(x) - HumanBaseline(x) | / HumanBaseline(x) where LLM}(x) is the model’s estimated “distance” between fields (e.g., robotics and microbiology) and HumanBaseline}(x) is a consensus measure of how distant experts judge those same fields. Higher S(x) indicates that the model identifies unusually surprising linkages. Figure: Hypothetical Surprise Scores for Interdisciplinary AI Suggestions
In this mock plot, each point denotes a pair of fields (e.g., Robotics + Microbiology). The y-axis marks the LLM’s Surprise Score, and the x-axis represents an objective measure of disciplinary distance. Notice how “Robotics + Microbiology” registers a higher Surprise Score than “Marketing + Chemistry,” implying that the AI finds more novel synergies in the former pairing than do human experts. This simple metric offers a concrete, tech-focused pathway to identify latent cross-disciplinary breakthroughs within AI outputs. It builds on Arthur’s concept of evolving technologies through recombination and aligns with the finding that “scientific outsiders” can foster especially impactful novelty (Shi & Evans 2023). As generative models mature, measuring and steering them toward surprising, interdisciplinary suggestions could greatly amplify their potential to drive meaningful innovation and growth. |
Agriculture remains one of the cornerstones of the African economy, inextricably linked with livelihoods and food security that underpin economic growth. Despite such importance, it has long been characterized by systemic challenges, including outdated technologies, inefficient methods of production, restricted access to resources, and susceptibility to climate change. These persistent issues have hampered productivity and restricted farmers' ability to scale operations or compete on a global level. But over the last couple of years, AI and automation have started to emerge as real game-changers: new ways in which these barriers can be tackled while fostering innovation and sustainability. This memo looks into the role of AI in African agriculture, as discussed in a project report titled "Inclusively Advancing Agri-Food Systems through AI and Automation", focusing on its application across the agricultural value chain, as well as at the Ethiopian Agricultural Transformation Agency's 8028 Farmer Hotline for an example of how AI works in practice. The Generalized Agricultural Value Chain and Use Case Framework (Figure 2) puts a helpful lens on the broad applications of AI in agriculture, highlighting how these technologies are integrated across different areas of the farming process. Genomic innovation and digital financial advisory systems are aiding farmers in their decisions regarding inputs, planting strategies, and financial management at the Planning & Production stage, with reduced risk and improved productivity. Automation makes the chain more efficient in the phase of post-harvest management where food waste is minimized and value is maximized by demand-supply matching and automated processing. In the Distribution & Logistics section, innovation is projected in the form of smart contracts and automated transportation systems, which will chart the clear line of supply chains, therefore minimizing inefficiencies. Connectedly, AI-powered solutions link these stages together to create a well-organized structure for agricultural production that addresses long-standing inefficiencies while opening up new avenues for growth. One particularly inspiring example is Ethiopia’s 8028 Farmer Hotline, which shows how much of a difference these technologies can make for smallholder farmers—the people who often struggle the most with accessing resources and information. The hotline, run by the Ethiopian Agricultural Transformation Agency, uses SMS and Interactive Voice Response (IVR) to provide farming advice in local languages, tailored to specific crops and regions. It even collects feedback from farmers and combines it with satellite data to deliver early warnings about crop diseases or pests, giving farmers time to act before disaster strikes. Since it started in 2014, the hotline has reached over 6.2 million registered users and handled more than 63 million calls. That’s not just impressive—it’s life-changing for the farmers who can make more informed decisions and avoid devastating losses. What really stands out when looking at the value chain framework and the success of the 8028 Farmer Hotline is how they show both the big picture and the on-the-ground impact of AI in agriculture. These aren’t just abstract ideas—they’re real solutions addressing real problems. However, much more needs to be done for this potential to reach everyone, especially smallholder farmers. Governments and organizations need to invest in infrastructure like reliable electricity and internet access, while also providing training to ensure farmers know how to use these tools. And policymakers have a responsibility to ensure the fair distribution of AI benefits, so small farmers aren’t left behind while bigger farms take all the gains. AI and automation are truly reshaping agriculture across Africa, and it’s not just about fixing inefficiencies—it’s about creating a future where farmers can thrive, communities can be more resilient, and economies can grow in ways we haven’t seen before. Ethiopia’s experience with the 8028 Farmer Hotline proves that when the right technology meets the right support systems, the results can be transformative. Scaling these innovations across the continent, thoughtfully and equitably, could unlock an entirely new chapter for African agriculture—one that’s smarter, more sustainable, and more inclusive for everyone involved. |
Post your memo in response any (or all) of the week's readings and an empirical case regarding artificial intelligence, innovation, and/or growth:
Post by Thursday @ midnight. By 1pm Friday, each student will up-vote (“thumbs up”) what they think are the five most interesting memos for that session. The memo should be 300–500 words (text) + 1 custom analytical element (e.g., equation, graphical figure, image, etc.) that supports or complements your argument. These memos should: 1) test out ideas and analyses you expect to become part of your final projects; and 2) involve a custom (non-hallucinated) theoretical and/or empirical demonstration that will result in the relevant analytical element. Because these memos relate to an empirical case students hope to further develop into a substantial final project and because they involve original analytical work, they will be very difficult to produce with generative AI and we strongly discourage you from attempting it. Some of the top-voted memos will form the backbone of discussion in our full class discussion and break-out room sessions.
The text was updated successfully, but these errors were encountered: