Skip to content

Commit

Permalink
Update 2024.09.impact.md
Browse files Browse the repository at this point in the history
  • Loading branch information
okhat authored Sep 4, 2024
1 parent 416ff8f commit 4d9b70f
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion 2024.09.impact.md
Original file line number Diff line number Diff line change
Expand Up @@ -97,4 +97,4 @@ When you read the previous guideline (#5), it's very natural to wonder: Where mi

The answer in practice is that it's possible for most of your time OSS time to be spend on doing new, exciting research. In fact, a major advantage of the style of research in this guide is that it creates recognizably important problems in which you have a very large competitive advantage. This is true in terms of the intuitive recognition of problems extremely early, having a far more instinctive understanding of the problem than others, having direct feedback on prototypes of your approaches, having access to wonderful collaborators who understand the significance of the problems, and also having great "distribution channels" that ensure that every new paper you do in this space will have a receptive audience and will reinforce your existing platform.

Just to illustrate, ColBERT isn't one paper from early 2020. It's probably around ten papers now, with investements into improved training, lower memory footprint, faster retrieval infrastructure, better domain adaptation, and better alignment with downstream NLP tasks. Similarly, DSPy isn't one paper but is a large collection of papers on programming abstractions, prompt optimization, and downstream programs. So many of these papers are written different, amazing primary authors. A good open-source artifact creates modular pieces that can be explored, owned, and grown by new researchers and contributors.
Just to illustrate, ColBERT isn't one paper from early 2020. It's probably around ten papers now, with investements into improved training, lower memory footprint, faster retrieval infrastructure, better domain adaptation, and better alignment with downstream NLP tasks. Similarly, DSPy isn't one paper but is a large collection of papers on programming abstractions, prompt optimization, and downstream programs. So many of these papers are written by [different, amazing primary authors](https://github.com/stanfordnlp/dspy?tab=readme-ov-file#dspy-programmingnot-promptingfoundation-models). A good open-source artifact creates modular pieces that can be explored, owned, and grown by new researchers and contributors.

0 comments on commit 4d9b70f

Please sign in to comment.