When I think about trying to forecast technology for the medium term future, especially for AI/AGI progress, it often crosses a bunch of technical boundaries.
These boundaries are interesting in part because they’re thresholds where my expertise and insight falls off significantly.
Also interesting because they give me topics to read about and learn.
A list which is probably neither comprehensive, nor complete, nor all that useful, but just writing what’s in my head:
Machine learning research—this is where a lot of the tip-of-the-spear of AI research is happening, and seems like that will continue to the near future
Machine learning software—tightly coupled to the research, this is the ability to write programs that do machine learning
Machine learning compilers/schedulers—specialized compilers translate the program into sequences of operations to run on hardware. High quality compilers (or the lack of them) has blocked a bunch of nontraditional ML-chips from getting traction.
Supercomputers / Compute clusters—large arrays of compute hardware connected in a dense network. Once the domain of government secret projects, now it’s common for AI research companies to have (or rent) large compute clusters. Picking the hardware (largely commercial-off-the-shelf), designing the topology of connections, and building it are all in this domain.
Hardware (Electronics/Circuit) Design—A step beyond building clusters out of existing hardware is designing custom hardware, but using existing chips and chipsets. This allows more exotic connectivity topologies than you can get with COTS hardware, or allows you to fill in gaps that might be missing in commercially available hardware.
Chip Design—after designing the circuit boards comes designing the chips themselves. There’s a bunch of AI-specific chips that are already on the market, or coming soon, and almost all of them are examples of this. Notably most companies that design chips are “fabless”—meaning they need to partner with a manufacturer in order to produce the chip. Nvidia is an example of a famous fabless chip designer. Chips are largely designed with Process Design Kits (PDKs) which specify a bunch of design rules, limitations, and standard components (like SRAM arrays, etc).
PDK Design—often the PDKs will have a bunch of standard components that are meant to be general purpose, but specialized applications can take advantage of more strange configurations. For example, you could change a SRAM layout to tradeoff a higher bit error rate for lower power, or come up with different ways to separate clock domains between parts of a chip. Often this is done by companies who are themselves fabless, but also don’t make/sell their own chips (and instead will research and develop this technology to integrate with chip designers).
Chip Manufacture (Fabrication / Fab) - This is some of the most advanced technology humanity has produced, and is probably familiar to many folks here. Fabs take chip designs and produce chips—but the amount of science and research that goes into making that happen is enormous. Fabs probably have the tightest process controls of any manufacturing process in existence, all in search of increasing fractions of a percent of yield (the fraction of manufactured chips which are acceptable).
Fab Process Research—For a given fab (semiconductor manufacturing plant—“fabricator”) there might be specializations for different kinds of chips that are different enough to warrant their own “process” (sequence of steps executed to manufacture it). For example, memory chips and compute chips are different enough to need different processes, and developing these processes requires a bunch of research.
Fab “Node” Research—Another thing people might be familiar with is the long-running trend for semiconductors to get smaller and denser. The “Node” of a semiconductor process refers to this size (and other things that I’m going to skip for now). This is separate from (but related to) optimizing manufacturing processes, but is about designing and building new processes in order to shrink features sizes, or push aspect ratios. Every small decrease (e.g. “5nm” → “3nm”, though those sizes don’t refer to anything real) costs tens of billions of dollars, and further pushes are likely even more expensive.
Semiconductor Assembly Research—Because we have different chips that need different processes (e.g. memory chips and compute chips) -- we want to connect them together, ideally better than we could do with just a circuit board. This research layer includes things like silicon interposers, and various methods of 3D stacking and connection of chips. (Probably also should consider reticle-boundary-crossing here, but it kinda is also simultaneously a few of the other layers)
Semiconductor Materials Science—This probably should be broken up, but I know the least about this. Semiconductors can produce far more than chips like memory and compute—they can also produce laser diodes, camera sensors, solar panels, and much more! This layer includes exotic methods of combining or developing new technologies—e.g. “photonics at the edge”—a chip where the connections to it are optical instead of electronic!
Anyways I hope that was interesting to some folks.
AGI technical domains
When I think about trying to forecast technology for the medium term future, especially for AI/AGI progress, it often crosses a bunch of technical boundaries.
These boundaries are interesting in part because they’re thresholds where my expertise and insight falls off significantly.
Also interesting because they give me topics to read about and learn.
A list which is probably neither comprehensive, nor complete, nor all that useful, but just writing what’s in my head:
Machine learning research—this is where a lot of the tip-of-the-spear of AI research is happening, and seems like that will continue to the near future
Machine learning software—tightly coupled to the research, this is the ability to write programs that do machine learning
Machine learning compilers/schedulers—specialized compilers translate the program into sequences of operations to run on hardware. High quality compilers (or the lack of them) has blocked a bunch of nontraditional ML-chips from getting traction.
Supercomputers / Compute clusters—large arrays of compute hardware connected in a dense network. Once the domain of government secret projects, now it’s common for AI research companies to have (or rent) large compute clusters. Picking the hardware (largely commercial-off-the-shelf), designing the topology of connections, and building it are all in this domain.
Hardware (Electronics/Circuit) Design—A step beyond building clusters out of existing hardware is designing custom hardware, but using existing chips and chipsets. This allows more exotic connectivity topologies than you can get with COTS hardware, or allows you to fill in gaps that might be missing in commercially available hardware.
Chip Design—after designing the circuit boards comes designing the chips themselves. There’s a bunch of AI-specific chips that are already on the market, or coming soon, and almost all of them are examples of this. Notably most companies that design chips are “fabless”—meaning they need to partner with a manufacturer in order to produce the chip. Nvidia is an example of a famous fabless chip designer. Chips are largely designed with Process Design Kits (PDKs) which specify a bunch of design rules, limitations, and standard components (like SRAM arrays, etc).
PDK Design—often the PDKs will have a bunch of standard components that are meant to be general purpose, but specialized applications can take advantage of more strange configurations. For example, you could change a SRAM layout to tradeoff a higher bit error rate for lower power, or come up with different ways to separate clock domains between parts of a chip. Often this is done by companies who are themselves fabless, but also don’t make/sell their own chips (and instead will research and develop this technology to integrate with chip designers).
Chip Manufacture (Fabrication / Fab) - This is some of the most advanced technology humanity has produced, and is probably familiar to many folks here. Fabs take chip designs and produce chips—but the amount of science and research that goes into making that happen is enormous. Fabs probably have the tightest process controls of any manufacturing process in existence, all in search of increasing fractions of a percent of yield (the fraction of manufactured chips which are acceptable).
Fab Process Research—For a given fab (semiconductor manufacturing plant—“fabricator”) there might be specializations for different kinds of chips that are different enough to warrant their own “process” (sequence of steps executed to manufacture it). For example, memory chips and compute chips are different enough to need different processes, and developing these processes requires a bunch of research.
Fab “Node” Research—Another thing people might be familiar with is the long-running trend for semiconductors to get smaller and denser. The “Node” of a semiconductor process refers to this size (and other things that I’m going to skip for now). This is separate from (but related to) optimizing manufacturing processes, but is about designing and building new processes in order to shrink features sizes, or push aspect ratios. Every small decrease (e.g. “5nm” → “3nm”, though those sizes don’t refer to anything real) costs tens of billions of dollars, and further pushes are likely even more expensive.
Semiconductor Assembly Research—Because we have different chips that need different processes (e.g. memory chips and compute chips) -- we want to connect them together, ideally better than we could do with just a circuit board. This research layer includes things like silicon interposers, and various methods of 3D stacking and connection of chips. (Probably also should consider reticle-boundary-crossing here, but it kinda is also simultaneously a few of the other layers)
Semiconductor Materials Science—This probably should be broken up, but I know the least about this. Semiconductors can produce far more than chips like memory and compute—they can also produce laser diodes, camera sensors, solar panels, and much more! This layer includes exotic methods of combining or developing new technologies—e.g. “photonics at the edge”—a chip where the connections to it are optical instead of electronic!
Anyways I hope that was interesting to some folks.