Silicon-Based vs Carbon-Based
Computing
step-by-step explanation of how silicon-based vs.
carbon-based computing works for AI, and why the "move" or
interaction isn't seamless yet.
1. The
Basics: How Any Computing Hardware Works (Silicon or Carbon)
AI algorithms (e.g., neural networks) are just math: lots of
additions, multiplications, and activations on numbers (weights and
activations). Hardware executes these as binary logic (0s and 1s) using transistors
— tiny electronic switches.
- A
transistor acts like a gate: it controls whether current flows (1) or not
(0) based on voltage.
- Billions
of transistors form logic gates (AND, OR, etc.), which build adders,
multipliers, memory, etc.
- The substrate
material (silicon, carbon nanotubes, etc.) determines how well
electrons move, how small/fast/reliable the switches can be, and how much
power/heat they use.
2.
Silicon Transistors (Current Standard)
- Structure:
In a silicon Field-Effect Transistor (FET), a gate voltage controls a
channel in doped silicon between source and drain. Electrons (or holes)
flow when the gate allows it.
- How
it works for AI: GPUs pack millions of these into cores optimized for
parallel matrix math. Data moves between memory and processors (the
"von Neumann bottleneck").
- Strengths:
Extremely mature manufacturing (photolithography on wafers), reliable,
cheap at scale.
- Limits:
As transistors shrink, quantum tunneling (electrons leaking), heat, and
power waste increase. Data movement eats most energy in AI.
3.
Carbon-Based (e.g., Carbon Nanotube — CNT) Transistors
Carbon nanotubes are rolled-up sheets of graphene
(single-layer carbon atoms in a hexagonal lattice) — tiny cylinders, ~1-2 nm
diameter.
- How a CNT transistor works:
- The
channel is a carbon nanotube instead of silicon.
- Electrons
move ballistically (with very little scattering/resistance) due to
the perfect structure — much faster and more efficiently than in silicon.
- Gate
voltage still controls the flow, but CNT versions can switch at lower
voltages, with steeper on/off transitions (less power waste in the
"gray" area).
- Advantages:
- Higher
carrier mobility → faster operation.
- Lower
power + less heat.
- Can
be made smaller and stacked in 3D more easily (lower processing
temperatures don't melt lower layers).
Recent
Chinese carbon-based AI chip example (Peking University et al.):
- Uses
CNT transistors with a special "source-gated" design.
- Creates
a controllable p-n junction by extending the source electrode into the
channel.
- This
allows ternary logic (three states: e.g., 0, 1, 2 or low/medium/high
current) instead of just binary (0/1).**
- Why
better? One ternary digit holds more information than a binary bit →
fewer operations/steps for the same computation, higher density, lower
energy for tasks like image recognition (which they demonstrated
perfectly).
4. How
AI Algorithms "Move" Between Them
- The
algorithms don't change — software (PyTorch, TensorFlow, etc.)
compiles to instructions the hardware understands.
- On
new hardware, you need new compilers, drivers, and optimizations.
- Transition
path: Start with hybrid systems (some CNT or graphene layers on
silicon chips) or accelerators.
5.
What Stops Full Interaction/Transition?
- Fabrication
nightmare: Aligning trillions of perfect CNTs on a wafer without
defects or metallic impurities is very hard (metallic ones short-circuit).
Silicon is uniform and predictable.
- Ecosystem:
Decades of tools, fabs, and expertise are built for silicon.
- Reliability
& variability: Early carbon devices can be inconsistent.
- For
biological carbon (real neurons): Completely different — wet, chemical
(ions, neurotransmitters), slow but ultra-efficient and adaptive.
Interfacing requires converting electrical signals to biochemical ones
(via BCIs). Not yet scalable for general AI.
Bottom line: Silicon won historically due to
engineering practicality. Carbon (CNTs/graphene) is a promising successor for
speed, efficiency, and scaling AI compute, with working prototypes already
doing AI tasks. Full replacement will come gradually through hybrids as
manufacturing improves. The "interaction" is advancing via better
materials and brain-computer bridges, but physics + economics set the pace.
Ternary logic
** Ternary
logic (three states: 0, 1, 2) in the Chinese carbon-based AI chip is
a key innovation that goes beyond traditional binary computing. Here's a clear
explanation of how it works.
Binary vs. Ternary Logic — The Core Difference
- Binary
(standard in silicon chips): Each bit has 2 states — 0 (low
voltage/off) or 1 (high voltage/on). All data, math, and logic build from
this. A single wire carries 1 bit of information.
- Ternary:
Each "trit" (ternary digit) has 3 states — typically 0
(low), 1 (middle), and 2 (high). One wire now carries ~1.58× more
information (log₂(3) ≈ 1.58 bits).
Benefits:
- More
efficient data representation → fewer operations, less memory, and lower
energy for the same computation.
- Simpler
circuits for some math (e.g., addition/multiplication can need fewer
steps).
- Higher
information density in the same physical space.
Real-world analogy: Binary is like a light switch (off/on).
Ternary is like a dimmer switch with three clear positions (off/low/high).
How the Chip Achieves Reliable Ternary States
Standard transistors are good at two clear states.
The Chinese team (Peking University + Beijing University of Posts and
Telecommunications) used Carbon Nanotube (CNT) Source-Gated Transistors
(SGTs) to create stable three-state behavior.
Key design trick:
- They
extend the source electrode into the channel of the CNT transistor.
- This
creates a controllable p-n homojunction (a junction within the same
material that behaves like a diode).
- The
bottom gate sweeps voltage across the device.
- The
extended source acts like an additional control, producing Negative
Differential Transconductance (NDT) — a region where current decreases
as voltage increases. This creates a distinct "middle"
current/voltage state.
Result: The transistor reliably switches between three
distinct current levels by adjusting gate voltage:
- Low
current → logic 0
- Medium
current → logic 1
- High
current → logic 2
These current states translate into voltage levels for logic
circuits.
How It Powers AI Tasks
The chip demonstrated image recognition with high
accuracy. Ternary circuits allow:
- More
compact neural network implementations (fewer transistors per operation).
- Efficient
matrix multiplications and activations common in AI.
- Overall
lower power and faster processing compared to equivalent binary silicon
designs.
Ternary logic gates (AND, OR, etc.) are built from these
transistors. For example, a ternary adder can handle more values per step than
a binary one.
Why Carbon Nanotubes Help Here
CNTs provide excellent electrical properties (high mobility,
sharp switching) that make the three states stable and distinguishable. Silicon
can do ternary in theory, but it's harder to get clean, reliable intermediate
states without leakage or variability — CNTs + the SGT design make it
practical.
Limitations (current stage): This is a research
prototype. Scaling to billions of reliable ternary transistors, integrating
with full systems, and competing with mature binary ecosystems are still
challenges. But it shows a promising path beyond binary limits.
In summary, the ternary approach uses clever transistor
physics (source extension creating a controllable junction) on carbon nanotubes
to unlock a middle state, packing more computing power into less energy and
space. This is why it's exciting for future AI hardware!
Comments
Post a Comment