into a curious era of artificial intelligence where size is wrongly equated with intelligence. The models get larger and larger to billions of parameters, the data centers become industrial in scale, and progress is measured by the megawatts of power used. However, some of the most ingenious intelligent systems ever created — such as interstellar spacecraft and the human brain — run under extremely tough constraints. They do not rely on their size but on their efficiency.
At the heart of modern data science, there is a division. On one hand, machine learning is in a race for scale. On the other hand and less loudly, a revolution is taking place in the backward direction: these are quantized models, edge inference, TinyML, and architectures that will survive on very limited resources. These are not limitations that lead to degradation of performance. They are the signs of a revolutionary change in the engineering of intelligence.
This piece puts forward a modest yet provoking notion: scarcity should not be viewed simply as a limitation to intelligence but rather as the most significant factor behind its development. Whether it’s Voyager 1, neural compression, or the very future of human civilization, the systems that survive are those that figure out how to get more out of less. Efficiency is not something that hinders progress. It is its ultimate form.
In 1977, humanity launched one of the most enduring autonomous engineering systems in history: Voyager 1.

A tiny ambassador from Earth, Voyager 1, sails through the silent grandeur of the cosmos.(Image generated by the author using AI)
Through the solar system, it has been sailing for almost 50 years, self-correcting its path and sending back scientific data from the space outside our solar system. It managed to perform all these feats with only 69.63 kilobytes of memory and a processor operating about 200,000 times slower than today’s smartphones.
Such limitation was not considered a flaw. It was an approach to the design.
Contrast this with the present moment. In 2026, we celebrate large language models that need gigabytes of memory just to write a limerick. We have taken for granted what can only be described as digital gigantism. Efficiency is almost forgotten; achievement is now measured by parameter counts, GPU clusters, and megawatts consumed.
If the Voyager 1 had been constructed using today’s software culture, it wouldn’t have made it beyond Earth orbit.
That aside, nature stays mercilessly efficient. The human brain — probably the smartest intellect out there — only consumes around 20 watts. The Voyager uses a nuclear source that produces even less power than a hairdryer. However, a significant part of what we refer to as AI at present necessitates energy consumption levels that are comparable to those of heavy industries.
In fact, we are manufacturing dinosaurs in an environment that is progressively favoring mammals.

The Efficiency Trap shows how biological intelligence runs on watts while digital intelligence runs on megawatts and becomes less efficient as it scales.(Image generated by the author using AI)
Currently, advanced language models possess tens or even hundreds of billions of parameters, therefore, only their weights can take up several hundreds of gigabytes just for the storage. For instance, GPT-3 in single-precision would take up around 700 GB. The energy consumption of training and running such systems is equal to that of a city.
This kind of design leads to different types of structural fragility:
Quite often, in real-life situations, these trade-offs are not needed. Smaller, more specialized systems most often can produce the bulk of functional value at a small fraction of the cost. Employing a model with a trillion parameters for a very specific job is becoming more and more like utilizing a supercomputer to run a calculator.
The issue is not the lack of capability. The issue is overkill.
Engineering tends to accumulate when resources are plentiful. However, it becomes very accurate when resources are scarce. Limitation makes systems become deliberate.
One good example is quantization — the process of lowering the numeric precision of model weights.

Evolution isn’t adding more data. It’s learning what to delete.(Image generated by the author using AI)
import numpy as np
np.random.seed(42)
w = np.random.randn(4, 4).astype(np.float32)
qmin, qmax = -128, 127
xmin, xmax = w.min(), w.max()
scale = (xmax - xmin) / (qmax - qmin)
zp = qmin - round(xmin / scale)
q = np.clip(np.round(w / scale + zp), qmin, qmax).astype(np.int8)
w_rec = (q.astype(np.float32) - zp) * scale
print("original:", w[0, 0])
print("int8:", q[0, 0])
print("reconstructed:", w_rec[0, 0])
print("error:", abs(w[0, 0] - w_rec[0, 0]))
The decrease of 75% in memory footprint by itself is not simply an achievement of efficiency; it is an essential change in the nature of the model. After removing the decimal noise, the inference speed goes up since the hardware works with integer arithmetic more efficiently than with floating-point operations. Industry studies have always shown that dropping precision from 32-bit to 8-bit and even to 4-bit leads to almost no accuracy loss. Hence, it is clear that a “smart” solution being limited is not turning into a low-level one; it is a concentration. The remaining signal is stronger, more capable of being moved, and finally more developed.
Imagine changing your location to the streets of Kolkata or the farmlands of West Bengal. The “Cloud-First” vision of Silicon Valley usually clashes with the reality of limited 4G and expensive data in much of the Global South. In these places, AI only becomes “useful” when it is local.
Out of such situations, TinyML and Edge AI came to be—not as small copies of “real” AI, but as special designs that can run on cheap hardware without a network connection.

Mobile technology and AI are bringing advanced crop disease detection directly to farmers in the field.(Image generated by the author using AI)
Just take the example of crop disease detection deployment with the PlantVillage dataset. A huge Vision Transformer (ViT) can reach 99% accuracy on a server in Virginia, but it is of no use to a farmer in a remote village without a signal. By using Knowledge Distillation, which is basically the large “Teacher” model training a small “Student” model like MobileNetV3, we could perform real-time leaf-rust detection on a $100 Android device.
In practice:
TinyML-style edge inference example
To deploy these “Student” models, we utilize frameworks such as TensorFlow Lite to transform models into a flatbuffer format which is optimized for mobile CPUs.
import tensorflow as tf
import numpy as np
interpreter = tf.lite.Interpreter(model_path="model.tflite")
interpreter.allocate_tensors()
input_details = interpreter.get_input_details()
output_details = interpreter.get_output_details()
data = np.array([[0.5, 0.2, 0.1]], dtype=np.float32)
interpreter.set_tensor(input_details[0]['index'], data)
interpreter.invoke()
output = interpreter.get_tensor(output_details[0]['index'])
print("Local inference:", output)
These are not compromises, but rather evolutionary advantages. A device of $50 can now perform work that previously required server farms. These systems do not pursue benchmark scores but rather concentrate on living. In terms of evolution, survival picks efficiency, and efficiency results in resilience.
It is only natural that the intelligence going in the direction of efficiency on Earth might also be a principle that applies to the universe at a large scale.
The Fermi Paradox poses the question of why the universe seems to be devoid of any signs of life even though statistically, there should be advanced civilizations out there. We have the belief that intelligence has to grow in the outward direction – Dyson spheres, megastructures, and interstellar broadcasting are some examples of how that may be done.

When intelligence matures, it stops screaming and starts optimizing.(Image generated by the author using AI)
However, what if the mature ones are capable not of expanding but of stabilizing?
A civilization that manages to conduct its computations with minimum waste production to the point of near-zero would leave hardly any trace that we could detect. It would be limiting the communication to a minimum possible level. As its intelligence expanded, its footprint would become smaller.
Under this case, silence is not being empty of life. It is being highly efficient.
As we move from Voyager 1 to the human brain and even to imagine superintelligences, the same pattern keeps on repeating: efficiency comes first, then sophistication.
If our most advanced machines can only do extremely narrow tasks and still need a whole city’s worth of energy, the problem is not that we are too ambitious, it is that our architecture is flawed. AI’s future won’t be a story of size but of grace in limitation.
It won’t be those systems that are the biggest that will survive, but the ones which are the most efficient.
Rather than by how much an entity consumes, intelligence is measured by how little it needs.
From Voyager 1 to the human brain to modern edge AI, one and the same idea keeps repeating: intelligence is not measured by how much it consumes, but by how effectively it works. Dearth is not a villain for innovation — it is the very engine that shapes it. If only a handful of resources are available, then living organisms become very intentional, precise, and resilient.
Quantization, TinyML, and on-device inference are no longer considered temporary solutions that engineering teams can use to patch things up; rather, they are the first signs of a major evolutionary path of computing.
AI’s future will not be determined by which model is the largest or which infrastructure is the loudest. It will be decided by the designs that provide significant functionality with little wasted resources. Genuine brainpower is born when energy, memory, and bandwidth are valued as scarce resources rather than treated as endless supplies. In that light, being efficient is no less than maturity.
The ones that will be here to tell the tale will not be those that merely scale continuously, but those that keep perfecting themselves to the level where nothing that is extra is left. Intelligence, at its finest, is beauty constrained by limitations.
Let’s optimize together
If you are working on making AI more sustainable, efficient, or accessible at the edge, I’d love to connect. You can find more of my work and reach out to me on LinkedIn.