Skip to main content
AI-архитектураАвтоматизацияРазработка ПО

AI Writing Machine Code Directly: Business Wins vs. Loss of Control

A radical shift is discussed where AI generates bytecode or machine code directly, skipping Python or Java to save tokens. For business, this increases execution speed but creates critical risks regarding quality control and security. Faster development does not mean verifiable or manageable code, requiring new auditing architectures.

Technical Context

The discussion includes an idea, popularized by Elon Musk among others, that "coding will die" because AI will eventually bypass human-readable languages and generate machine code directly. In its extended version, this argument claims the model "thinks" in a latent space and isn't required to "output tokens" (text), meaning it could produce bytecode, object files, or CPU instructions directly—a "metalanguage" optimal for the machine.

It is crucial to separate engineering reality from futurology. As of February 2026, there are no robustly confirmed studies or generally accepted architectures in open sources where the direct generation of correct machine code is solved for mass industrial use (with debugging, portability, security, and provability). However, there is clear engineering logic behind why this is discussed: current LLMs work via text tokenization, and code is also text, which adds overhead. Theoretically, the "human syntax" layer could be reduced.

What Can Actually Be "Generated Directly"

  • Bytecode: e.g., JVM bytecode, .NET IL, WebAssembly. This is a more structured "low level" than source code but not yet specific CPU instructions.
  • Intermediate Representation (IR): Similar to LLVM IR. In practice, this is often the best compromise between optimization and portability.
  • Machine Code: Specific instructions for an architecture (x86-64, ARM64), considering ABI, linking, memory, exceptions, and system calls.
  • Graph/Vector Program Representation: Not standardized, but the idea of "code as an object" rather than text actively surfaces in architectural discussions.

Why "Fewer Tokens" Sounds Logical

  • Context Savings: Source code is verbose; bytecode is compact.
  • Less "Noise": Whitespace, names, formatting, comments—useful for humans, but not for machines.
  • Structural Ambiguity: Many syntax-level errors disappear if the model immediately outputs a correct structure.

However, these benefits are not free. Syntax is not the only source of complexity. The main complexity in development is semantics: correctness, safety, edge cases, integration, testability, and maintenance.

Key Technical Limitations Business Will Feel First

  • Verification: How do you prove the generated bytecode does exactly what is needed, not just "almost"? At the machine code level, explainability and diff-reviews practically vanish.
  • Determinism and Reproducibility: It is vital to stably rebuild artifacts, compare versions, and rollback. For "one-step generation," build chains and controls must be reinvented.
  • Debugging: Without source code, debugging turns into work with disassemblers/IR and symbols; this is more expensive and requires different competencies.
  • Supply Chain Security: The lower the artifact level, the harder it is to spot backdoors and unexpected side effects.
  • Portability: Machine code is tied to the platform; bytecode/IR is a compromise but still requires strict environment contracts.

Business & Automation Impact

If we imagine direct bytecode/machine code generation becomes widely available, the market will change not because "programmers are no longer needed," but because responsibility contours and development architecture will shift. For business, this means speed may increase, but the cost of errors and the cost of control will also rise.

Where Business Can Really Win

  • Generating "Micro-Components": Small functions, data transformations, serialization/deserialization, filters, rules—where behavior is easily specified by tests.
  • Optimizing Hot Spots: If a model can spawn optimized low-level code for target hardware, this is interesting for high-load, edge computing, embedded systems, and fintech calculations.
  • WASM as a Universal "Logic Container": For products with plugin architectures (extension marketplaces, integration platforms), aiming for a sandbox is safer than a native binary.
  • Reducing "Intermediate Routine" Costs: Tasks like "rewrite from language to language," "build SDK," or "align interfaces" may shift to artifact generation modes.

Who Is at Risk

  • Regulated Industries (finance, medicine, industry): Audit, traceability, and explainability requirements conflict with a "black box" at the binary level.
  • Companies with Long Software Lifecycles: Maintenance for 5–15 years requires readable artifacts, documentation, and stable standards.
  • Organizations without Mature QA/SecOps: Direct low-level output will strictly multiply the probability of vulnerabilities and incidents.

How Development Architecture Will Change

If "code as text" gives way to "code as artifact," the main asset becomes not the IDE or the language, but contracts and tests. Consequently, companies will compete not on the ability to "write code," but on the ability to formalize requirements and verify results.

In practice, this leads to three shifts:

  • Specification > Implementation: The value of formal specifications, property-based testing, generative tests, and API contract testing grows.
  • Verification in CI/CD Becomes Core: Static analysis, SAST/DAST, fuzzing, sandbox execution, and version behavior comparison.
  • Policies and Guardrails: What exactly the model is allowed to generate, where and how it runs, who signs the artifact, and what quality metrics are mandatory.

This is where AI implementation initiatives often break: companies buy a "code generator" but do not rebuild the control contour. As a result, AI is either banned after several incidents or remains a toy for prototypes. At Nahornyi AI Lab, we regularly see this pattern: the technology is strong, but without proper AI architecture and engineering regulations, it doesn't reach production.

Expert Opinion: Vadym Nahornyi

The biggest mistake is thinking that the "programming language" is the main bottleneck. The bottleneck is verifiability and responsibility. If we remove the human-readable layer, we gain seconds and tokens, but we risk losing auditability, debugging, and security.

In our practice at Nahornyi AI Lab, when a client asks to "create AI development automation," it almost always turns out they need not to "generate more code," but to reduce defects, accelerate releases, and maintain quality. This is achieved through process architecture:

  • Unified Requirements Contour: User stories → test scenarios → API contracts → automated checks.
  • Zone Separation: Where AI is allowed to generate (low risk) versus where it only assists (high risk).
  • Sandboxes: Execution with restrictions (e.g., WASM/containers) so that even a faulty artifact doesn't become an incident.
  • Observability: Code quality and behavior metrics (latency, errors, drift) so "fast" changes don't break business processes.

My forecast: Utilitarian value will appear sooner in bytecode/IR (WASM, JVM, LLVM IR) than in "raw" machine code. It is easier to standardize contracts, ensure portability, and build defensive perimeters there. The idea that "a model writes x86-64 directly and replaces development" is more likely a hype formulation. Reality will be hybrid: humans define specifications and constraints, AI generates the artifact, and the industrial pipeline proves, measures, and admits it to production.

And one more practical point: even if the model "thinks in latent space," business still lives in a world of documents, contracts, regulations, and liability. Therefore, the key asset is not generation magic, but the integration of Artificial Intelligence into the development process such that the result can be defended in an audit and explained to leadership.

Theory inspires, but results require practice. If you want to evaluate where AI will truly accelerate development and operations, and where it will create new risks, discuss your case with Nahornyi AI Lab. I, Vadym Nahornyi, am responsible for architectural quality and driving AI initiatives to measurable business effects.

Share this article