• Home  
  • US Military AI Deals with 7 Tech Companies
- Artificial Intelligence

US Military AI Deals with 7 Tech Companies

US Department of Defense partners with Google, Microsoft, and others to integrate AI into classified systems

US Military AI Deals with 7 Tech Companies

AI Models Are Getting More Accurate — But Also Much Heavier

Artificial intelligence models are achieving record-breaking accuracy across vision, language, and multimodal tasks. Systems like OpenAI’s GPT-4, Google’s Gemini, and Meta’s Llama 3 now handle complex reasoning, code generation, and real-time dialogue with striking fluency. These improvements come at a steep cost: model size and computational demand have exploded. What once required a single GPU now needs clusters of high-memory accelerators. Training runs consume millions of dollars in compute and weeks of uninterrupted processing. Even inference — running the model after training — has become prohibitively expensive for smaller organizations.

Why Model Size Isn’t Just a Technical Detail

Model size directly impacts accessibility, deployment, and environmental sustainability. Large language models (LLMs) like GPT-4 are estimated to have over a trillion parameters. Meta’s Llama 3-70B, publicly released in mid-2024, requires at least 140GB of GPU memory just to run a single inference pass. That rules out deployment on consumer hardware or edge devices like smartphones and IoT systems. Only cloud providers with access to Nvidia H100 or AMD MI300X clusters can realistically host these models at scale.

The financial barrier is steep. Running a single LLM inference on a cloud instance can cost between $0.001 and $0.01 per request, depending on context length and model size. For high-traffic applications, those costs compound rapidly. A chatbot serving one million users per day could incur daily inference costs exceeding $10,000. Startups and academic labs struggle to compete with tech giants who can absorb these expenses. Google, Amazon, and Microsoft have integrated their AI models directly into cloud offerings — AWS Bedrock, Google Vertex AI, Azure OpenAI — turning model scale into a profit engine.

Energy use is another consequence. A 2023 study from the University of Massachusetts, Amherst estimated that training a single large transformer model can emit over 280 metric tons of CO₂ — equivalent to five average cars over their lifetimes. As models grow, so do their carbon footprints. Some organizations, including Hugging Face and EleutherAI, have begun publishing energy consumption metrics alongside model releases, pushing for transparency. But without regulatory pressure or industry-wide standards, such disclosures remain voluntary.

Compression and Efficiency: The Quiet Innovation Race

While headline-grabbing models grow ever larger, a parallel effort focuses on making AI smaller, faster, and more efficient. Techniques like quantization, pruning, and distillation are gaining traction. Quantization reduces the precision of model weights — from 32-bit floating point to 8-bit integers — cutting memory use by up to 75% with minimal accuracy loss. Meta has applied quantization to Llama 3, enabling versions that run on Apple’s M2 chip. Google uses a variant called float8 in its TPU v5 hardware, achieving faster inference across its data centers.

Model distillation trains smaller “student” models to mimic larger “teacher” models. Microsoft’s Phi-3 series, released in 2024, includes a 3.8-billion-parameter model trained via distillation on synthetic data. Despite its size, Phi-3 matches the performance of models ten times larger on benchmarks like MMLU and GSM8K. The model runs efficiently on mobile devices and supports on-device AI features without relying on cloud connectivity.

Startups are also entering the efficiency space. Californian firm Neural Magic develops sparse models that activate only a fraction of their parameters per inference. Their DeepSparse engine runs dense LLMs like Llama 2 on commodity CPUs, reducing reliance on GPUs. In early 2024, they reported a 4x speedup over baseline CPU inference for a 7B-parameter model. Meanwhile, MIT’s SparseGPT method allows pruning up to 50% of weights in models like Llama and OPT without fine-tuning. These tools aren’t just academic — they’re being adopted by enterprises looking to reduce cloud spend and latency.

The Bigger Picture: AI’s Sustainability and Equity Challenge

The trend toward larger models isn’t just a technical trajectory — it’s reshaping who gets to participate in AI innovation. When training a model costs $100 million and requires exclusive access to thousands of GPUs, only a handful of corporations can lead. OpenAI, Google, Meta, and Anthropic dominate the frontier model landscape. Even national research initiatives, like France’s Maison de l’Intelligence or Germany’s GAIA-X, struggle to match their scale. This concentration risks locking in biases, limiting model diversity, and reducing accountability.

Smaller models offer a path to democratization. Africa-focused startup Lelapa AI trains compact language models for African languages using transfer learning and low-rank adaptation (LoRA). Their Setswana and isiZulu models run on laptops and local servers, serving communities often ignored by global tech firms. Similarly, India’s Sarvam AI released OpenHathi, a family of Hindi and regional language models under 7B parameters, optimized for local deployment. These efforts prove that high performance doesn’t require massive scale — especially when tailored to specific linguistic and cultural contexts.

Regulatory frameworks are starting to respond. The EU AI Act, expected to take full effect by 2026, includes provisions for environmental impact assessments of high-resource AI systems. France’s data protection authority, CNIL, has already required companies to disclose energy use for AI training runs exceeding certain thresholds. In the U.S. the Department of Energy funded a 2024 study on AI’s electricity demand, projecting that data centers could consume up to 6% of national power by 2028 — up from 2% in 2020. These signals suggest that efficiency may soon become a compliance issue, not just an engineering goal.

What Competitors Are Doing Differently

Not all major players are betting on scale. Mistral AI, the French startup founded in 2023, has gained attention by releasing smaller, open-weight models that punch above their weight. Their Mixtral 8x7B model uses a sparse mixture-of-experts (MoE) architecture, activating only two of eight experts per token. This reduces compute needs while maintaining performance close to Meta’s Llama 2 70B. The model runs on a single H100 GPU, making it accessible to more developers. Mistral licenses its models commercially, but allows broad usage rights — a contrast to OpenAI’s restrictive API-only approach.

Alibaba Cloud’s Qwen series takes a hybrid path. Their Qwen-72B model competes with Llama 3 and GPT-4 in benchmarks, but they’ve also released Qwen-1.8B and Qwen-7B for edge deployment. Alibaba integrates these models into its cloud ecosystem, offering tiered pricing based on size and latency requirements. In China, where data sovereignty laws limit foreign cloud access, this local availability is a strategic advantage.

Apple has taken a different stance altogether. Rather than chasing frontier LLMs, they’ve focused on on-device intelligence. Their 2024 iOS 18 update features a new generative model for Siri and text editing that runs entirely on iPhone hardware. Details are scarce, but analysts believe it’s under 10B parameters and uses aggressive quantization. By keeping data local, Apple avoids cloud costs and strengthens privacy — a clear differentiator in markets wary of data harvesting. They’ve also acquired startups like DarwinAI, which specializes in neural network optimization, signaling long-term investment in efficient AI.

What Comes Next: Efficiency as a Core Metric

The AI industry is beginning to treat efficiency as a first-class metric, alongside accuracy and speed. MLPerf, the benchmarking consortium backed by Google, Nvidia, and Stanford, now includes inference efficiency tests across mobile, data center, and edge devices. In the June 2024 round, entries from Qualcomm, Intel, and Tenstorrent showed dramatic improvements in tokens-per-second per watt. These benchmarks are shaping procurement decisions — automakers, for instance, use MLPerf Tiny to evaluate AI chips for autonomous driving systems.

Hardware is adapting too. Groq, a startup founded by former Google TPU engineers, launched a Language Processing Unit (LPU) in early 2024 that delivers 500 tokens per second for Llama 2 70B — faster than most GPU setups. Their architecture eliminates traditional bottlenecks by using a static scheduling model, reducing memory latency. While currently limited to a narrow range of models, Groq’s performance highlights that specialized hardware can outperform general-purpose GPUs when efficiency is the goal.

Looking ahead, the balance between scale and efficiency will define the next phase of AI. Environmental limits, economic pressures, and regulatory scrutiny are aligning to favor leaner, smarter models. The winners may not be those with the biggest budgets, but those who can deliver strong performance with minimal resource use. That shift won’t make headlines like a new GPT release — but it could reshape the industry more profoundly.

About AI Post Daily

Independent coverage of artificial intelligence, machine learning, cybersecurity, and the technology shaping our future.

Contact: Get in touch

We use cookies to personalize content and ads, and to analyze traffic. By using this site, you agree to our Privacy Policy.