Antreas Antoniou, AI Researcher and Engineer

psst... click me

Antreas Antoniou

AI Researcher & Engineer

Cofounder at Axiotic AI

About

Who I Am

I'm an AI researcher and engineer with a PhD in Machine Learning and Meta-Learning from the University of Edinburgh. I've worked at Google, Amazon, and as principal scientist across AI startups in the UK and US. I specialize in large language models, multimodal learning, and self-supervised methods.

I believe intelligence emerges from structure, interaction, and information—not from sheer scale alone. The field has converged on one dominant recipe: make it bigger. But that's not how natural intelligence works. Brains compress experience, maintain state, and build hierarchical representations. They don't re-ingest the universe every time they need to think.

My work focuses on the first principles of learning—how structure, interaction, and information theory can replace massive parameter counts. I'm driven by that Cambrian Explosion spirit of the 2010s: new architectures, new paradigms, genuine exploration of the research tree.

What drives me:

  • Smarter learning, not bigger models — efficiency through insight, not brute force
  • First principles — understanding why things work, not just that they work
  • Open science — releasing code, models, and datasets to push the field forward
  • Building things — ideas are cheap, working systems are what matter

Current Work

What I'm Building

🔬

Axiotic AI

Cofounder & Research Lead

An open-science frontier AI company. We're exploring what's beyond the current scaling paradigm—better learning signals, novel architectures, and systems that actually understand.

Axiotic is also where I turn these ideas into deployed systems: LLM-assisted expert consensus platforms, research tooling, creative-companion workflows, and practical AI infrastructure for people who need more than a chatbot.

🎯

Advisory & Technical DD

Open to a small number of advisory and technical due-diligence engagements for VCs and AI startups: eval design, data moats, model strategy, efficiency claims, and research-roadmap pressure-testing.

iam@antreas.io →

Philosophy

How I Think About Research

Overarching Theme

My goal is to emulate the trajectory of human-like representation learning—starting from foundational representations akin to those seen in human infants, but without requiring eons of evolutionary fine-tuning. This serves as a springboard for my broader ambition: investigating how these infant-like representations can be fine-tuned in concert with higher-level abstract concepts to pave the way for general artificial intelligence. To achieve this, my research focuses on scalable, data-efficient, and generalizable self-supervised learning in a multimodal setting. I integrate insights from neuroscience and evolutionary computation to explore optimal learning sequences and curricula, paying close attention to architectural choices and their corresponding training recipes.

Research Focus

Leading my research interests is Multi-Modal Learning—the synergistic integration of text, images, audio, and video. This is followed by Self-Supervised Methods inspired by infant learning and evolutionary computation.

Multi-Modal Learning Self-Supervised Methods Meta-Learning Adversarial Learning Evolutionary Optimization Computational Efficiency Memory-Augmented Networks

Research Philosophy

I operate within a pragmatic framework, aiming to identify high-leverage focal points conducive to in-depth investigation. This allows for efficient allocation of both computational and cognitive resources. In line with evolutionary tenets and the Pareto Principle, my methodology focuses on the "fittest" 20% of research avenues likely to contribute 80% of impactful results.

Research Agenda

What I Think About

I'm fundamentally inspired by information theory, evolution, and biology—how can we use what we observe about human learning and nature to build better artificial intelligence?

Improved Tasks

Beyond next-token prediction. Predicting masked past, present, and future states to force causal structure, long-horizon planning, and robust memory.

Multimodal Learning

Inspired by the brain's multimodal neurons—"synesthetic" architectures processing text, images, and audio in a shared latent space.

Memory & Recursion

Explicit memory and latent reasoning loops turning transformers into differentiable computers—infinite-context learning without expensive CoT tokens.

Hierarchical Attention

Fractal attention patterns operating at multiple resolutions simultaneously—seeing the forest and the trees without brute-force scaling.

Novel Architectures

New connectivity patterns, activation mechanisms, and dense routing schemes where deep layers attend directly to shallow features.

AI Flywheels

Data quality is the new frontier. Closed-loop engines where models generate, filter, and curriculum-sort their own training data.

Embodied Learning

Intelligence requires grounding. Using simulation as a primary data source, training agents on interaction physics before sim-to-real transfer.

Evolution meets SGD

SGD is greedy; evolution generalizes but is slow. Researching how to synergize them for robust, efficient learning.

Information & Distillation

Distillation through an information-theoretic lens—cracking the physics of compression to train small models to their maximum potential.

Local Learning Rules

Efficiency in nature is achieved through local learning rules. Leveraging this for efficient learning without global backprop.

Asynchronous Dynamics

Rhythms across multiple temporal scales define natural systems. Exploring how neural architectures can internalize these for efficient learning.

Expert Consensus Systems

LLMs as coordination and compression engines: mapping agreement, disagreement, uncertainty, and evidence across expert groups without replacing judgement.

Bio-Inspired Learning

Nature solved learning over billions of years. Drawing from biological neural networks, synaptic plasticity, and adaptive mechanisms.

Publications

Selected Research

A curated view: the three latest papers plus three highly cited foundations. See all on Google Scholar →

Teaching

Teaching, Lectures & Workshops

A single home for lecture materials, invited talks, workshops, tutorials, and teaching recognition.

Teaching Materials

Lecture pack 2026

GenAI as Creative Companion

A teaching package on using generative AI as a collaborator for writing, music, visual storytelling, and agentic creative workflows.

Workshop 2026

AI-Assisted Development Crash Course

A standalone teaching session on using Cursor, Claude, Codex, and agentic workflows for practical software development.

🎤

Guest Lectures & Invited Talks

University of Edinburgh and invited sessions · 2025–2026

Quest Lecture: LLM 101 and Model Compression

Machine Learning Systems Course · 2026

Quest Lecture: LLM 101 and Model Compression

Machine Learning Systems Course · 2025

GenAI Superpowers

Teach-A-Thon presentation

Slides

LLMs for Teaching

Workshop on using LLMs in education

Slides

3000 Hours with ChatGPT

What I learned from extensive LLM usage

Slides

AI-Assisted Development: A Crash Course

Practical guide to coding with AI assistants

Slides

EIDF Cluster, Kubernetes & Docker Primer

Introduction to UoE's compute infrastructure

Slides

Machine Learning Practical (2017-2019)

Lead TA at University of Edinburgh. Created comprehensive tutorials, coursework materials, and supervised research projects. Course Website →

Google Cloud Platform

Teaching Awards & Recognition

Staff Award for MLP TA Service

2020-21

University of Edinburgh — Recognition for exceptional teaching assistance

View Award →

Teaching Award Nominations

View Letters →
Best Practice in Inclusive Learning Award 2019
Best Support Staff Award 2019
Best Student Who Tutors Award (2 nominations) 2019
Best UK PhD Tutor Award 2019
Best Student Who Tutors Award 2018

Background

Experience

Cofounder & Research Lead

2025 - Present

Axiotic AI

Building an open-science frontier AI company. Exploring efficient learning, novel architectures, and what's beyond scaling.

Principal AI Scientist & Head of ML

2025 - 2026

Pieces for Developers

Led ML research agenda. Built nano-models and foundation models for efficient deployment across CPUs, GPUs, NPUs, and LPUs.

Lead Research Scientist

2024

Malted AI

Efficient LLM training, distillation, synthetic data generation, and LLM-as-a-Judge systems.

Machine Learning Scientist

2021 - 2022

Amazon

Applied machine learning research and development.

Research Intern

2020

Google

Machine learning research internship during PhD studies.

Research Associate

2019 - 2024

University of Edinburgh

Supervised by Prof. Amos Storkey. Part of BayesWatch and the Adaptive and Neural Computation (ANC) institute.

PhD in Machine Learning

2016 - 2020

University of Edinburgh

Thesis: "Meta Learning for Supervised and Unsupervised Few-Shot Learning"

Community

Outreach & Open Source

Building communities and democratizing ML research

🤖

LeRobot Edinburgh Hackathon

7th Worldwide

Lead Organizer & Competitor · 2025

Originally conceived and initiated bringing the global LeRobot Hackathon to the University of Edinburgh. Built the organizing team and managed 60% of organizational work. Sourced and provided all specialized hardware and high-performance GPU infrastructure. Served as expert mentor for all teams throughout the 30-hour competition. Also competed and led my team to 7th place out of 1000+ teams worldwide.

🖥️

EIDF A100 GPU Cluster Community

Early Adopter & Community Lead · 2023

One of the early adopters and key community support members for Edinburgh's EIDF A100 GPU cluster. Created the Slack server, answered hundreds of user questions, scheduled community meetings for key issues, and served as a bridge between users and developers. Authored early documentation and developed kubejobs, a Python package simplifying Kubernetes job specifications.

🎯

Workshop & Event Organizing

Co-organizer · 2024-2025

Co-organized the International Workshop on Efficient Generative AI 2024, a GAIL-funded event bringing together leading researchers from academia and industry to discuss efficient approaches to generative AI.

📚

Research Talks & Presentations

To Learn or Not to Learn

A journey across modern neural network inductive biases

Slides

Transferable Representation Learning

Discovering learning priors across domains, tasks, and modalities

Slides

TALI Dataset

Temporally and semantically Aligned Audio, Language and Images

Slides

Better DL Benchmarks & Datasets

...and other mythical creatures

Slides

Continual Few-Shot Learning

NeurIPS Workshop paper presentation

Slides

Learning to Learn via Self-Critique

NeurIPS 2019 paper presentation

Slides

Parting Talk: UoE 2024

Farewell presentation at University of Edinburgh

Slides

How to Cat

Cat ownership, biology, evolution & psychology

Slides
🎙️

Podcasts & Interviews

Nano Models, Transformers, and Long-Term Memory AI

Pieces for Developers Podcast · 2025

Watch
✍️

Writing & Articles

Academic Writing Tips and Resources

A practical collection of academic writing habits and go-to paper-writing resources

Read

Beyond the Cloud: SLMs, Local AI, and Agentic Constellations

Pieces Blog · August 2025 — A vision for local-first AI built on biology-inspired architectures

Read

Too Much of a Good Thing: How Chasing Scale is Stifling AI Innovation

Pieces Blog · July 2025 — The "Great Amnesia" of AI's scaling monoculture

Read
🌐

Open Source Contributions

My conviction in the democratization of ML research stems from the irreplaceable value of individual expertise and the power of collective collaboration. Open source is not merely a development model—it's a fundamental necessity for driving innovation and maintaining ethical standards in the field.

🔧

Infrastructure & Tooling

University of Edinburgh · 2021-2024

  • • Secured $20K in Google Cloud Platform Research Credits and a Google TRC Compute Grant
  • • Procured a £50K deep learning research server through market analysis and vendor negotiations
  • • Built and deployed a Kubernetes cluster for the research group with Python tooling and tutorials
  • • Authored a minimal ML research framework following best practices
  • • Built comprehensive wiki documenting best practices, tools, and resources for the group

Playground

Demos & Datasets

Interactive explorations of my research. Dive in!

🚀 Featured Dataset

TALI: Temporally & Semantically Aligned Audio, Language and Images

TALI is my response to the growing demand for multimodal understanding in deep learning. It's a large-scale, tetramodal dataset that aligns text, video, images, and audio—a playground for innovative self-supervised learning tasks and multimodal research. With TALI, we're exploring how different modalities and data/model scaling affect downstream performance. I'm excited about the diverse research ideas it will inspire!

🧭 Applied AI Platform

Symphonia: LLM-Assisted Expert Consensus

Symphonia is a platform for synthesising structured expert judgement. Experts submit positions, reasoning, evidence, and confidence; an LLM drafts a transparent map of agreement, disagreement, and uncertainty; participants then refine it through iterative feedback cycles. The point is not to automate judgement — it is to make collective judgement cheaper, clearer, and more scalable.

This is the same research story in applied form: LLMs are most powerful when used as coordination and compression engines over human expertise, preserving disagreement rather than flattening it.

More demos coming soon...

Contact

Let's Talk

Interested in research collaboration, consulting, or just want to chat about AI?