Skip to content

Insights, Breakthroughs, and Stories from Sony AI

Explore how our teams are pushing the boundaries of artificial intelligence.

Close-up of a person wearing glasses with a blue digital screen light reflecting on the lenses
female-professional-ping-pong-athlete
A hand tightly grips a gaming controller, surrounded by vibrant neon hues and swirling lines
A robotic arm delicately holds a microchip, showcasing precision technology
A scientist in a lab coat and safety goggles operates advanced lab equipment under blue lighting, focused and meticulous in a modern laboratory setting

Explore Our Global Impact

Every day, our globally recognized researchers explore the most complex questions in AI. Our mission—to unleash human imagination and creativity—shapes a distinctive perspective, one that opens new creative pathways while honoring past innovation and expanding the boundaries of what is possible.

Creating an Ecosystem of Fairness and Protection with AI

Today’s narratives would have you believe that AI’s only role is in creating content that takes advantage of and replaces human creativity. Current opt-out systems place the burden on creators to proactively restrict the use of their work – a near-impossible task given how widely content is distributed online. Additionally, those opt-out signals are frequently ignored or circumvented by AI developers. 

Our researchers are working to develop AI technologies and new breakthroughs in AI science that use the unique power of artificial intelligence to empower artists and rights holders with tools that protect past and future creative pursuits. 

Sony AI’s researchers are working to develop the blueprints for AI technologies that can use the unique power of artificial intelligence to help artists and rights holders understand when and how their work appears in generated music, and enable the creation of tools that support attribution and protection at scale. Recent work opens a discussion about  'nuanced opt-in', which flips the default of ‘opt-out’ systems: no work can be used unless a creator explicitly permits it. and the need for Attribution by design, rethinking the way artists are compensated in a world of AI generated content. 

A person with braided hair sits on a couch, playing an acoustic guitar with a joyful expression

Research Highlights: AI Ethics

At Sony AI, our AI Ethics team has spent years orienting our research to challenge the status quo of dataset construction.

Our work has shown how existing human-centric computer vision datasets often rely on non-consensual web scraping, lack critical demographic metadata, and perpetuate representational gaps—leading to models that can be both unfair and unreliable (Andrews et al., Ethical Considerations for Responsible Data Curation, NeurIPS 2023).

We’ve also highlighted a deeper paradox: protecting privacy by limiting data collection can sometimes leave marginalized groups “unseen,” yet this very absence increases the risk of being “mis-seen” by AI systems—misclassified, misrecognized, or misrepresented (Xiang, Being Seen Versus Mis-Seen, Harvard JOLT 2022). The harms of invisibility and mis-visibility are inseparable, and solving them requires a shift in how datasets are built.

And we have argued that fairness cannot be treated as an afterthought. From our ethical audits to our analysis of how demographic data shapes bias detection, our research has made clear that datasets must be designed with purpose, consent, and diversity from the outset (Xiang et al., Mirror, Mirror: Reflections on Dataset Bias and Fairness, 2021).

Together, these works revealed not just the importance of ethical data collection, but also the difficulty of reconciling best practices, conflicting priorities, and technical specifications of fairness. 

This is evidenced in The Fair Human-Centric Image Benchmark (FHIBE). FHIBE’s dataset comprises 10,318 consensually-sourced images of 1,981 unique subjects, each with extensive and precise annotations. These annotations capture demographic and physical attributes, environmental factors, and camera settings – enabling nuanced assessments of fairness and bias across a wide range of demographic attributes and their intersections. The images were collected from subjects in over 81 countries/regions, making it one of the most globally diverse and most comprehensively annotated datasets in existence.

A collage of diverse groups engaging in activities, swimming, dancing, eating, and embracing

FHIBE – Bias and Fairness

Sony AI’s FHIBE provides the first globally diverse, consensual dataset to build fairer, more transparent computer vision technologies.

Related Resources

Blog

Sony AI Ethics Flagship: Reflecting on Our Progress and Purpose

Blog

Introducing FHIBE: A Consent-Driven Benchmark for AI Fairness Evaluation

Protective AI, Music Attribution, IP-Level Tracking

New research papers accepted to NeurIPS, ICML, and INTERSPEECH 2025, are focused on musical integrity in the age of machine learning, exploring attribution, recognition, and protection.

A man and woman smile warmly at each other in a music studio

Research Highlights: AI for Creators

This research is part of a growing body of work exploring how AI can unlearn what doesn’t belong to it, how connections between musical segments can be identified, and how effective current audio authentication methods are in verifying a track’s integrity after typical alterations, like compression or file conversion.

Abstract image of floating, colorful translucent squares against a blurred background

Memorization

To protect licensed data, Sony AI researches ways to prevent verbatim memorization in LLMs. This ensures AI remains a creative partner rather than a duplicator.

Related Resources

Blog

Protecting Creator’s Rights in the Age of AI

Publication

Supervised Contrastive Learning from Weakly-Labeled Audio Segments for Musical Version Matching

Publication

Reductive, Exclusionary, Normalising: The Limits of Generative AI Music

Learning Play, not just Prediction

From sentiment analysis to interactive robotics, AI tackles a range of challenges. While some tasks involve parsing patterns or generating outputs from static data, others require systems to make decisions, adapt in real time, and learn from experience. This is where reinforcement learning thrives: at the core of learning how to act, not just predict.

To research and develop GT Sophy, we collaborated directly with Kazunori Yamauchi, the CEO of the video game studio Polyphony Digital, known for creating the popular racing simulation game Gran Turismo. 

Yamauchi-san and his team were involved from the beginning of the project and through every other stage, helping our team define the challenge we wanted to address, research the utility of our novel deep reinforcement learning approach within Gran Turismo, assess how human players responded and performed against GT Sophy, and release GT Sophy as a feature within Gran Turismo back in late 2023.

What started out as a grand challenge to create an AI agent that could beat the world’s best Gran Turismo drivers quickly evolved into a quest for our team to help game developers deliver new and exciting gaming experiences to players around the world through AI.

A person with long turquoise hair wears a headset, focused intently on steering a simulated racing game

Research Highlights: Gaming & Robotics

These breakthroughs come from our core research in gaming and robotics. Our teams use deep reinforcement learning and sophisticated sensing to move beyond virtual simulation and into the physical world, analyzing agents that compete and collaborate at advanced levels.

GT Sophy: Interactive RL

Sony AI's GT Sophy evolves reinforcement learning from pure competition to sportsmanship. Integrated into Gran Turismo 7, this agent delivers realistic, collaborative, and human-centric racing experiences.

A focused individual works on a robotic hand with intricate metal parts and orange wires

Beyond Simulation: Real-World RL

Our research shifts AI from virtual benchmarks to the physical world. By combining robotics and sensing, Sony AI demonstrates that agents can achieve expert-level performance in unpredictable, adversarial environments.

A collage of images showing vision detection capabilities, people engaging in various outdoor activities-1

Collaborative Human-AI Play

Focusing on "Learning Play," Sony AI develops agents that augment the human experience. These systems prioritize player engagement and refined interaction over simple algorithmic dominance.

Related Resources

Blog

Sony AI’s Deep RL Team on Why the Hardest Problems Still Matter

Blog

How to Train Your Race Car

Amplifying Creativity

By expanding what AI can see, hear, and translate, we are researching  new tools creators can use to explore new pathways to amplify creativity and efficiency. Explore our work in audio, sound creation, and seamless multilingual communication.

In many ways, AI – especially generative AI – remains in its infancy. Researchers, AI developers, and creators are working to understand how this technology will influence the creative process.

Since many questions still remain around real-time efficiency and controllability, AI researchers and developers need to dedicate significant time and resources to understanding this.

From our perspective at Sony AI, we are working to understand how AI influences the work of artists – not only from a technical or process perspective, where it can be a tool they choose to utilize – but also in how it helps them push boundaries to explore new mediums or genres, as well as how it affects their rights and the ways we can protect them.

A person in headphones speaks into a microphone, seated at a desk with a laptop

Research Highlights: Imaging, Sensing, & Music

Much of our research in this area focuses on imaging, sensing, and music. Our teams refine how AI interprets the world and generate high-quality content, providing creators with tools that act as extensions of their craft.

Man wearing headphones and sunglasses, bathed in neon pink light

MMAudio: Video-to-Audio Synthesis

Sony AI’s MMAudio generates synchronized sound from video. Using deep learning, it aligns audio with visual timing, providing creators with a contextually accurate sound-design tool.

Colorful digital soundwave pattern with blue grid background, featuring vibrant pink, purple, and orange waves

Similar Sound Search: Creative Workflow

In collaboration with Audiokinetic, Sony AI launched Similar Sound Search. This tool allows designers to find assets by example or text, streamlining the discovery process.

A man with long hair, wearing a white T-shirt, sits on the floor using his phone in a cozy living room

Controllability in AI for Creators

Sony AI’s Flagship team enhances real-time controllability for professionals. These tools ensure AI acts as an intuitive partner, providing granular control over music and 3D creation.

Sound Engineering & DisMix

Addressing pitch and timbre, Sony AI’s DisMix and VRVQ research advances audio mixing. These innovations offer professionals precise control, pushing the boundaries of digital artistry.

Related Resources

Blog

Sights on AI: Yuki Mitsufuji Shares Inspiration for AI Research into Music and Sound

Blog

New Research at ICCV 2025: Expanding the Boundaries of Vision and Generative AI

Latest Updates from Sony AI

Stay current with our latest news, media coverage, and technical insights. Explore the research and people driving our mission forward.

A woman with glasses and a pink jacket smiles while using a laptop on a dark sofa

Shape the Future of AI with Sony AI