MEI-Trumpet Landscape Cover

MEI-Trumpet is an educational web application that supports users in mastering musical competencies. The app facilitates learning through guided repetition, structured reflection, and multi-sensory engagement, addressing critical technical and auditory skills.

By allowing students to set custom parameters, MEI-Trumpet adapts to different skill levels, creating a personalized learning experience. Starting with a beginner-friendly interface, users can customize their practice by selecting:

The image below shows the interactive controls where students can set these parameters before starting an exercise.

Each exercise is structured in three steps:

On the top of the screen, an image showing the correct trumpet fingering, including adjustments for the third valve slide, is displayed. This helps students easily associate the note shown on the musical staff with the corresponding pitch and valve combination.

App Screenshot 1

MEI-Trumpet is available as a free beta version for both web and mobile devices:

Click HERE to try MEI-Trumpet on your browser now!

Advancing Your Skills!

Once users get the hang of playing the right note, scaffolding helps them take it to the next level!

Intonation and Dynamics

Adding a chromatic tuner and dynamic markings, users are encouraged to fine-tune their pitch and volume control bringing more expression and style to their playing.

Tuner Screenshot

Clear Articulation

The latest feature of the Music Education Interface I am working on introduces articulation assessment. A state-of-the-art digital signal processing algorithm measures the attack transient duration in milliseconds, providing users with real-time feedback on their articulation clarity.

"Attack clarity is finding the purity and the best beginning of sound"

Achieving a clean and controlled attack is a crucial skill for musicians, influencing both technical precision and artistic expression. The articulation analysis tool enables users to experiment with different attack styles (e.g., staccato, détaché, martelé) while receiving real-time feedback on their performance. By analyzing the transient development of their sound, students gain deeper control over articulation, allowing them to make more informed musical choices and refine their expressive capabilities.

Audio Examples!

Compare the difference between an unclear and a clean articulation:

Unclean Attack

Clean Attack

Enhancing Sound Quality with Machine Learning

To further refine the student's technique, a machine learning model was integrated into the learning interface to classify sounds based on their timbral quality. This functionality introduces a new layer of complexity, helping users develop their musical competencies.

Watch the DEMO:

What's next?

Future developments of this project focus on enhancing its capabilities through adaptive learning features. The Music Education Interface aims to integrate generative AI to provide a more personalized and inclusive learning experience. Key advancements include:

The objective is to make music education more engaging, effective, and adaptable to individual needs.

Research Background

MEI-Trumpet is part of an open-source research project led by Alberto Acquilino at the Schulich School of Music of McGill University, in collaboration with the International Neuroinformatics Coordinating Facility (INCF). Ongoing studies aim to further define the pedagogical foundations of the interface and discuss its usability and social validity through groups of participants.

The source code for the app is publicly available here.