6 September 2022: Symbolic Sound announces the Pacamara Ristretto Audio Processing Unit, representing a giant leap forward in performance, energy efficiency, and connectivity over previous generation APUs.
PRESS RELEASE
September 6, 2022
Symbolic Sound Unveils “Pacamara Ristretto” Audio Processing Unit (APU)
2.5 times performance boost, 20% lower price, expanded connectivity, and inter-application audio
CHAMPAIGN, ILLINOIS — Symbolic Sound announces a new Audio Processing Unit (APU) representing a giant leap forward in performance, energy efficiency, and connectivity over previous generation APUs. At less than half the size, the new Pacamara “Ristretto” delivers, on average, a 2.5 times jump in performance, a 20% reduction in cost, cooler operating temperature, expanded connectivity options, and seamless inter-application routing of audio and MIDI to/from applications running on a host computer.
Performance
Pacamara Ristretto’s compute-cycles and memory are solely dedicated to sound computation, delivering unprecedented levels of performance capabilities to sound designers for film and games, live electronic music performers, sound artists, educators, researchers, and audio professionals who require the highest-quality real-time audio, responsive live interaction, and unbounded creative flexibility from their tools.
“The Pacamara Ristretto represents a radical transformation, not just because it’s Symbolic Sound’s most powerful external APU yet, but because of the ease with which it communicates with the rest of the world,” said Symbolic Sound’s president Carla Scaletti. “Audio, MIDI, and OSC streams can be freely routed between Kyma and other audio applications on your computer and audio devices in your studio, including across LANs and out onto the Internet, creating a combinatorial explosion of sound creation possibilities.”
According to game/film sound design consultant and Pacamara field tester, Alan M. Jackson, “It’s the latest, most powerful revision of the most important audio DSP environment of the last 30 years.”
Powered by Kyma
Audio professionals can leverage the full Kyma 7 accelerated sound design environment for rapid prototyping, testing, creative exploration, and refinement of production-quality sound synthesis and processing algorithms to run in real time on the Pacamara Ristretto.
Kyma, described as a “recombinant sound design environment,” gives sound designers access to extensive libraries of unique audio synthesis and processing modules, along with tools for creating limitless combinations of those algorithms. Galleries of automatically-generated signal flow graphs help speed development, spark the imagination and boost the creative flow — providing creators with the advanced models needed to solve sound design challenges for games, film, live music, data sonification, voice processing, multi-sensor controllers, and more.
Deep integration between hardware and software has always been central to the Symbolic Sound approach. Hardware design is continually informed by software development and vice versa. As a result, the Kyma 7 sound development environment is optimally tuned to take full advantage of the huge increases in processing and memory bandwidth made available by the new Pacamara APU.
Sound designer and musician Cristian Vogel describes the Kyma software environment as “the best for innovating, prototyping and designing innovative digital sound processes… frictionless for ideation and research. Then, reliable and consistent at the outcome stage.”
Say Hello to My Little Friend
The Pacamara Ristretto is as comprehensive in its connectivity as it is diminutive in size. The back panel of the rugged, roadworthy enclosure includes:
• USB-C port for audio and MIDI transfer to/from host computer
• Two USB 2.0 ports for connecting audio interfaces and MIDI controllers
• 3.5 mm jack sockets for mic input and stereo headphone output
• RJ-45 network port to connect via a gigabit Ethernet cable to a host computer
• Wi-Fi antenna jack
• USB-C power input port
An APU does for Real-time Audio what a GPU does for Real-time Graphics
Audio professionals require high quality, low latency, reliable real-time audio synthesis and processing. That’s what an APU can deliver.
“The question is not whether a CPU is capable of real-time graphics generation and processing. Everyone understands the advantages of offloading real-time graphics computation onto one or more GPU(s) running in parallel with your main CPU. Not only does it free up memory and computing resources on your CPU, a GPU guarantees the highest-quality, glitch-free interactive real-time graphic performance. The same is true for audio!”
“With the Pacamara you have high quality, configurable algorithms running on dedicated, uncontended hardware,” explains Jackson, “The result is that it sounds great and it’s reliable.”
Soft Hardware
Unlike traditional audio hardware, which has a fixed functionality that never changes, the Pacamara Ristretto can adapt to the evolving needs of users. Regular, significant software updates ensure that Symbolic Sound APUs continue to gain features and functionality so that they actually improve with time.
Jackson adds, “there’s a subtle and important difference between hardware that is upgradeable… and hardware that DOES get upgraded continuously and significantly. Kyma is [Symbolic Sound’s] only product and [they’ve] been working on it, passionately, for decades. [They] have a proven history of constant and significant upgrades.”
What’s in a Name?
The name, Pacamara, comes from a hybrid coffee plant bred to produce larger beans on smaller trees planted closer together — evoking both the ‘recombinant sound’ design philosophy of Kyma and the smaller size, higher-yield of the Pacamara APUs. According to Scaletti, “Coffee aficionados know that a ristretto shot delivers twice the caffeine per milliliter as a standard shot of espresso. The Pacamara Ristretto is actually 2.5 times more powerful and less than half (44%) the volume of previous generations, meaning that the computational performance by volume has increased 5.7-fold.”
Summary
The Pacamara Ristretto is an APU, designed from the ground up for real-time sound synthesis and processing; it works in parallel with your main computer to deliver high-quality real-time audio signal computation, freeing up resources on the CPU for managing your overall studio workflow.
Audio professionals can leverage the full Kyma 7 accelerated sound design environment for rapid prototyping, testing, creative exploration, and refinement of production-quality sound synthesis and processing algorithms to run in real time on the Pacamara Ristretto.
The Pacamara Ristretto APU is the most recent development in a consistent flow of innovation from Symbolic Sound — from Capybara, to Capybara 33, Capybara 66, Capybara 320, Paca(rana) — the sixth generation Pacamara Ristretto APU is groundbreaking in terms of raw compute power, delivering a 2.5-fold leap forward in performance, connectivity, and dependability backed by extensive field testing in real world performance environments.
Pricing and Availability
The Pacamara Ristretto is available today at https://kyma.symbolicsound.com/order-now Initial supplies are limited due to ongoing (and inexplicable) global electronic component shortages.
The price for a 4-processor Pacamara Ristretto is $3,818 (US)
An entry level, 2-processor system is available for $3,302 (US)
The Kyma 7 software for macOS or Windows is included with the APU.
From its origins in 1990 when the first generation of APUs were assembled at a kitchen table in a student apartment, Symbolic Sound’s current software and hardware offerings represent over 30 years of continuous development, creation, refinement, and close interaction with customers, inspired by a passion for sound and an obsession with learning new things. Our business model is to share the results of our research and creative explorations in order to make it possible for us to continue this work and share future results with others so they, in turn, can build upon these results in their own creative and scientific work.
Composer, performer, reviewer, writer, Barton McLean has studied composition with Henry Cowell, performed on double bass in jazz ensembles and the Hudson Valley Philharmonic Orchestra, taught music composition and theory at Indiana University, and directed the Electronic Music Center at the University of Texas at Austin. In 1974, he began touring as the McLean Mix, performing electro-acoustic music with fellow composer Priscilla Taylor. We caught up with him to ask about his newest piece, which he composed in Kyma using his “Melodic Sculptor” technique.
Thanks for speaking with us, Bart! Your recent work, Illusions, true to its name, sometimes conveys the “illusion” of an acoustic ensemble in a physical space. Yet, on careful listening one hears moments of “physical impossibility” achievable only through synthesis or electronic manipulation. I assume this is at least one of the “Illusions” referred to by the title. But in the notes, you go further to say that “In the larger sense, every time a work of art or a musical composition is created, an illusion of reality accompanies it…” which reminds me of a quote attributed to Einstein that “Reality is merely an illusion, albeit a very persistent one.” Can you comment further on this idea?
I would turn Einstein’s famous quote on its head; an illusion IS reality. For example, Modal realism is the view, notably propounded by David Kellogg Lewis, that all possible worlds are as real as the actual world. In short: the actual world is regarded as merely one among an infinite set of logically possible worlds, some “nearer” to the actual world and some more remote. In music and art we do actually create our own illusion-reality, which, when perceived by others, is a reality to them as well, albeit one tempered by their own experiences. Perhaps that’s why music is such a powerfully emotion-engaging art.
I hear “Illusions” as a kind of sonic ‘curiosity cabinet’ with many doors, each opening into a different space with its own color and atmosphere — some playful, some melancholy, some ominous, and all intriguing! Did you name the individual sections?
I chose the five section titles from Priscilla McLean‘s poem (see below): (Golden, Dreamy, Colder, Music, Fantasy).
The illusion of summer flashes in the mist
of a golden morning, keeping fantasy alive
through dark, colder times, evoking the
dream of flowers and music.
One of my objectives was to make each of the five sections as different as possible (which, as it so happens also highlights the scope of multiple possibilities that Kyma has to offer. With the thousands of individual Sounds that I’ve created and collected over seven years of working with Kyma, I had much to choose from).
What is the connection between the music and the visuals in the video?
The visuals were created by my co-collaborator in this work, composer and media artist Stephen Dankner, who generously opened his library of fractal-generated visuals for my choosing. These visuals share my own concept of creating work which engages the emotional response of a live audio or visual experience but yet has its own set of perception parameters which go beyond what can be achieved by merely representational art or live performance. Is it or is it electronic? Hard to tell, but why would anyone care? Just relax and absorb the emotional responses.
Can you talk a little bit more about your “Melodic Sculptor” technique and how you use it to create these hybrid physical/acoustic/imaginary spaces and worlds?
My Melodic Sculptor takes an existing melodic line and scrambles it in an unprecedented variety of controllable ways. For the raw material, I begin with an improvisatory sample — jazz lick or contemporary extemporaneous single line, for instance. In the introduction to my Melodic Sculptor Tutorial (available, along with other tutorials I’ve written for the Insights magazine), I describe it in more detail:
Once sampled and converted into a spectrum sample, this is placed in a SumofSines shell to be manipulated and molded in such ways that the original may no longer be evident, instead migrating into a sea of possibilities of creation constrained only by the composer’s imagination and patience as each stage of the 16-stage StepSequencer affords a unique sound-sculpting opportunity. The basic procedure is to isolate each sequencer stage and develop it using the many parameters available here (sample start, forward or reverse sample and sequence, pitch transposition independent of sample speed, stretching sample stage time or sequence stage time, along with many random features attached to each area).
In other words, the finished melodic result can be extremely different from the original but yet retain its live-generated illusion. You can, for example, take the Star Spangled Banner and convert it into a Messiaen or Coltrane or Arnold Schoenberg-like melodic phrase. My tutorial gives examples and takes the user through each type of transformation.
How would you describe the state of being “in the flow” or in the creative flow? How do you get there? Do you have any tips or tricks that you’ve learned for staying there once you’ve gotten there? Have you set up your studio in a way to achieve or maintain an environment conducive to creative flow?
Because Kyma is so rich in possibilities and complexities, I have three* rules:
I never begin a work with rigid preconceived notions of what the piece will contain. Kyma is set up most amazingly to allow for boundless discovery, and the first step should be, not to begin a Timeline and start inserting Sounds in it, but rather to set up conditions in which I can discover combinations that have never been heard or which I could never imagine. I DO NOT START A COMPOSITION YET.
I first began with the Kyma Sound Library and, as I became more familiar with it, I began to develop my own stock library apart from that. These were all individual Sounds. I would take an existing Sound and try different samples. I would take a spectral Sound (SumofSines) and explore the galleries that appear when I convert an AIFF sample to spectrum, etc. I DO NOT START A COMPOSITION YET.
With the Multigrid, which allows me to instantly realize all sorts of different combinations of Sounds and their relation to each other, I take these Sounds and experiment with different groupings and presets. This further expands my understanding of what works and what doesn’t. When I have developed a rich understanding of capabilities and have collected a library of Sounds and Multigrids, then I use the Multigrid as a prototype, I find combinations that seem most promising…
THEN I CAN BEGIN A COMPOSITION. It could be a Timeline, or perhaps I might want to try an improvisation work using the Multigrid only. Or, on rare occasions, I can compose a work with one Sound only, as I did in my “Dreamy” section in Illusions.
*I said there were three rules, but there’s actually one more! I DO NOT PUBLISH THE COMPOSITION UNTIL I HAVE COMPLETED RULE 4:
Find someone you can trust to tell you the unvarnished truth about your work, and take to heart any criticism. Personally, my wife and fellow composer and collaborator Priscilla McLean (who also has been a professional reviewer for the Albany Times-Union, Musical America, and other publications) is my tormentor whose criticism is on the mark much of the time, presenting a viewpoint that I perhaps have not considered. By and large, much of the work I have composed since we were married (in 1967) has had to pass the “Priscilla test.” With this in mind, for every two works I finish, I will (roughly) discard one of them as being not up to my standards.
ONLY AFTER THIS REVIEW AND SUBSEQUENT REVISION WILL I PUBLISH MY COMPOSITION. As an aside, before publishing my “Illusions,” I spent a year of chasing after possibilities only to discard them as not fulfilling their initial promise. I’m reminded of the Thomas Edison quote that comes up on the front of the Pacarana — something about how his most productive activities in inventing were his failures.
Was it this one?
I have not failed. I’ve just found 10,000 ways that won’t work. —Thomas Edison
Right, Illusions is one of my successes. But it was arrived at only after finding and discarding dozens of ways that did not work.
In the past you’ve written that “without Kyma, I probably would have quit composing a few years ago.” Do you recall why it was that you had gotten discouraged with composing at some point? And what it is about Kyma that drew you back in? (I’m reminded of Al Pacino’s famous quote from the Godfather “Just when I thought I was out, they pull me back in!”)
Ha! My discouragement has more to do with my own creative approach, which puts often impossible demands on performers and also the current state and limitations of music notation. During my initial composing career, like most composers I wrote exclusively for traditional performers in orchestral, wind ensemble, vocal and chamber group configurations. Too often the performances just would not produce what I really wanted to say, either due to the notation limits, limits of performers’ time, or limits of the instrument itself. Particularly worrisome to me were the balances in larger ensembles, which, to my taste had to be perfect in order to satisfy my intent. And there are my own quirks as a composer, which dictate that I often need to hear something before I can write it down. And finally, the limitations of my electronic instruments from 1996-2013 increasingly did not keep up with my ideas and aspirations. (Just before switching to Kyma in 2014 my studio consisted of ARP, Ensoniq ASR-10, Yamaha TX81z, Kincaid modules, Macintosh, Alesis keyboards, 2 Korg Wavestations, 2 Prophet 2000s, etc, migrating in 2011 also to MAX/MSP as well, with an 8 track and 4 track and two 2 track decks, and using Studio Vision and other software).
With Kyma, I found that I could not only have exact control over all these limitations (instrument ranges and capabilities, balances, performer issues, and above all new sounds never heard before) but I could custom design each Sound with dozens of parameters that could then be exquisitely controlled in real time using keyframe protocols (as in video). Most remarkably with Kyma 7, the Multigrid provided an authentic way to rearrange, posit, change a group of Sounds as they work together, again speaking to my quirk of having to hear something before I enter it in a composition. For me, a whole new world of possibilities came before me—a world that I will never exhaust. To celebrate this new way of music creation, I wrote two versions of Discoveries which present the work in video, first with its Timeline, and then with selected Sounds used in the piece diagrammed. To me, if I were forced to produce one word that most encapsulates Kyma, it would be “discoveries”.
You’ve worked as a professional composer both inside and outside of academia. What are some of the advantages and rewards of life as an independent artist? Are there any aspects of university teaching that you miss?
Our freedom from academia resulted in our 39 years of touring and reacting directly with all sort of communities — concert series, university residencies, museums, installations, etc. In all these cases we were able to interact directly with listeners on the spot— a gift unavailable to those composers who limit themselves to performing for academic audiences only, and one which we valued highly (we retired from touring in 2013). There was also the creative freedom, knowing that all we had to do was to compose our best work and the only judge would be our listener communities, not academic committees. Being a free spirit and sort-of vagabond has its drawbacks (see “bad weather”) but also brings forth a kind of kinetic energy — a “not knowing where your income is coming from next year” kind of fear plus excitement — that cannot be reproduced in academia. I did, however miss my composition and some theory students, many of whom have become lifelong friends and colleagues.
What (if any) advice would you give to a composer just starting a career in 2021?
I wish I could be more optimistic about university teaching careers today. But even in the 1970s when we graduated, the supply-demand situation for university teaching began to turn negative. It is much worse now. And so, it is my opinion that current graduating composers should turn to other means to gradually open the door. I have written extensively about this topic, and I invite interested readers to take a look at “In Search of an Audience: Outside the Bubble,” as well as “HOW TO SURVIVE AND PROSPER AS REVEALED BY THE “NEWSLETTER NINE,” in which as the Independent Composer Representative on the SCI Executive Committee, I interviewed nine independent composers with insights as to how they started and maintained their careers outside the university.
You’ve chosen to live in an area of great natural beauty which also offers some extreme weather challenges. Have you always enjoyed the woods and the mountains?
We are fortunate to have a small mountain in back of our house, with an entry trail owned by my good friend Jim Dahoney. From my house to the top, it is 3 miles round trip and 650 feet in elevation. I do this every other day, no matter what the weather. My doctor says that it is this, more than anything else, that keeps my medical age “at least ten years younger” than my real age, which is 83. A few times each summer we go mountain hiking in the Lake George Wild Forest. I am an Adirondack 46er, by the way, which I attained mostly in the 1960s. During the early 2000s, Priscilla and I would spend our summers in British Columbia hiking around Jasper, Banff, Glacier, and Revelstoke parks, as well as Vancouver Island.
What is your favorite part of using Kyma? Your least favorite?
My greatest thrill in Kyma is in creating new Sounds. Next is fooling with the Multigrids, which always produce surprises, sometimes ugly, sometimes ho hum, but sometimes sublime. My least favorite activity in Kyma is the actual composing in the Timeline, where I really have to stop having “fun” and get serious. But I guess they say all play and no work makes a composer … ?
I sometimes get the impression that the artists who really take to Kyma also seem to be independent-minded, self-reliant individualists and non-conformists. Do you concur with this observation?
If I have any sort of independent streak it must have been nourished by study with Henry Cowell in 1963-4 at Eastman (I may very well be his last surviving student, as he had a crippling stroke soon after that and could not teach). When you studied with Henry it was an all-consuming affair; you ate dinner with him, you had informal gatherings at a student’s house, you played bridge with him and his lively wife Sidney, and you enjoyed his informal conversations where he assumed the mantle of being at the forefront of virtually every historical development in the early – mid 20th century. His independence from any stale academic trend was legendary, as he happily migrated between atonality, experimentation, Asian influence, polytonality, and at times unabashed tonality and even banality. When the whole composer world zigged into post-Webern serialism, Henry was deeply zagging with Indian ragas. The stodgy Eastman composition faculty endured him, while the students loved him. I hope that some of this has rubbed off on me, but the determination on whether or not I am a “self-reliant individualist and non-comformist” will, I’m afraid, have to be made by others. I just do what I do, day by day (as Henry would also say), and don’t really think about whether or not I’m that kind of person. I guess I’d like to be, but I really don’t know.
You interact with a lot of Kyma users through your tutorials. Have you noticed any commonalities among the user community?
As I consider my Kyma friends and read the posts, there may be two kinds of Kyma users: The ones who use Kyma as the vehicle for serious composition are the kinds of people who think long-term and can smoothly make the connection between the left and the right brain skills. The other group of Kyma users are what I would call the “process” users who are more interested in the technical challenges Kyma affords, and are not hesitant about pushing the envelope of what is possible, even beyond what might be practically useful or evident in the ultimate result. Kyma is fortunate to have both kinds of users working side by side, and in having a leadership team that nourishes both areas.
Who would have thought that, when I was struggling with algebra in the 8th grade, that I would someday be using it to fashion a spaceship through the universe of sonic possibilities!
In closing, I’d like to come back to Priscilla’s opening lines from Illusions:
The illusion of summer flashes in the mist
of a golden morning, keeping fantasy alive
through dark, colder times, evoking the
dream of flowers and music.
Do you think that music (listening to music, creating music, remembering music) can, in some way, help us get through “dark and colder times”?
Yes, definitely. And music (such as that produced with Kyma) that does not require us to sit shoulder to shoulder in an audience setting is particularly relevant in today’s pandemic world. In our living room, we have an Optomic video projector which projects onto a nine foot screen, with a state-of-the-art stereo sound system. I have noticed that, as each year has passed recently, more interesting music has appeared on YouTube. Every evening just before dinner I take an hour or less to listen to something. This allows me peace, energizes me, excites me, calms me—in fact this listening experience is my most enjoyable activity. Recently on YouTube I have been quite excited by the work of Austrian composer Olga Neuwirth and the Rumanian Horațiu Rădulescu, along with my perennial favorites Berlioz and Sibelius, not to mention (well, the list goes on and on). One of my favorites is Angelo Badalamenti’s Industrial Symphony #1, which we heard decades ago at BAM and happily discovered that same performance recently streamed on YouTube. In other words, the electronic arts which do not require audiences may very well have a more important place to play in our post-pandemic future.
Bart, thank you for taking the time to share your some of your musical philosophy and compositional approaches with us today! It’s all about Discovery!
As part of an on-going series of interviews with artists adapting to and emerging from the disruptions of 2020-22, we had a chance to speak with independent sound artist and photographer Will Klingenmeier to ask him about how he continued his creative work in spite of (or because of?) the restrictions last year, what he is up to now, and some of his hopes and inspirations for the future.
Will is a self-described omnivore of sound and noise, living as a borderline hermit and wanderer in order to focus on developing his unique artistic voice.
Thanks for taking the time to speak with us, Will!
Thank you so much for having me. I’m honored and happy to have the opportunity to talk about my work.
The colors and arrangement of your studio photo seem to capture just how central your studio is in your life (you’ve said that you spend more time there than in in any other place, and I think a lot of us can identify with that). Can you give us a brief description of how you’ve set up your studio workflow and say a bit about how you feel when you’re in the studio and in the flow?
My studio is really sacred ground for me. It’s a special place that I have spent thousands of hours in, as well as making. All of the acoustic treatment, gear racks, speaker stands and shelves I made with my dad over the course of many years. When I’m in the flow it’s like I’m on a tiny planet in my own little world, and it’s immensely satisfying. It’s a source of constant joy and a never-ending project.
At the moment, a lot of what I do is based in Kyma so it is the heart of my studio. I try to perform everything in real-time without editing afterwards; even if it takes longer to get there, it’s more satisfying. I have a few different set-ups including laptops and an old Mac Pro. I have mostly the same content on multiple computers not only for back-ups, but also to make it easier for traveling with a mobile rig. When I can’t avoid using a DAW, I have an old version of Pro Tools on my Mac Pro. Outside of Kyma, I have a handful of microphones, instruments, synthesizers, outboard gear and tape machines that I use. There’s a ton of cabling to the closet turned machine room which looks confusing, but my set-up is actually quite straightforward. Also, I leave all the knick-knacks out so my studio looks like a bricolage, but that way I know where everything is when I need it, and it’s only an arm’s reach away.
During the lockdowns and travel restrictions of spring 2020, you were caught about 7000 miles away from your studio. How did that come about?
I don’t go looking for tough situations, but I always seem to find them. On February 1st, 2020, I left Colorado for Armenia where I was scheduled to volunteer teaching the next generation of noisemakers; two week intensives at four locations around the country, including Artsakh, for a total of two months. I booked a one-way flight so I would be able to travel Armenia and the region, and I had another flight afterward to visit a friend in India, but needless to say that’s not what happened.
When I left the states COVID-19 was around but every day things seemed to stay about the same. For the first month in Armenia things were normal, there wasn’t even a documented case of COVID until March. I made it through two schools and I was at the third when I got a call that the World Health Organization had increased the epidemic to a pandemic. That’s when things changed. A driver came for me and took me back to Yerevan where I was told things were going to change and ultimately go into lockdown. The school said I could stay in their apartment or do whatever I felt I needed to, including going home. I didn’t really know what to do; who did? Throughout my travels I’ve learned not to be reactionary and instead concentrate on options and to realize a working solution. As such, I figured if the virus was going to spread it would be mostly in Yerevan, the biggest city, so I left Yerevan on my own and went to the Little Switzerland of Armenia, Dilijan. It’s an incredibly beautiful place in the mountains and I thought “I’ll lay low here for a few days and plan my next move.”
After five days I decided to go back to Yerevan, and found out that I could have a three bedroom apartment all to myself indefinitely. I thought, “that sounds a lot better than scrambling and traveling 7,000 miles across the world right now”, so I talked with my family and I decided to stay put. I was in that apartment by myself for six months and it was an incredible opportunity to turn inwards. As it turned out, I stayed until one day before my passport stamp expired so I took the experience as far as I possibly could. In hindsight, I don’t think it was ever impossible for me to leave, but I never actually pursued it until I absolutely had to. And I wouldn’t change any of it.
Can you describe how your sound work was modified by your “quarantine” in Armenia? How is it different from the way you normally work in Colorado? Has it changed the way you work now, even after you’ve returned home?
Basically everything I’m doing now has come out of exceedingly difficult situations that I persevered through. Several things happened that I’m aware of and probably even more that I’m unaware of and still digesting. I definitely breached a new threshold of understanding how to use my gear. And I absolutely learned to make the most of what’s on hand—to repurpose things, up-cycle and reimagine. I remember taking inventory of everything I had, laying it all out in front of me and considering what it was capable of doing, the obvious things to start like the various connection ports and so on, then I moved to the more subtle. By doing this I was able to accomplish several things that I didn’t think I could previously based solely on my short-sighted view.
This same awareness and curiosity has stayed with me and I’m really grateful for it. Essentially, it was realizing there’s a lot to be gained in working through discomfort. Moreover, I’ve heard that creativity can really blossom when you’re alone, so maybe that is some of what happened. Not that I held back much before this time, but now I really swing for the fences—I love surreal, far-out, subjective, ambiguous sounds. Creatively speaking, I’m most interested and focussed on doing something that’s meaningful to me, and then I might share it.
Once I returned to Colorado a different level of comfort and convenience came back into my life. Some things were left in Armenia and some things were gained in Colorado. I’ve made it a point to make the most of wherever I am with whatever I have. That’s something I’ve lived by, and now it is a part of me. It is nice to be back in my studio though. I’ve spent more time in this room than any other so I know it intimately. To be sure, I’ve always enjoyed my space and I thrive in solitude so having the situation I did in Armenia really brought out the best in me. Back in the states there are definitely more distractions, so since returning I’ve become even more of a night owl to help mitigate them.
Do you have Armenian roots? Can you explain for your readers what’s going on there right now?
I don’t have Armenian roots, at least none that I know of. That said, I have been told by my Armenian friends that I am Armenian by choice, and I will agree with that. They are a wonderful people and it breaks my heart with what is happening there now. A lot is going on and I don’t claim to understand it all, but there is definitely a humanitarian crisis—thousands dead, thousands of refugees and displaced peoples, and many severely injured people from a 44 day war launched by Azerbaijan. As a result, there are both internal and external problems including on-going hate from the two enemy nations Armenia is sandwiched between. It is a tough time and when I ask my Armenian friends about it they don’t even know how to put it into words. I can see the sadness and worry in their eyes though.
Throughout the pandemic, you’ve maintained connections and collaborations with multiple artists. Please talk about some of those connections, how you established them, how you maintained them, how you continued collaborating (both in terms of technological and human connections).
The first of these collaborations began when I received a ping from one of my good friends and fellow Kyma user Dr. Simon Hutchinson. He said he would like to talk about YouTube. I’d recently started ramping-up my channel (under the name Spectral Evolver) and he was looking to do the same. We discussed the potential for collaboration and cross-promoting our channels. At the time, I was making walking videos around Armenia and we eventually decided to create a glitch art series which used these videos as source material for Simon’s datamoshing. Additionally, we decided to encode the audio for binaural and to use Kyma as a part of the process. We started with short videos up to a minute long with varying content to run some tests, and then started doing longer videos around ten minutes once we figured out the process. We used Google Drive to share the files. It was immensely satisfying and a totally new avenue for me; he is incredibly creative and talented and I was thrilled to be taken along for the datamoshed ride.
Another collaboration during this period was with my long-time friend from college Tim Dickson Jr. He is one of my favorite pianists. Everything he plays is very thoughtful and he has developed this wonderfully minimalist approach. He was going through tough times and wanted to share as much of his creations as possible so I asked him if he wouldn’t mind sending me some of his compositions for me to reimagine. He uploaded about 15 pieces and I used only them and Kyma in creating our album ‘abstractions from underground‘.
‘The Unpronounceables‘ as a trio was yet another collaboration that started during lockdown. Somehow I finagled my way into the group comprised of İlker Işıkyakar and Robert Efroymson, both friends of mine from the Kyma community. At that point, there was still hope for KISS 2020 in Illinois and for us performing as a trio, therefore we needed to start practicing. The only way that was possible at the time was online. As I remember we had a few chats to check-in and say “˜hi’ then we jumped in and started feeling our way forward. The only guideline we ever had was to split the sonic spectrum in thirds, one of us would take the lows, one of us the mids, one of us the highs—that was İlker’s idea I think. However, that quickly disappeared and we all just did what we could. I remember using a crinkled piece of paper as an instrument and opening my window to let the sound of birds and children playing in the courtyard come into the jam. I didn’t have my ganjo yet so I used whatever I could to be expressive, it was really beneficial to work through that. To jam, we used a free application called ‘Jamulus’. Robert made a dedicated server and the three of us would join in once a week and play for 30 minutes or so. Surprisingly the latency wasn’t unbearable. We were doing mostly atonal, nonrhythmic sounds though and I suspect if we needed tight sync it would have been painful.
Finally, despite the lockdown, I was able to continue volunteering with the school once they made the transition to online. While brainstorming with the school on how it would be possible to offer workshops for kids in lockdown at home with seemingly limited resources, I realized I could immediately share what I had just learned: to use whatever you have where you are! The results were spectacular: they created a wonderful variety of sound collages created, recorded and edited mostly through cell phones. I assigned daily exercises to record objects of a certain material, for example, metal, glass, wood etc. I asked them to consider the way the object was brought into vibration by another object. Don’t just bang on it! We also discussed the opportunity to use sound to tell a story even if it’s a highly bizarre, surreal, subjective story. In the end, I learned far more from them than they did from me. I was asking them how they made sounds and I was taking notes. For these classes, we met through Google Hangouts and shared everything on Google Drive.
Your recent release, “the front lines of the war,” is the result of another collaboration, this time with Pacific rim poet/multi-disciplinary artist Scott Ezell. Visually, tactilely, sonically and verbally, you and Scott employ multiple senses to express the disintegration and despair of the “shatter zone”. What exactly is a shatter zone?
Shatter Zone has at least a double meaning. It started off in geology to describe randomly cracked and fissured rocks then came to also be used to describe borderlands usually with displaced peoples, so for me, it carries both of those resonances.
The sounds you created for this project are also intentionally distressed and distorted through techniques like bit-crushing, digital encoding artifacts and vocal modulations. Particularly striking is the section that begins “In the shatter zone too long…” whose cracking dripping conveys a miasma of putrescence and disease carried along on a dry wind. Can you describe how you did the vocal processing in this section (and any other techniques you employed in this section that you’d like to share)?
True, there’s a lot of different audio processes happening to make the album sound the way it does including all of the things you’ve mentioned. The piece you’re talking about is called “Shatter Zones” and it came to sound the way it does through Kyma granulation. Scott’s reading of the poem is one take of about 8 minutes and I used a SampleCloud to read through the recording in a few seconds longer than that and used a short grain duration of less than 1 second. What results is a very slightly slower reading, but the pitch of his voice stays the same. It’s just slow enough and granulated enough that we can feel uneasy but not so much that the words aren’t intelligible. It’s super important that we understand what he’s saying so I had to find the line between making an evocative sound and leaving his voice intelligible. I thought there was an obvious correlation to Shatter Zones and granulation which is why I decided to use that technique. The sound around his voice is granular synthesis as well but with long grains of 5 or more seconds. I had this field recording I did on my cell phone of a dwindling campfire, down to the embers, and it always sounded really interesting to me. I brought it into Kyma, replicated it, added some frequency and pan jitter so each iteration wouldn’t sound the same, and used the long grain technique.
What other techniques did you use to “degrade” and “distort” the sound? Did you intentionally choose to distribute the sound component on cassette tape for that reason?
We arrived at putting the sound on a cassette for several reasons. I think a physical element is a really meaningful necessity for music and with the chapbook there was already going to be a physical element so we knew we wanted physical music. These days there’s basically three options: CD, vinyl, cassette — all of which will have a noticeably different sound. We thought about this from the beginning instead of it being an afterthought. Initially, we were thinking vinyl but because of the cost and the length of the sides we realized it probably wasn’t the best solution, if it was even possible. Between a CD and a cassette, this content lends itself to the cassette format more than a CD, not that CD would have been a bad choice. We welcome what the cassette tape might do to the sound—a further opportunity for the medium to influence the art. Besides, over the last year or so I’ve been buying music on cassette—they are really coming back—and they don’t sound nearly as undesireably bad as you might think. In fact, some music sounds best on cassette. I think it’s really cool if an artist considers all the different formats and decides which one is best for the art—like a painter choosing what they will paint on, it matters, it changes the look. When listening to this project on cassette there is a noticeably different experience than listening digitally, and with the exception of one track I think it is all more desirable on the cassette version. As for other techniques for degradation and distortion, I used a lot of cell phone field recordings, there’s a few databent audio files and I recorded out of Kyma into another recorder and during that process a bunch of noise came along. I also have a 1960’s style tone-bender pedal I made a long time ago that’s super special to me, and all of the guitar tones went through that. Then, of course, there’s the stuff inside Kyma which you’ve already hinted at: Bitwise operators, granulation, cross-synthesis and frequency domain haze.
Do you think it’s ironic to use Kyma, a system designed for generating high-quality sound, to intentionally degrade and distort the audio?
Kyma is definitely known for high quality, meticulous sound, but that’s not actually why I got into it, or why I continue to use it. Over the years, I’ve developed and found a way of working that’s meaningful to me and it’s centered around Kyma: real-time creation, no editing, and performing the sound. Kyma thrives in that situation and is extremely stable in both the studio and live use, therefore it’s where I feel really creative and capable of expressing the sound in my head.
One way I think about Kyma is like a musical instrument, say a guitar. What does a guitar sound like? What could it sound like? When does it stop being a guitar sound? Seems to me there’s no end to that and definitely no right or wrong. Kyma is the same way for me. In fact, it never occurred to me that I was making lo-fi sounds in a hi-fi system. I think there’s a sadness when data compression is used for convenience and to miniaturize art, but when it’s used from the onset for a unique sound or look, like we’ve done here, that’s different.
Is this piece about Myanmar? Or is it about “every war” in any place?
Both. The spoken words are specifically based on Scott’s first-person experience of a Myanmar Army offensive against the Shan State Army-North and ethnic minority civilians in Shan State, Myanmar in 2015. Scott was smuggled into Shan State under a tarp in the back of an SUV. The Shan State Suite (side one of the cassette) tells this story. At the same time, the project is expressing a bigger realization as it explores the ways that global systems implicate us all in vectors of destruction and conflict in which “everyone is on the front lines of the war.” (This last sentence is Scott’s words, he said it perfectly for our liner notes, so I’ve just repeated it here.)
Can you offer your listeners any suggestions for remediating action?
It is incredibly difficult not to be lost in despair and discouraged by the things that we, especially Scott, have seen and experienced, but I think the work of art itself is actually meant to be a positive thing. It’s once we fully ignore and disregard things that hope is lost, or rather that we are taking an impossible chance hoping that it “works out for the best.”
I believe we have to make an effort and that effort can come in many forms. This project cracks open the door and offers a start to a conversation, a very difficult, long conversation, but a necessary one, I think. Scott and I have had success bringing this kind of content into different University environments including an ethics of engineering course at the University of Virginia. We shared a piece of art we made and shared our personal experiences and relationships to contested landscapes and marginalized peoples. The response we got from the professors and students suggested there would be on-going consideration for the topic.
Finally, there are lots of good people in the world and lots of human rights organizations seeking to stop violence both before and after it has started, so that’s something encouraging. There is hope and we can start to do better immediately even if only on a very small, personal level. In fact, that’s actually where it all needs to start, I think.
In an ideal world, what would you love to work on for your next project?
An installation of some kind. I really want to do something on a bigger, more immersive, more tactile scale using Kyma and sound as a medium of expression along with some of the visual forms I’ve been getting into. In the meantime I’ll continue to do what I’m doing!
What do you see as the future direction(s) for digital media art and artists? For example, you’ve gone all-in on developing your youtube channel for both educational and artistic purposes, video, and live streaming. How does youtube (and more) figure into your own future plans?
I think we are going to continue to see extremes and new forms of art. Artists, as a whole, always challenge and question which leads to new horizons. Digital media has certainly created many incredible and wonderful on-going opportunities for artists. I see VR/AR getting more and more capable as well as popular. Also, with the advent of everything becoming “˜smart’, new needs have arisen for artists. Like Kyma being used in the design of the Jaguar I-PACE, that kind of stuff. As for my YouTube channel, I have every intention of continuing it. I don’t know exactly what all the content is going to be, but that’s why I put ‘Evolver’ in the name.
As a coda, we strongly recommend that everyone check out the detailed description of ‘the front lines of the war’ on Will Klingenmeier’s website, where you can also place an order for a copy of this beautifully produced, limited-edition chapbook and cassette for yourself or as a gift for a friend.
It’s not an easy work to categorize. It’s an album, it is a video, it’s a chapbook of poetry, yet it’s so much more than that: it is a meticulously crafted artifact which, in its every detail, conveys the degradation, despoilment and degeneration of a “shatter zone”. It arrives at your mailbox in a muddied, distressed envelope with multiple mismatched stamps and a torn, grease-marked address label, like a precious letter somehow secreted out of a war zone, typed on torn and blood/flower stained stationary and reeking of unrelenting grief, wretchedness, and inescapable loss. It deals with many topics that are not easy or comfortable to confront, so be sure to prepare yourself before you start listening.
We had a chance to speak with lead sound-effects editor, Marco Lopez, to find out more about how he used sound to enhance the narrative. Born in Leipzig Germany to Cypriot and Chilean parents, it seems inevitable that Lopez would become a multi-lingual citizen of the world. A solid background of 7 years of classical piano lessons and music theory led to sound engineering studies in Santiago, but it was almost by chance, during a short course entitled ‘Sound Design for Film’, added in his final semester, where he discovered his true passion for sound design, launching him on what he describes as “an endless search for knowledge and techniques”.
You come from a Cypriot, Chilean, and German background, but what about the Swedish connection? How did that come about?
In 2013, I attended Randy Thom’s sound design masterclass in Hamburg. Prior to the masterclass, each of the participants received a 5 minute sequence from the film “How To Train Your Dragon” and given the assignment of adding sound design to that sequence. By the end of the masterclass and after listening to my sound design, the Europa Studio (Filmlance International) sound design team invited me to visit them at their studio in Stockholm the next time I was in town. Eventually I took the decision to take the next step in my professional growth and move to Sweden, and I was fortunate enough to start working right away with the Europa Sound Studio/Filmlance team.
When Filmlance International mixer and sound designer, Erik Guldager, who was doing sound design for the first two episodes of “A Class Apart”, invited me to join the team, I immediately agreed! It’s always great working with them. Due to the pandemic the communication was done mainly by email, or Zoom. It was very effective, as if we were in the same place.
Is the dialog in Swedish? How does language influence sound design?
The dialog is indeed in Swedish. For the last five years, I have been speaking exclusively in Swedish with my girlfriend which has helped me a lot to learn the language. I think that it is important to understand the dialog and the underlying codes that sometimes might be carried along in this way. It becomes easier to support the story with the proper sound effects and build a better sound around them.
From the title and the synopsis, it sounds like class differences and privilege are a central “character” in this story. Did you try to create a “sound” or ambience that would convey privilege, exclusivity and power for some of the scenes? How did you go about doing that?
Yes, that is correct and that becomes even more prominent due to the mysterious death of one of the students of the boarding school: Tuna Kvarn. The central character is very well described both with the picture and with the dialog, so we began by highlighting those moments and, once we were happy with our first attempt, we then started adding details around those moments and enhancing them.
As part of this process, the director requested that we use “unnatural sounds”, sounds that would not be normally present in a certain room or an exterior space. This request made the whole project even more exciting for me, because it allowed us to open an extra door of creativity and gave us the opportunity to experiment and create elements (which I, unofficially referred to as “non-musical drones”) that functioned well in the overall context.
One of the guidelines from the sound supervisor of the project, Boris Laible, was that we were after a feeling. That is an inspiring place for me to be in, because sometimes it takes several attempts to nail it, and it’s interesting to be able to witness the different versions that can be created with different sound effects. Eventually we selected a few of those non-musical drones, based on the fact that they were blending well with the rest of the sounds and they were supporting the scenes properly, but most importantly, they were not distracting the viewer away from the storytelling. We kept tweaking and readjusting the sound design the whole time until the very end.
How did you use Kyma on this project?
I used Kyma both as an external FX processor where it receives and sends a signal to a DAW, and for offline processing (for example, to generate the non-musical drones).
One interesting sound design challenge was to create the sound of a grandfather clock ticking that, during some scenes, would slow down or accelerate to imply that something that was being said or some behavior was off. For that, I imported the sound effect in the Tau Editor and after creating a Gallery folder, I found a Sound where I could shift the tempo without affecting the pitch of the sound.
Then I thought of adding a clock bell and stretching its ringing, in a similar way as in the scene from the “Barton Fink” by the Coen brothers, where Barton taps the bell to register his arrival at the hotel. For that I used the Sum Of Sines Sound where I would modulate its pitch and give some sort of movement in the sound.
I even used Kyma in order to add an extra element to a CCTV electrical interference noise. By combining an FFT analysis with Christian Vogel’s ZDFResonatorBank prototype from his ZDF Filters Kyma Capsules, I was able to create some variations that blended very well with other sound effects recordings that I already had in my SFX library.
For the non-musical drones I would create Galleries and go through all the options given and if a Sound sounded interesting to me, I would spend more time experimenting with creating presets. This procedure was the most time consuming but it definitely gave fantastic results! By the end of the project, I realized I had used Kyma to create 96 non-musical drones along with a few extra sound effects.
Every space had its own defined character and within a certain situation we would introduce the non-musical drones and blend them with the rest of the sounds.
Are there any things that are easier (or faster or more natural) to do in Kyma than in other environments?
Just by importing a sound in Kyma, creating Gallery folders of Kyma Sounds it’s luxurious, because you can choose which one best suits your idea. Also the fact that I can control a Kyma Sound with my Wacom tablet, a microphone or my keyboard, gives me the freedom to perform the sound however I want to, or according to what is happening in the picture.
Could you describe your sound design studio setup?
I work on a 5.1 system, both on Pro Tools and Nuendo with the RME UCX audio interface. I use the MOTU Traveler mk3 connected to Kyma. I recently started using Dante which allows me to share the interface that the DAW is connected to and it gives a stereo format with 48Khz. Otherwise, I’ll just connect the interfaces of Kyma and the DAW via ADAT.
Do you usually work to picture? Do you use any live controllers or MIDI keyboards?
I always work to picture. I sometimes use a keyboard but for Kyma, I use the Wacom tablet more often.
How do you build your sound library?
If there’s a sound effect that I don’t have in my library, I’ll go out and record it, or I’ll use Kyma to create what I am after.
Any advice for Kyma sound designers? Any resources you recommend?
The fastest way to get into Kyma is to open a sound effect in Kyma and create a Gallery folder based on the options you choose. Then go through each folder and see the different Sounds that Kyma has created for you.
Personally I think of Kyma as an instrument in that, the more you practice, the more you will start seeing results. At the same time you also need the theory, so you understand the powerful possibilities and philosophy behind Kyma. That is why I would strongly recommend to read the manual. Once you begin to understand how it works you will be able to start building your own Sounds based on what you envisioned in the first place.
Check the Kyma Q&A periodically and also ask questions there. You should also feel free to join the Kyma Kata Group! There’s a lot of great people that practice and share their knowledge on Kyma. I’d like to thank Charlie Norton, Andreas Frostholm, Alan Jackson and Pete Johnston, of the Kyma Kata group, who generously offered valuable suggestions and helped me out when it was needed.
What is the function of sound for picture?
Sound helps define the picture and brings up emotions that support the storytelling. In “A Class Apart” there were scenes where sound was underlining what was going on visually, but in other moments we would create something a bit differently from what was going on in the picture. I would say that in the last episode sound helped build up the tension gradually, right from the beginning until the very last scene.
Any tips for someone just starting out in sound design?
Give the best you can on the project you are working on, because your performance will open the door to the next project. Allow yourselves to make mistakes and learn from them. Especially in the beginning nobody expects from you to know everything. Later on, it can also happen that something that we might consider as a mistake, might trigger an idea to create something new and exciting. In every moment experience the world around you through your ears and hold those experiences in your memory. You never know when they will be a source of inspiration for you. Study as much as you can about sound design and meet other sound designers. Watch films. A lot! Two or three times the same film. Study them and listen to what the sound is doing in relation to the picture.
Tell us what you’re the most proud of in “A Class Apart” (don’t be shy!!)
I am proud because we delivered an exciting sound. The overall process was creative and fun. There were moments when it seemed overwhelming like there was too much to do, but I trusted the creative process and decided to enjoy it.
What kind of project would you most love to do the sound for next?
I would like to have the chance to work on an animation, a scifi or a thriller.
Finally, the most important question, where can we all binge watch “A Class Apart (Zebrarummet)”? It sounds really intriguing!!
A taste of some of Marco Lopez’ “non-musical drones” from A Class Apart (Zebrarummet)
Marco credits his first piano teacher, Josefina Beltren, with teaching him various ways to “perform the silence” in a piece of music. Clearly that early training has translated to his talent for creating different forms of meaningful “silence” to advance the story and lend character to rooms and spaces:
Opportunities for learning, listening, connecting, and inventing new ways to make music together in 2021:
Kyma software updates
If you haven’t done so in a while, please check the Help menu in Kyma for software updates to ensure that you have all the new (and free) features and enhancements we’ve been adding! Here are a few highlights:
A New Year’s gift from Domenico Cipriani: Samples 3rd party > Domenico Cipriani > drumstation to discover the new samples. Thanks, Domenico, and Happy New Year to all!
Kyma 7+ is compatible with Apple macOS 11 Big Sur (released November 25, 2020).
Kyma 7++ is a universal binary and runs on Apple Silicon M1 processor.
SampleTimeIndex, FunctionGenerator, MemoryWriter can now take a sample-rate triggers or gates.
OscillatorTimeIndex Frequency field can now update at the sample rate.
Displays in VCS now update when the mouse is down (when moving a fader, for example).
Substantial speed-ups & optimizations in Oscilloscope, Spectrum Analyzer, 2d & 3d aggregate faders controlled by SoundToGlobalControllers.
Wormhole now works with an Ethernet connection to the host computer (substituting for the FireWire connection from Paca(rana) to host computer.
A gift to Kyma users from Stephen Taylor of violin pizzicato and col legno samples performed by Olivia Taylor.
New Capytalk message tripWires: and segmentOfGrid:
Speed ups in the reading & writing of all Kyma file types.
and those are just the highlights. When you download, please check the release notes for more details.
Opportunities for learning Kyma:
Personal Kyma Coaching with Alan Jackson
Alan Jackson is providing online individual coaching sessions and small group classes in Kyma. There are 3 slots still available on Fridays for 1 hour individual Kyma coaching sessions. Sessions are personally tailored for each student and conducted online through screen-sharing and video conferencing. Jackson assigns homework or suggests a direction for self-study during the week (when help is available through Slack chat if you get stuck). At the end of the week (or every two weeks) you meet online to discuss what you’ve done then look at what you want to do next. You do everything in Kyma yourself, sharing your screen with Alan’s help and guidance. The cost is £60 (GBP) per session. More information…
Online Kyma Course led by Alan Jackson
Alan Jackson is also running a series of weekly online Kyma classes starting in early January 2021. There is a 2-hour online teaching session once a week on Friday at 16:00 UTC. Some self-study tasks are set and there is support over Slack throughout the course. The numbers are kept low (5 or 6 people) to ensure that the classes are participatory and collaborative. The provisional syllabus is:
8 January: Introduction to Capytalk
12 February: Introduction to Smalltalk in Kyma
19 March: Microsound (sound particles etc.)
Cost: £240 (GBP) for the 4 week course.
At Carnegie Mellon University, Professor Joe Pino plans to continue using remote-access Kyma in his Advanced Sound Skills course, so students can log in from their homes. Last semester he discovered an unexpected side-benefit to remote access: his students could also use their own laptops during in-person classes, thus saving time cleaning and sanitizing before, after and during class. In the spring semester, he will be trying to minimize the time that people have to be in a room together, although he still hasn’t figured out how to deal with analog modular synths, since half the attraction of that is having one’s hands on the devices.
At the University of Oregon, Professor Jeff Stolet has put together an online Pacarana farm — 7 Paca(rana)s running on Mac Minis — that Future Music Oregon students can access for coursework via the Splashtop remote desktop software. Students in Jeff Stolet’s Electronic Composition and Digital Audio & Sound Design courses as well as those in Jon Bellona’s Data Sonification course will use the Pacarana farm for the 2021 winter term (beginning 4 January).
At Indiana University, Professor Chi (“Irisâ€) Wang has two Pacaranas that she has set up for hybrid online and in-person access for her undergraduate and graduate courses in COMPUTER MUSIC: DESIGN and PERFORMANCE.
At New York University (NYU), Professor Joel Chadabe has set up two Pacas online for students in his Creating with Interactive Media: Kyma course for spring semester 2021 (classes start at the end of January). Chadabe’s course will be structured around the history, composing philosophy, and ideas presented on his site: https://joelchadabe.net/
At Dankook University in Cheonan (Seoul), Professor Kiyoung Lee is using a Kyma/Pacarana system for his courses on Advanced Sound Design-Kyma System, Electronic Music Ensemble, Computer Music Composition. During his 2021 sabbatical, he is finishing a book on Kyma.
Wave terrain on Lines
Ben Phenix (aka wave.terrains) has a tutorial on Lines taking you step by step through how to create a granular engine from scratch in Kyma
Summer course at SPLICE Institute
SPLICE Institute is an online, week-long, intensive summer program for performers, composers, and composer-performers interested in music that combines live performance and electronics. Applications are now being accepted for SPLICE Institute 2021, which will take place online 28 June – 2 July 2021. This year, SPLICE includes daily workshops with Carla Scaletti on the Kyma system, with remote desktop access to Kyma systems for those who do not already have access to their own systems. Registration is $200 for the week (scholarships available).
Roland Kuit book coming soon…
Roland Kuit is working on a book about Kyma to be finished in time for KISS 2021. In the meantime, there is an outline, loaded with sound examples, on his site.
Jeffrey Stolet’s new work in response to events this summer.
Jim O’Rourke live Steam(room)
Inferential: Sounds produced / processed in Kyma by Jim O’Rourke, recorded at the Steamroom in Japan, November 2020 with a shout-out to the Kata: Alan, Simon, Ian, Pete, Domenico, Ben and Charlie
Robert Jarvis: sonoraV19
Sound artist Robert Jarvis’ interview describing his evolving sound art piece based on numbers of confirmed cases of COVID-19, using Kyma to map the numbers to the amplitudes of harmonics (one per country) to create a slowly evolving timbre. Listen to the (regularly updated) sound here.
Lucretio live stream
On 29 December 2020, Domenico Cipriani (Lucretio) performed a 45-minute live stream using Kyma on Seoul Community Radio. Listen again…
Anne La Berge and the poetics of wind turbines
Site V, new sound art from data & the poetics of wind turbines by Anne La Berge (Flute & Kyma) & Phil Maguire (Doepfer Synth), inspired by Borssele Wind Farm Site V floating wind turbines stationed where water depths are too great for fixed foundations.
Ben Phenix on SoundCloud
On SoundCloud, you can listen to Ben Phenix’s extensions to physical modeling and granular synthesis: “The primary goal was to move beyond the normal per grain controls we see (amplitude, pitch, pan, grain size, etc) and be able to insert other processing at the individual grain level. For example, each grain could have its own filter settings, delay, or wave shaper. At that point, we may be departing from Curtis Road’s classical definitions but that is where the fun is!”
Looking back to move forward: Gualtieri’s BTF Project
Composers and performers have been adapting to the constraints of the pandemic in various ways while inventing new paradigms for live musical performance. Composer Vincenzo Gualtieri has chosen to focus on one of the essential aspects of making music together: attentive and reciprocal listening. Gualtieri’s work can now be heard on a new album: the (BTF) project, released on EMA Vinci Records.
Mei-ling Lee‘s Doctoral Recital ranges from gun violence to darkly poignant children’s stories.
Matteo Milani: Wasted Planet
Matteo Milani and Giovanni Dettori – Wasted Planet, an album created during the first lockdown, electronic layers synthesized or processed in Kyma, released on Miraloop Records in May 2020.
Will Klingenmeier’s SpectralEvolver channel includes an ever-evolving collection of sounds-of the-day, binaural walks, Kyma tutorials and more!
News from Kyma users:
Silence and a sense of space
Laura Tedeschini Lalli (Università di Roma Tre) presented the keynote lecture “Silence and a sense of space: Rome, Italy during complete lockdown and after†for “Sounds of the Pandemic†virtual conference 16-17 December 2020.
Delora KymaConnect 2.0
Doug Kraul’s KymaConnect 2.0.0 BETA 6 adds support for macOS 11.0 “Big Sur†and includes, for the first time, native operation on Apple Silicon Macs. This release was built with the latest releases of the macOS tools and SDK. The application binary is in Apple’s “Universal Binary†format that places native executables for both Intel and Apple Silicon processors in the same application bundle. macOS will automatically choose the Intel version if your Mac has an Intel processor, or select the Apple Silicon version if you have one of the new Macs with the M1 processor.
Amazon acquires Zoox
Amazon and Zoox announced in June 2020 that they’ve signed an agreement for Amazon to acquire Zoox — the pioneer ride-hailing company dedicated to designing autonomous technology from the ground up. Aicha Evans, Zoox CEO, and Jesse Levinson, Zoox co-founder and CTO, will continue to lead Zoox as a standalone business. Zoox has a very special sound-design team creating a sonic user experience for passengers and pedestrians, and rumor has it that a few of those sounds may have made their way into the backing music for the Zoox 14 December reveal video.
Craig Vear awarded a European Research Council Grant
Congratulations to Craig Vear for being awarded a 5-year European Research Council grant for frontier research into the Digital Score. Of 2506 applicants, 327 researchers received ERC_Research funding for research, to build up their teams & have far-reaching impact. Vear plans to continue to expand upon the foundational work begun in his book The Digital Score: Musicianship, Creativity and Innovation which features interviews with several Kyma practitioners and a photo of Anne La Berge on the cover.
In remembrance of those we have lost in 2020…
With sadness, we remember two friends and colleagues. We feel their absence and we salute them for their contributions to sound and music and the world:
Stanley Cowell (5 May 1941 — 17 December 2020)
A brilliant composer, pianist, and friend, Stanley Cowell’s masterful use of Kyma to enhance and extend the piano in ways that feels both musical and inevitable can be heard on his album Welcome to this New World (https://news.symbolicsound.com/2013/07/welcome-to-this-new-world/) and other recordings. Soft-spoken and deep-thinking, Cowell fervently believed that artists can have an affect on the world: “We are not just artists…we are citizens of the world. In our own personal ways, and when necessary, in unity with others, we should add our ‘fuel’ to the cleansing fire against injustice.â€
https://www.wbgo.org/post/stanley-cowell-pianist-composer-and-educator-kaleidoscopic-view-jazz-dead-79#stream/0
Eugenio Giordani (21 March 1954 — 4 April 2020)
Eugenio Giordani’s reverberation algorithm (EuVerb) has touched the sounds of nearly every composer/sound designer in the Kyma community. We were saddened earlier this year to hear that Eugenio died of COVID-19 in April 2020. In September and October, there were several concerts held in honor of this polymath engineer/composer/performer https://www.lacertezza.it/commossi-omaggi-a-un-maestro/. And Kyma users will remember Eugenio each time they hear his exquisite, smooth reverberation enhancing their sounds and music.
Thanks to all of you for continuing to create and to give voice to the experiences of our times. We wish you a healthy, creative and productive 2021 filled with new ideas, new energy, and new sounds!
Over the course of 2020/21, composers and performers have been adapting to the constraints of the pandemic in various ways and inventing new paradigms for live musical performance. Composer Vincenzo Gualtieri has chosen to focus on one of the essential aspects of making music together: attentive and reciprocal listening.
“BTF†stands for feed-backto feed-forward, because of the extensive use of feedback principles in this work. In a deeper sense, though, BTF also stands for looking back to move forward.
Based on the premise of attentive and reciprocal listening, there is, in (BTF), an encounter between “electronic sounds” (processed in real time) and “acoustic soundsâ€. This interaction takes place within an environment of self-regulated feedback. The DSP system responsible for the digital treatment of sound metaphorically “listens” not only to the external acoustic energy state but also to its own own internal states. The same is required of the musical instrument performer.
As Gualtieri puts it, “Sound-producing systems – human and machine – create networks of circularly causal relationships and chains of self-generating (self-poietic processes), mutually co-determining sound events.”
Composing in this way requires an adaptation, a paradigm shift. If the digital processing of sound is linked to the acoustic properties of the environment, the quantity (and quality) of the generated sound events is no longer entirely predictable. Therefore the musical score asks the performer to take into account both the performance instructions and, at the same time, to continuously listen to the product of electronic processing and be freely influenced by it.
For (BTF) 1-4, Gaultieri worked in tandem with other musicians, with the composer managing live-electronics. Starting with (BTF) -5, he decided to perform the acoustic instruments himself. So, from (BTF) -5, onward, Gualtieri has automated the entire DSP process, resulting in two types of sound events: one managed by the digital system responsible for electronic processing (Kyma) and the other guided by music notation or word processing.
Capturing the spirit of invention inspired (and compelled) by worldwide lockdowns and isolation, Gualtieri writes:
La temporanea rinuncia a collaborare con altri musicisti mi ha permesso di concentrarmi diversamente
sull’esplorazione sonora e le sue forme organizzative.
– V. Gualtieri
(In English: “Temporarily giving up on collaborating with other musicians allowed me to focus differently
on sound exploration and its organizational forms.”)
An inspiring manifesto for what has and continues to be a challenging time for composers and performers worldwide.
Networked collaboration, telematic performances, and online learning have been growing in popularity for several years, but the lockdowns and social-distancing guidelines precipitated by the global COVID-19 pandemic have accelerated the adoption of these modes of interaction. This is a brief (and evolving) report on some of the solutions your fellow Kyma-nauts have found for practicing creative collaborations, live performances, private tutoring, consulting, and teaching large online courses. Thanks for sharing your input, feedback and alternative solutions for distance-collaboration with the Kyma community!
Note: For example configurations showing how to get audio onto your computer and out onto a network, read this first.
Kyma Kata
One of the earliest ongoing examples is Alan Jackson’s Kyma Kata, a regular meeting of peers who practice Kyma programming together, that has been operating online in Google Hangouts for over a year before the crisis and recently celebrated their 100th session! (Did they know something the rest of the world didn’t?) The Kyma Kata currently meets twice a week, on Mondays and Tuesdays. They begin each session with “Prime Minister’s Question Time” (open question-and-answer session on how to do specific tasks in Kyma), followed by an exercise that each person works on independently for 30 minutes, after which they share and discuss their results. Ostensibly the session lasts for 2 hours, but when people really get interested in a problem, some of them stick with it for much longer (though there is no honor lost if someone has to leave after two hours).
Kata participants focus on how to do something together, so screen-sharing is important and audio quality has been less important: they often play over the air, using the computer’s built-in microphone to send the audio.
For higher quality audio, Alan uses a small USB mixer plugged in to the Mac as the Hangouts audio source. Using the mixer, he can mix the Paca’s output and a microphone which provides a lot better quality than over-the-air through the laptop’s mic, although it’s still limited by Hangout’s audio quality, delay and bandwidth.
What is the Kata and How do I sign up?
The term, kata, which comes to us by way of karate, has been adopted by software engineers as a way to regularly practice their craft together by picking a problem and finding several different solutions. The point of a kata is not so much arriving at a correct answer as it is to practice the art of programming.
In the Kyma Kata, a group of aspiring Kymanistas come together regularly via teleconferencing and the facilitator (Alan) introduces an exercise that everyone works independently for about half an hour, after which people take turns talking about their solutions. All levels of Kyma ability are welcome, so why not join the fun?
Improvisation with the Unpronounceables
The Unpronounceables are Robert Efroymson in Santa Fe New Mexico, Ilker Isikyakar in Albuquerque New Mexico, and Will Klingenmeier (ordinarily based in Colorado, but due to travel restrictions, on extended lockdown in Yerevan Armenia). To prepare for a live improvisation planned for KISS 2020, Robert proposed setting up some remote sessions using Jamulus.
Collaboration platform
Using the Jamulus software, musicians can engage in real-time improvisation sessions over the Internet. A single server running the Jamulus server software collects audio data from each Jamulus client, mixes the audio data and sends the mix back to each client. Initially, Robert set up a private server for the group, but they now use one of the public Jamulus servers as an alternative. One of the amusing side-effects of using the public server is that they are occasionally joined by uninvited random guests who start jamming with them.
During a session, the Unpronounceables use a Slack channel to communicate with each other by text and Jamulus to time-align and mix the three audio sources and for sending the mix to each of the three locations.
Audio from Kyma
Each Unpronounceable uses a second interface to get audio from Kyma to the host computer. Robert uses a Behringer to come out of Kyma, and an IO/2 to get to his Mac. Ilker sends his MOTU Track 16 audio outputs to a Behringer; then selects the Behringer as an I/O in the Jamulus preference tab. Will uses a ZOOM H4n as his Kyma interface and sends the audio to an M-Audio Fast Track Pro which acts as the interface for Jamulus.
Ecosystemic audio in virtual rooms
Scott Miller and Pat O’Keefe’s HDPHN project has always been an exploration of what it means to be alone together — it’s a live public concert where each member of the audience wears headphones, rather than listening through speakers. When stay-at-home orders made it impossible for Scott in Otsego to meet in person with Pat in St. Paul, Minnesota, they started looking into how to move the live HDPHN performance onto the Internet.
When Earth Day Art Model 2020 shifted from a live to an online festival, Scott and Pat used this as an opportunity to dive in and perform HDPHN along with one of their older pieces Zeitgeist live through the Internet.
Audio from Kyma
Scott describes the audio routing for HDPHN as follows:
Pat’s mic comes over Zoom and out of my desktop headphone audio. It also goes into Kyma input 1 on my Traveller. With Zoom, I can’t get/send stereo from a live source. With two people (did this Friday) I bring the second person in on a separate Skype/Zoom/Facetime session on another device, and into Kyma input 2. With 2 inputs, I then mathematically cross-processing them in a virtual room.
I am sending Kyma’s processed/mixed output (Main 1-2) back into my desktop via Lynx E22 audio card, going into DSP-Quattro for compression and EQ, then to iShowU virtual audio interface 1) —> to Zoom for Pat’s monitoring, and 2) —>OBS and then to YouTube synced with Pat’s Zoom video. YouTube latency very bad and it wrecked chamber music with duo, but was fun for free improv with a different duo.
Live coding at a Virtual Club
On Monday, 11 May 2020, beginning at 20.30 CET Time, Lucretio will be live-coding in Kyma using a new Tool of his own design at The Circle UXR Zone. The Circle is a virtual club that is a UXR.zone — a decentralized social VR communication platform offering secure communication, free from user-tracking and ads. A spinoff project of the #30daysinVR transhumanist performance by Enea Le Fons, a UXR.zone allows participants to share virtual rooms and communicate via avatars across a range of devices: from VR headsets (main platform) to desktop and mobile phone browsers.
The avatars of people attending the UXR events are either anonymous (robots) or follow a strict dress code based on the CVdazzle research to stress the importance of cyber camouflage via aesthetics against dystopian surveillance measures happening in the real world. Click to enter the club
The belly of the BEAST
Simon Smith (who is actually quite a slender chap at the BEAST of Birmingham) has recently been tasked with researching online collaboration platforms for the BEAST, so we asked him for some tips from the front lines. He just completed a Sound and Music Workshop with Ximena Alarcon (earlier he helped Alarcon on a telematic performance using the Jacktrip software from Stanford).
In the workshop, she also mentioned:
Artsmesh A network music and performance management tool. Content creators run the Artsmesh client which streams live media point-to-point; audiences run a light Artsmesh client to watch the shows.
SoundJack is a realtime communication system with adjustable quality and latency parameters. Depending on the physical distance, network capacities, network conditions and routing, some degree of musical interaction is possible.
How have you been collaborating, teaching, consulting, creating during the lockdown? We’re interested in hearing your stories, solutions and experiences.
Have you used YouTube or Vimeo for live streaming with Kyma?
What’s your preferred video conferencing software for sending computer audio (Zoom, BigBlueButton, Meet)?
Have you been using remote desktop software (like Chrome Remote Desktop, Jump) to access your studio computer from home?
We welcome hearing about alternate solutions for this ongoing report.
Generate sequences based on the patterns discovered in your MIDI files
Symbolic Sound today released a new module that generates endlessly evolving sequences based on the patterns it discovers in a MIDI file. HMMPatternGenerator, the latest addition to the library of the world’s most advanced sound design environment, is now available to sound designers, musicians, and researchers as part of a free software update: Kyma 7.25.
Composers and sound designers are masters of pattern generation — skilled at inventing, discovering, modifying, and combining patterns with just the right mix of repetition and variation to excite and engage the attention of a listener. HMMPatternGenerator is a tool to help you discover the previously unexplored patterns hidden within your own library of MIDI files and to generate endlessly varying event sequences, continuous controllers, and new combinations based on those patterns.
Here’s a video glimpse at some of the potential applications for the HMMPatternGenerator:
What can HMMPatternGenerator do for you?
Games, VR, AR — In an interactive game or virtual environment, there’s no such thing as a fixed-duration scene. HMMPatternGenerator can take a short segment of music and extend it for an indeterminate length of time without looping.
Live improvisation backgrounds — Improvise live over an endlessly evolving HMMPatternGenerator sequence based on the patterns found in your favorite MIDI files.
Keep your backgrounds interesting — Have you been asked to provide the music for a gallery opening, a dance class, a party, an exercise class or some other event where music is not the main focus? The next time you’re asked to provide “background†music, you won’t have to resort to loops or sample clouds; just create a short segment in the appropriate style, save it as a MIDI file, and HMMPatternGenerator will generate sequences in that style for as long as the party lasts — even after you shut down your laptop (because it’s all generated on the Paca(rana) sound engine, not on your host computer).
Inspiration under a deadline — Need to get started quickly? Provide HMMPatternGenerator with your favorite MIDI files, route its MIDI output stream to your sequencer or notation software, and listen while it generates endless recombinations and variations on the latent patterns lurking within those files. Save the best parts to use as a starting point for your new composition.
Sound for picture — When the director doubles the duration of a scene a few hours before the deadline, HMMPatternGenerator can come to the rescue by taking your existing cue and extending it for an arbitrary length of time, maintaining the original meter and the style but with continuous variations (no looping).
Structured textures — HMMPatternGenerator isn’t limited to generating discrete note events; it can also generate timeIndex functions to control other synthesis algorithms (like SumOfSine resynthesis, SampleClouds and more) or as a time-dependent control function for any other synthesis or processing parameter. That means you can use a MIDI file to control abstract sounds in a new, highly-structured way.
MIDI as code — If you encode the part-of-speech (like verb, adjective, noun, etc) as a MIDI pitch, you can compose a MIDI sequence that specifies a grammar for English sentences and then use HMMPatternGenerator to trigger samples of yourself speaking those words — generating an endless variety of grammatically correct sentences (or even artificial poetry). Imagine what other secret meanings you could encode as MIDI sequences — musical sequences that can be decrypted only when decoded by the Kyma Sound generator you’ve designed for that purpose.
Self-discovery — HMMPatternGenerator can help you tease apart what it is about your favorite music that makes it sound the way it does. By adjusting the parameters of the HMMPatternGenerator and listening to the results, you can uncover latent structures and hyper meters buried deep within the music of your favorite composers — including some patterns you hadn’t  even realized were hidden within your own music.
Remixes and mashups — Use HMMPatternGenerator to generate an never-ending stream of ideas for remixes (of one MIDI file) and amusing mashups (when you provide two or more MIDI files in different styles).
Galleries of possibilities — Select a MIDI file in the Kyma 7.25 Sound Browser and, at the click of a button, generate a gallery of hundreds of pattern generators, all based on that MIDI file. At that point, it’s easy to substitute your own synthesis algorithms and design new musical instruments to be controlled by the pattern-generator. Quickly create unique, high-quality textures and sequences by combining some of the newly-developed MIDI-mining pattern generators with the massive library of unique synthesis and processing modules already included with Kyma.
How does it work?
If each event in the original MIDI file is completely unique, then there is only one path through the sequence — the generated sequence is the same as the original MIDI sequence. Things start to get interesting when some of the events are, in some way, equivalent to others (for example, when events of the same pitch and duration appear more than once in the file).
HMMPatternGenerator uses equivalent events as pivot points — decision points at which it can choose to take an alternate path through the original sequence (the proverbial “fork in the roadâ€). No doubt you’re familiar with using a common chord to pivot to another key; now imagine using a common state to pivot to a whole new section of a MIDI file or, if you give HMMPatternGenerator several MIDI files, from one genre to another.
By live-tweaking the strengths of three equivalence tests — pitch, time-to-next, and position within a hyper-bar — you can continuously shape how closely the generated sequence follows the original sequence of events, ranging from a note-for-note reproduction to a completely random sequence based only on the frequency with which that event occurs in the original file.
Other new features in Kyma 7.25 include:
▪ Optimizations to the Spherical Panner for 3d positioning and panning (elevation and azimuth) for motion-tracking VR or mixed reality applications and enhanced binaural mixes — providing up to 4 times speed increases (meaning you can track 4 times as many 3d sound sources in real time).
▪ Multichannel interleaved file support in the Wave editor
• New granular reverberation and 3d spatializing examples in the Kyma Sound Library
and more…
Availability
Kyma 7.25 is available as a free update starting today and can be downloaded from the Help menu in Kyma 7. For more information, please visit:Â symbolicsound.com.
Summary
The new features in the Kyma 7.25 sound design environment are designed to help you stay in the creative flow by adding automatic Gallery generation from MIDI files, and the HMMPatternGenerator module which can be combined with the extensive library of sound synthesis, pattern-generation, and processing algorithms already available in Kyma.
Background
Symbolic Sound revolutionized the live sound synthesis and processing industry with the introduction of Kyma in 1990. Today, Kyma continues to set new standards for sound quality, innovative synthesis and processing algorithms, rock-solid live performance hardware, and a supportive, professional Kyma community both online and through the annual Kyma International Sound Symposium (KISS).
Kyma, Pacarana, Paca, and their respective logos are trademarks of Symbolic Sound Corporation. Other company and product names may be trademarks of their respective owners.
Sound designers, electronic/computer musicians and researchers are invited to join us in Busan South Korea 29 August through 1 September 2019 for the 11th annual Kyma International Sound Symposium (KISS2019) — four days and nights of hands-on workshops, live electronic music performances, and research presentations on the theme: Resonance (공명).
“Resonanceâ€, from the Latin words resonare (re-sound) and resonantia (echo), can be the result of an actual physical reflection, of an electronic feedback loop (as in an analog filter), or even the result of “bouncing†ideas off each other during a collaboration. When we say that an idea “resonatesâ€, it suggests that we may even think of our minds as physical systems that can vibrate in sympathy to familiar concepts or ideas.
At KISS2019, the concept of resonance will be explored through an opening concert dedicated to “ecosystemic†electronics (live performances in which all sounds are derived from the natural resonances of the concert hall interacting with the electronic resonances of speaker-microphone loops), through paper sessions dedicated to modal synthesis and the implementation of virtual analog filters in Kyma, through live music performances based on gravity waves, sympathetic brain waves, the resonances of found objects, the resonance of the Earth excited by an earthquake, and in a final rooftop concert for massive corrugaphone orchestra processed through Kyma, where the entire audience will get to perform together by swinging resonant tubes around their heads to experience collective resonance. Sounds of Busan — two hands-on workshops open to all participants — focus on the sounds and datasets of the host city: Busan, South Korea. In part one, participants will take time series data from Busan Metropolitan City (for example, barometric pressure and sea level changes) and map those data into sound in order to answer the question: can we use our ears (as well as our eyes) to help discover patterns in data? In part two, participants will learn how to record, process, and manipulate 3d audio field recordings of Busan for virtual and augmented reality applications.
Several live performances also focus on the host city: a piece celebrating the impact of shipping containers on the international economy and on the port city of Busan; a piece inspired by Samul nori, traditional Korean folk music, in which four performers will play a large gong fitted with contact mics to create feedback loops; and a live performance of variations on the Korean folk song: Milyang Arirang, using hidden Markov models.
Hands-on Practice-based Workshops
In addition to a daily program of technical presentations and nightly concerts (https://kiss2019.symbolicsound.com/program-overview/), afternoons at KISS2019 are devoted to palindromic concerts (where composer/performers share technical tips immediately following the performance) and hands-on workshops open to all participants, including:
• Sounds of Busan I: DATA SONIFICATION
What do the past 10 years of meteorological data sound like? In this hands-on session, we will take time series data related to the city of Busan and map the data to sound. Can we hear patterns in data that we might not otherwise detect?
• The Shape Atlas: MATHS FOR CONTROLLING SOUND
How can you control the way sound parameters evolve over time? Participants will work together to compile a dictionary associating control signal shapes with mathematical functions of time for controlling sound parameters.
• Sounds of Busan II: 3D SOUND TECHNIQUES
Starting with a collection of 3D ambisonic recordings from various locations in and around Busan, we will learn how to process, spatialize, mix down for interactive binaural presentation for games and VR.
Networking Opportunities
Participants can engage with presenters and fellow symposiasts during informal discussions after presentations, workshops, and concerts over coffee, tea, lunches and dinners (all included with registration). After the symposium, participants can join their new-found professional contacts and friends on a tour of Busan (as a special benefit for people who register before July 1).
Sponsors and Organizers
Daedong College Department of New Music (http://eng.daedong.ac.kr/main.do)
Dankook University Department of New Music (http://www.dankook.ac.kr/en/web/international)
Symbolic Sound Corporation (https://kyma.symbolicsound.com/)
Busan Metropolitan City (http://english.busan.go.kr/index)
Composer/saxophonist Andrew Raffo Dewar will be at Mills College Center for Contemporary Music on Monday, February 4th, 2019 to perform his work, Ghosts in the Uncanny Valley II — a 35-minute composition for acoustic quartet improvising with live electronics programmed in the Kyma sound design environment. Kyma analyzes, manipulates, and expands upon the sounds of the acoustic instruments in real time, creating an electronically “extended†(and altered) quartet. The quartet features the composer on sax, Gino Robair (prepared piano), Kyle Bruckmann (oboe/English horn), and John Shiurba (acoustic guitar).
On January 31st 2019, Dewar will be at the Santa Monica Library to perform the premiere of his new piece for soprano sax and Kyma as part of their Sound Waves series. This performance starts at 7:30 pm and is free and open to the public.