Notes: Since Mavericks, the Soft66lc SDR external has not been working in Max. Although I was not able to update the external, there is a temporary workaround.
Mac OS is hijacking the FTDI USB device with its own driver. You can unload the driver from terminal:
sudo kextunload -bundle com.apple.driver.AppleUSBFTDI
To reload the driver use “kextload”.
Here is article from Sparkfun with details about this workaround: https://learn.sparkfun.com/tutorials/how-to-install-ftdi-drivers/mac
And a more elaborate workaround that removes the Apple driver: http://www.mommosoft.com/blog/2014/10/24/ftdi-chip-and-os-x-10-10/
Notes about latency and FTDI http://openbci.com/forum/index.php?p=/discussion/199/latency-timer-os-x-new-info-plist
The real solution involves using the new Apple driver to communicate with the device: https://developer.apple.com/library/mac/technotes/tn2315/_index.html
Or spoofing the driver with a codeless kext: http://stackoverflow.com/questions/7263648/codeless-kext-loading-problem
Open source support for Softrock, SDR-IQ, K3, Lp-pan, and others.
By James Ahlstrom
Installation guide: http://www.hamradioandvision.com/installing-quisk-for-linux/
Three part tutorial.
By Andrew Pask at Cycling 74
Information between words: “Palin Biden Silences” David Tinapple, 2008: http://vimeo.com/38876967
Reverse engineering example: Naoto Fushimi https://reactivemusic.net/?p=18228
attributes : attrui
Also, be familiar with the difference between messages, attributes, and arguments with Max objects
“Palin Biden Silences” or “We used to be Friends”
Either sketch out the design, or make an example using Vizzie that demonstrates some aspect of the process.
midi osc thing / chat
voice cancellation thing
max for live – granulator or convolution reverb
show basics of max
make a spectrum analyzer
make a pitch detector tweet thing
in Max. It should look amazing. It should be the coolest control panel you can imagine. Use any objects, colors, shapes that you can find. But… it shouldn’t actually control anything.
Be ready to show it in class next week.
Pick any two.
Case study: Adapting and transforming an interactive video performance.
Video and programming by Naoto Fushimi
Here is a case study of how you might approach an interactive video project. For example, if you had an opportunity to design a realtime visualization tool for a band.
Knowing that cool video things can be designed with Max/MSP/Jitter, I opened up Google web search and typed in: jitter max
Here are the results:
The second entry looked interesting: [Share Patch] Simple Audio Reactive Set [Max/MSP Jitter]. (It is the cool video at the top of this post). There is a link to the Max patch in the video description – but the link is broken. So I entered the title of the video into Google: Share Patch Simple Audio Reactive Set
Here are the results:
The second entry is a link to the Cycling 74 forum. https://cycling74.com/forums/topic/looking-for-share-patch-simple-audio-reactive-set-maxmsp-jitter/ Here I was able to find a link to a modified version of Naoto Fushimi’s Max patch – in a post by Giorgio. http://1cyjknyddcx62agyb002-c74projects.s3.amazonaws.com/files/2014/12/NAOTO.maxpat
It appears that audio input is supposed to trigger the interactive graphics. But immediately on starting the audio, it begins to feedback, and there is apparently no way to stop it. I notice there is an ezadc~ object as well as an IO abstraction – but can’t quite see how they are hooked up due to the maze of segmented overlapping patch cords.
By selecting everything (<cmd> a) you can restore normal patch cords by selecting Arrange | remove all segments. After doing this, it appears that the ezadc~ is redundant, so I deleted it. Now the patch looks like this:
Much easier to understand. In the IO abstraction I loaded an audio file and then was able to get graphic output by toggling on the metro at the top of the patch and using the 3 toggles in the center of the patch.
3 openGL subpatches
“KSP is a game where the players create their own space program.”
Online archive 1939-2005.
teacher: Tom Zicarelli – http://tomzicarelli.com
You can reach me at: [email protected]
Office hours: Tuesday 1-2 PM, or Tuesday 4-5PM, at the EPD office #401 at 161 Mass Ave. Please email or call ahead.
Assignments and class notes will be posted to this blog: https://reactivemusic.net before or after the class. Search for: ep-426 to find the notes
Examples, software, links, and references demonstrated in class are available for you to use. If there is something missing from the notes, please ask about it. This is your textbook.
Everybody calls this course “The Jitter class” – referring to Max/MSP jitter from Cycling 74. You will learn to use Jitter. But the object is to create interactive visual art. Jitter is one tool of many available.
The field of interactive visual art is constantly evolving.
After you take the course, you will have designed projects. You might design a new tool for other artists. You will have opportunities to solve problems. You will become familiar with how others make interactive art. You will explore the connection between sound, video, graphics, sensors, and data. You will be exposed to to a world of possibilities – which you may embrace or reject.
We will explore a range of methods and have opportunities to use them in projects. We’ll look at examples by artists – asking the question: How does that work?
Topics: (subject to change)
Grades are based on two projects that you will design – and class participation. Please see Neil Leonard’s EP-426 syllabus for details. I encourage and will give credit for: collaboration with other students, outside projects, performances, independent projects, and anything else that will foster your growth and success.
I am open to alternative projects. For example, if you want to use this course as an opportunity to develop a larger project or continue a work in progress.