Difference between revisions of "User:Jerryestie"

From DigitalCraft_Wiki
Jump to navigation Jump to search
 
(49 intermediate revisions by the same user not shown)
Line 7: Line 7:
 
0902472@hr.nl
 
0902472@hr.nl
  
==Making is Connecting ==
+
==Practice Q7: Making is Connecting ==
 
For my second year project please look at [[User:Jerryestie/Making | Making is Connecting 2016]]
 
For my second year project please look at [[User:Jerryestie/Making | Making is Connecting 2016]]
  
 
[[File:workjr7.jpg|500px|Caption]]
 
[[File:workjr7.jpg|500px|Caption]]
  
=Quarter 9: Workshops & Assignments=
+
==Practice Q9 & Q10: Craft & Radiation ==
==Inspiration and Own Work ==
+
For my third year project please look at [[User:Jerryestie/Practice | Craft and Radiation 2017]]
  
===Breakfast Glove===
+
[[File:Gifcensor.gif|500px|Caption]]
One of my own projects together with Joëlle was the breakfast glove for paper strings and electronic things. We wanted to make a freeze frame high five device. A song will play when you high five each other, but only if you stand still during the song. Check the video here:[https://youtu.be/UF43N295ig0 Breakfast Glove]
 
  
===Dunes===
+
=Minor: Human?=
[https://www.studioroosegaarde.net/project/dune/info/ Studio Roosengaarde's Dune Project] was a project here in Rotterdam with lights that responded to users walking next to them.
+
[[File:Westworld-skele-fb-1.jpg‎]]
 +
[[File:Bladerr.jpg]]
  
==Radio Experiments==
+
==Week 1: Kick-Off==
===Part 1===
+
===Beginning===
With the tuner I had the idea of trying to find 'weird' noises and translating them into more humane sounds. So instead of looking for actual voices, I wanted to create them myself. I tried to do this with the program Audacity, using effects like Pitch etc.
+
My group with Jochem and Naomi had the body parts of index finger, pink and the eye(s). We started out with topics that we found fitting before coming up with a more concrete subject. Things like Identity and security came up often and also folklore and myths involving hands and eyes. For instance: How in Japan would cut off their pinks as a way of apology, or how ISIS followers point up their index finger to the sky.  
  
[[File:audajerry1.png|500px]]
+
===Initial Ideas===
[[File:audajerry2.png|500px]]
+
There were a few quick concepts we came up with before ending up with our hand-eye dis coordination box.  
  
[http://vocaroo.com/i/s0Mv0mqfzFku Original Sound File: First Test] & [http://vocaroo.com/i/s12rHJqiyUAR Edited Sound File: First Test]
+
=====Pointing fingers=====
 +
[[File:Wewantu.jpg]]
  
The first test I mostly played with pitch, I tried to remove the static and see if that created a human feeling. It didn't, in my opinion. Instead it just sounded more flat.  
+
When a finger points at you it creates this immediate connection. It can be a bit uncomfortable and we wanted to play with this feeling. One of the ideas was to walk into a room where fingers would follow you (through the use of a Kinect) and point at you. Another was a closed-off space where you would have to point into the camera before entering and then being confronted by the previous visitor pointing at you. There was a variation where it would connect the people pointing.
  
[http://vocaroo.com/i/s0UbY5D8OQfX Original Sound File: White Noise] & [http://vocaroo.com/i/s0YrF0LFZAmK Edited Sound File: White Noise Talking]
+
[[File:Sketchjer1.jpg]]
 +
[[File:Sketchjer2.jpg]]
  
Next, I tried to find a weak signal, where maybe you couid almost here a conversation but would be too confusing. It already begins to sound more human, yet some stretches of the 'conversation' have this glitchy sound to them that I couldn't get rid of. But just removing the noise and changing the pitch really helped.
+
=====Games=====
 +
There were several variations where we played with the idea of a 2-screen/2-user installation. This set-up would have two players use the installation against each other in several ways. The 'controller' would force the player to use only the pink and index in a rapid order to force a sort of carpal tunnel syndrome. The player who would use the controller the fastest would induce a flashing/annoying eye-strain creating image to the other player. Another variation to this was the cheat screen. The cheat screen would give one player, unbeknownst to the other player, a sort of cheat screen that could make things worse or better for both the player and its competition.
  
[http://vocaroo.com/i/s0EVstmv3W5b Original: Noise] & [http://vocaroo.com/i/s1CJcNlD55gH Edited: Noise in Deep Space]
+
[[File:Screeeeeeeeb.png]]
  
This was a 'happy accident'. Trying to make this noise more human I instead got the sound of a starship in deep space... I changed the pitch to a very low sound and changed the speed of the noise. This, together with messing around with the bass and treble (I was actually trying to decrease the bass) got this humming space sound.
+
===Hand-Eye Discoordination===
===Part 2===
+
With some time pressure from Tim we ended on a final concept of putting the coordination between eyes and hand on trial. The question basically became "what would happen if you can't see what your hand is doing anymore."
[[File:Buitenjr2.jpg| 300px |Looking for signals near the maritime museuml]] [[File:Buitenjr1.jpg | 300px | ]]
 
  
For the second workshop week we went outside to try and pick up some more signals. At first it was hard to find anything but eventually we were able to find two truckers communicating with each other. The conversation was very casual with some mentions of dissapointing children.
+
=====Buttons=====
 +
We decided to find several buttons and hide them from plain sight. By using these buttons you would get a visual feedback of what you were doing. We went to the piekfijn to try and fide specific buttons so that each one would be a different sensation.
  
==Workshop with Jon==
+
=====Arduino Coding=====
[[File:Workshopjr.jpg | 500px]]
+
I focused on coding the Arduino and getting an input from the buttons that are pressed. There had to be a way to have each button have an unique identifier so Processing would know which button was pressed, since all input came in as the same numbers. I did this by adding a line to the serialprint that adds a different letter per button. This way Processing was able to differentiate all the different inputs.
  
During this workshop we switched ideas around the table to the group in front of us. In our case, we first came up with the idea to create a sort of souvenir that allowed you to connect to the country of the souvenir through NFC or another connection method. Our given instructions were to make a device that changes sound into water waves. By connecting a speaker underneath foamboard we were able to make the sound move, either by music or voice call (on max volume)
+
=====Exhibition=====
 +
filler
 +
soon im too lazy to export pics atm
  
==Imaginary Radio Assignment==
+
=====Observations and Feedback=====
====Introduction====
+
One of the things that was immediately clear during the exhibition is that people didn't get that they had to turn around and watch the screen. They would either just look at the box and play around and then turn around after moving something or pressing a button. We did not communicate this properly or force the user to stand in the way we intended.  
For the assignment I worked together with Naomi van Maasakkers. For inspiration we decided to look into the past instead of into the future and looked up a database of sci-fi devices in books from the 40s to the 70s. We used the site [http://www.technovelgy.com/ Technovelgy] for this.
 
  
====Mediaglyphic====
+
In general most people didn't (immediately or at all) get what the project was about or what was going on. The interactivity helped spark people's interest but the concept wasn't easy to see or understand through our installation.  
After looking at several options, me and Naomi decided to go with the "mediaglyphic". A device for people not bothered to learn how to read: Translating words into pictures. In a way, this is happening right now on the web: A lot of information is distilled into quick headlines with 1 image, since nobody seems to take the time to sit down and read something anymore. From the 10-second time limit on Snapchat (which dissappears afterwards) and 140 characters-news on Twitter.  
 
  
====Mediaglyphic in the internet age====
+
The buttons were also too hard to find, whether this was actually a good or bad thing I haven't made my mind up about yet. I feel like it's in a middle ground right now, it could either be a lot harder and become this sort of brutal installation or easier but then how do you add a challenge to the lack of eye coordination?
We decided to combine this new era with the idea of the mediaglyphic by creating a web plugin that changes news articles to articles in GIFs. Each new color represents a new sentence and headlines are bigger than normal text. All GIFs were found by using giphy.com's search for the word that's displayed. We also made a poster together. We used a New York Times Article as an example.
 
  
=====Website=====
+
The visuals also had a disconnect with what you were doing, it didn't seem to be a whole product but rather two separate things.
[[File:Gifspage.jpg | 800px]]
+
==Sensor==
 +
For the sensor I want to use sound as an unlocking mechanism. Using a distinct sound or sounds made in a distinct order.
 +
===Research Links===
 +
[https://forum.processing.org/two/discussion/14502/create-tap-tempo-metronome Getting BPM]
  
The website can be found here: http://imaginaryradio.businesscatalyst.com/index.html
 
  
=====Poster=====
+
----
[[File:Poster_digitalcraft2.jpg | 500px]]
+
float bpm = 80;
 
+
float minute = 60000;
A moving poster can be found on: https://indd.adobe.com/view/f33c7246-1217-4a1e-afed-bf7d2b91051b
+
float interval = minute / bpm;
 
+
int time;
==Workshop with MICA==
+
int beats = 0;
...one early tuesday morning we were put into mixed groups of WdKA and MICA students. I was in the group post-radio aesthetics. We started our morning by splitting the group into different parts: Builders and concept. The builders were on their way to build the given the device while the rest of us were put in several other groups but in the end we all just decided to go by the name 'conceptual group' ;). There we came up with the idea (after sparring across a few ideas) to try and make radio waves a tangible object, while keeping it's visual experience. First, we used a hand-built radio receiver to record a radio sound. Next, we opened that sound file in Photoshop and converted it into 6 different images. For one of the objects we used fabric and the lasercutter: The lasercutter was used to engrave unto fabric the 6 radio imagery parts. The other technique was knitting (an old friend from Q7 assignment last year!) where we used the bitmap to create 2-color knitting pattern.
+
 
 
+
 
Though in the end the engraving worked well, in hindsight I would have loved to spend a full week on this assignment to get the full potential out of this assignment. We used a lot of 'conversion' (Record audio> open audio in photoshop > save as different format) and this meant that everything was sort of 'planned'. In my own vision I would have loved to see this signal being picked up by software or hardware that we created and would convert this to imagery in real life without any manual labour from our part. I sort of had a plotter in mind for this, but again, time constraints.
+
void setup() {
 
+
  size(300, 300);
Another idea was to use the knitted material and make it a video that shows the timeline of audio moving along with the knitted scarf, showing the visual and audio at the same time would make it easier to understand what these visuals meant (in my opinion.)
+
  fill(255, 0, 0);
 
+
  noStroke();
All in all a fun day and morning, I have kept contact with a few students as well and maybe we'll work together again in the future!
+
  time = millis();
 
+
}
[[File:Eyyyyok.jpg | 500px]]
+
 
 
+
void draw() {
=Quarter 10: Own project (Pre-Research)=
+
  background(255);
 
+
 
===Early Thoughts===
+
  if (millis() - time > interval ) {
For Quarter 10 and this project in general I haven taken an interest in NFC. This, relatively, new technology. In fact, I want to explore this very specific part of NFC products: The 'toy to life' Industry. Several major brands have been releasing these toys over the years that have NFC chips embedded in them. What these chips do is bring a toy into a video game world. While I have no interest to own these toys myself, I have always found them interesting. Especially the marketing and thoughts behind them that is being communicated to kids. The idea of a living 'thing' inside  your toy coming to life is of course very exciting.
+
    ellipse(width/2, height/2, 50, 50);
 
+
    beats ++;
I got inspired by the project of Boris and company (sorry, don't know everyone's name) with the sound recording of different technologies. I want to try and recieve this audio of an NFC chip inside this figure and find a way to automatically change this electronic sound to a more 'living' or 'organic' sound. To basically give this idea of actual 'life' inside these toys. This is jut my initial idea and I hope that during the coming weeks it will expand as I learn more and more about this technology.
+
    time = millis();
 
+
  }
 
+
 
===Checking out NFC tags===
+
 
[[File:Jerrrr.jpg | 700px ]]
+
  text(beats, 30, height - 25);
 
+
}
By using the app Any NFC Launcher on Android you can see the ID tag of NFC cards. By using this app you can launch a certain app when you connect an NFC card like your OV chip card. One weird discovery is the Dutch Identity Card: It seems to send out weird NFC ID codes, and it's almost impossible to pick up the same one. I've had 6 different codes so far but haven't gotten the same one twice?
+
----
 
 
===Random Bits===
 
[[File:Magnetic-fields.png]]
 
 
 
[https://www.youtube.com/watch?v=XECHcbakMIg  RFID Installed into arm]
 
 
 
[http://stackoverflow.com/questions/18121707/nfc-tag-to-count-number-of-taps Tracking the amount of times an NFC tag is used]
 

Latest revision as of 11:33, 4 September 2017

Main Information

Jerry Estié

0902472

0902472@hr.nl

Practice Q7: Making is Connecting

For my second year project please look at Making is Connecting 2016

Caption

Practice Q9 & Q10: Craft & Radiation

For my third year project please look at Craft and Radiation 2017

Caption

Minor: Human?

Westworld-skele-fb-1.jpg Bladerr.jpg

Week 1: Kick-Off

Beginning

My group with Jochem and Naomi had the body parts of index finger, pink and the eye(s). We started out with topics that we found fitting before coming up with a more concrete subject. Things like Identity and security came up often and also folklore and myths involving hands and eyes. For instance: How in Japan would cut off their pinks as a way of apology, or how ISIS followers point up their index finger to the sky.

Initial Ideas

There were a few quick concepts we came up with before ending up with our hand-eye dis coordination box.

Pointing fingers

Wewantu.jpg

When a finger points at you it creates this immediate connection. It can be a bit uncomfortable and we wanted to play with this feeling. One of the ideas was to walk into a room where fingers would follow you (through the use of a Kinect) and point at you. Another was a closed-off space where you would have to point into the camera before entering and then being confronted by the previous visitor pointing at you. There was a variation where it would connect the people pointing.

Sketchjer1.jpg Sketchjer2.jpg

Games

There were several variations where we played with the idea of a 2-screen/2-user installation. This set-up would have two players use the installation against each other in several ways. The 'controller' would force the player to use only the pink and index in a rapid order to force a sort of carpal tunnel syndrome. The player who would use the controller the fastest would induce a flashing/annoying eye-strain creating image to the other player. Another variation to this was the cheat screen. The cheat screen would give one player, unbeknownst to the other player, a sort of cheat screen that could make things worse or better for both the player and its competition.

Screeeeeeeeb.png

Hand-Eye Discoordination

With some time pressure from Tim we ended on a final concept of putting the coordination between eyes and hand on trial. The question basically became "what would happen if you can't see what your hand is doing anymore."

Buttons

We decided to find several buttons and hide them from plain sight. By using these buttons you would get a visual feedback of what you were doing. We went to the piekfijn to try and fide specific buttons so that each one would be a different sensation.

Arduino Coding

I focused on coding the Arduino and getting an input from the buttons that are pressed. There had to be a way to have each button have an unique identifier so Processing would know which button was pressed, since all input came in as the same numbers. I did this by adding a line to the serialprint that adds a different letter per button. This way Processing was able to differentiate all the different inputs.

Exhibition

filler soon im too lazy to export pics atm

Observations and Feedback

One of the things that was immediately clear during the exhibition is that people didn't get that they had to turn around and watch the screen. They would either just look at the box and play around and then turn around after moving something or pressing a button. We did not communicate this properly or force the user to stand in the way we intended.

In general most people didn't (immediately or at all) get what the project was about or what was going on. The interactivity helped spark people's interest but the concept wasn't easy to see or understand through our installation.

The buttons were also too hard to find, whether this was actually a good or bad thing I haven't made my mind up about yet. I feel like it's in a middle ground right now, it could either be a lot harder and become this sort of brutal installation or easier but then how do you add a challenge to the lack of eye coordination?

The visuals also had a disconnect with what you were doing, it didn't seem to be a whole product but rather two separate things.

Sensor

For the sensor I want to use sound as an unlocking mechanism. Using a distinct sound or sounds made in a distinct order.

Research Links

Getting BPM



float bpm = 80; float minute = 60000; float interval = minute / bpm; int time; int beats = 0;     void setup() {   size(300, 300);   fill(255, 0, 0);   noStroke();   time = millis(); }   void draw() {   background(255);     if (millis() - time > interval ) {     ellipse(width/2, height/2, 50, 50);     beats ++;     time = millis();   }       text(beats, 30, height - 25); }