- Le tricodeur
- Journées du code créatif
- Processing Paris 2013
- Atelier OpenData à Nantes
- Cadavre exquis
- Processing Abidjan
- Processing Bordeaux Season #02
- Processing Paris 2012
- Processing Paris #03
- Processing Bordeaux #02
- Processing Bordeaux #01
- Formation Openframeworks
- Processing & Open Data à Rennes
- Processing Rennes
- Post Algo. Sérigraphiques
- Algorithmes sérigraphiques
- Post Processing Paris #02
- Processing Paris #02
- Processing, livres en français
- Contour map & blob detection
- Workshop “Exposer” à Pau
- panOramas 2010
- Award tour
- Processing Paris #01 photos
- Processing à l’EESI #02
- Processing à l’EESI #01
- Listen to my pictures
- Visuals @ Cool Catz party
- Club visuals
- *deggv @ l’atelier hypermedia
- Sevenfive visuals
- Processing @ Tweak
- Processing sketches 2002-2004
Post Processing Paris #02
Talented and creative people, a sunny weather and a careful organization were the ingredients of the highly successful second edition of Processing Paris.
Three workshops with different levels (beginners, intermediate & masterclass) were given at La Fonderie de l’Image during two days (22nd & 23rd of April).
I had the chance to teach the intermediate class, alongside with Christian Delecluse (beginners) and Harmut Bohnacker (masterclass).
First day : classes, containers & controls
The first day was dedicated to the creation of basic classes used to manage particles with position, velocity, acceleration and mass concepts.
Particles were then gathered into Arrays and then ArrayLists to manage as many instances as one wanted.
Application parametrization was shown through controlP5 library, emphasizing the relationships between user interface, data and their visual representation.
Second day : Twitter data visualization
On the start of the second day, typography sketches were practiced so that one becomes more and more familiar with the concepts behind arrays of objects while trying to revisit previous particles sketches. This led to some nice visual experimentations.
Then, twitter streaming API was explored, being able to retrieve messages linked to a particular hash tag for building a dictionary of word occurrences.
First of all, the data model was assembled and refined, the main function being a messages parser, without regular expression though. Then we ended building a simple visual representation of this data model, time lacking to add some nice physics to it.
Sketches by Alex LeGuillou & Emilie, Photo by Mark Webster.