Project description (written by the amazing Huiyi Chen):
The Evolution of COVID-19 is a web-based Interactive project that visualizes the evolution of the global Covid-19 Crisis since December 1st, 2019, the day when the symptom onset of the first patient was identified(1). Each date on the timeline will contain information of major events related to Covid-19, as well as worldwide data including number of total cases, active cases and total deaths.
For this project, I worked with Huiyi, Rui, Qice, and Viola. I mainly focused on the visual design and creating the p5js virus-y code effect. Huiyi and Viola worked on the data collection, Rui worked on the three.js part/front-end, and Qice worked on the back end.
This is a longer term project than we had imagined, partly because many of us (including myself) have been focused on completing our thesis. However, I managed to finish the the p5js ‘creative coding’ portion for the home and about page. Also, created the sketch designs, sharing some mockups below. Will aim to help with front-end + data collection/editing now.
Much more to come with this project!
Some Code Sketches I created of the Virus Logo
p5 sketch of the about page background
p5 sketch for the logo
Coding Resources for Making Drop Shadows with HTML5 API + p5js:
All thesis process can be found: here!
Link to site
Link to code
This dataset represents ferrous metal mines in the United States. It is based on the data collected from the Minerals Information Team (MIT) of the U.S. Geological Survey, and the operations are those considered active in 2003 and surveyed by the MIT. Dataset from data.world.
For the assignment to create a data visualization (using non-digital tools) on a postcard that shows a ‘snapshot’ of a particular slice of a public data set, I chose to use the NYC Department of Health and Mental Hygiene’s dataset on NYC Restaurant Inspection Results.
I was thinking of sending it to my cousin, who got food poisoning when he visited NYC a couple months ago.
Legend shows: (1) each violation type has their own pattern, (2) each restaurant type has their own color, and (3) critical/not critical changes the box’s drop shadow direction.
Below is the final postcard:
1.) Turning the Data Around
This article’s general sentiment to “consider the human” resonates strongly with me. The idea that data visualization should benefit the community or people in which the data is taken from seems only right. Reading this article, also makes me reflect on all the past data visualization projects I have done. I can’t help but wonder if our ‘New Yorkers who Clean Up and New Yorkers who Complain’ piece helped sanitation workers in any way – did it give back to them? Did any of my previous ‘Data + Art’ projects do that? I would like to think that it helped shed light on the fact that garbage collection is a dangerous job, but did it? Did it benefit the workers? How would I improve upon my previous data projects using this mindset of ‘turning the data around’? How would they feel if they saw the data visualization? Would they feel like their issues were being heard? Would they feel like they had been part of a study they didn’t agree to be a part of? Or worse, would they feel uncomfortable seeing this? I cringe thinking about it, but it’s very possible that it made them uncomfortable and not heard. It would have been better to get to know sanitation workers, to get to know the people first – not just make a data visualization. Good to be self-critical. It’s helpful reading this article and keeping in mind the community in which the data is taken from. I’d also like to know more ways to empower the people whom data is collected from other than putting it in a public setting.
2.) Chapter Two: On Rational, Scientific, Objective Viewpoints from Mythical, Imaginary, Impossible Standpoints
A great saying the article repeated was “data visceralization not visualization.” Designing for viscerlizations requires a much more holistic embodiment from the person. As the article mentions multiple times, it’s important to keep in mind that objective, rational “truths” are impossible. The argument to keep data “objective” is faulty in that everyone’s “truth” is different. We live in a subjective, non-binary world full of many unique truths. Another great point is the idea that novel representations of data are much more memorable than a typical bar graph. Even the “chart junk”, which are illustrated data visualizations, are more unique and therefore more effective. It was helpful to read this and remember that when making my own data visualizations it is important to a.) show not tell, b.) think of unique forms to represent this data, c.) convey the emotions of the dataset, and d.) design for “data viscerlizations” which require a “holistic conception of the viewer” and treats viewers as more than just a pair of eyes.
My NYU hosted website doesn’t work anymore, so I’m trying to get my previous live web + data art projects that use an ssl certificate up and running again. In hopes that I don’t forget how to do this, I’m documenting my process of hosting my own domain name, getting my own server, and registering for ssl certificates. Should be simple, but let’s see!
Step 1: registered for a new domain name on Dreamhost. My new domain name: dadododoes.tech
Step 2: Created a new droplet on digital ocean.
Step 3: Added a SSH key. This video was quite helpful in understanding how to get a SSH key on my DigitalOcean. With all this set up I am now able to ssh in using terminal
I am also able to log into fetch server:
Step 4: point my DigitalOcean name server from common Domain Registrar. This was a helpful video and article.
Step 5: updating my previous code with new ssl certificate and rsa private key that DreamHost has provided.
Step 7: ran into a lot of issues with not being able to run “node server.js” in my terminal once my file was in fetch. realized i have to install node in it. must use these commands to install node. live web class page has a helpful tutorial for this.
apt-get install nodejs
apt-get install nodejs-legacy
apt-get install npm
Step 7: also had to reinstall forever
Step 8: on Digital Ocean it needs to have the ‘www’ added to the url name that points to the ip address.
To get ssh key content:
pbcopy < ~/.ssh/id_rsa.pub