Week 5: How to Clean Without a Broom
Hi everyone! Welcome back to Week 5 of my blog! I’m glad to see you all tune back in for some exciting updates on my progress with my project on the commodification of gua sha. Now, I hope this week’s title wasn’t too misleading for you all: today’s blog post isn’t about that type of cleaning but actually about the last step of my data collection process: cleaning the data set.
If you don’t remember from last week, I had just finished collecting my full sample of data. I collected 200 scripts total: 50 each from Instagram, TikTok, Facebook, and X. I thought it was a tedious process on Instagram and TikTok because I had to manually transcribe the video audio. What I didn’t realize then was that cleaning the data set would be even more tedious. I went through each piece of data from each app. First, I consolidated what I had transcribed from the video audio and whatever text had appeared on screen/in the description box. I had decided at the beginning of data collection to analyze the video’s audio and text, so I kept that consistent throughout the whole process. Then, I went and deleted any filler words in the transcribed scripts (i.e. “um”). Since I had transcribed each content creator’s exact words, there were a lot of filler words that would’ve made analyzing the data more difficult. Now that I have a more optimized data set, I’m confident that my analysis and coding process will be a lot more manageable!
Speaking of the analysis and coding process, that’s what I plan to start next week! I’ve already prepared two coding charts: one for authentic language and one for commodified language. I’m planning to finish coding at least half of my data set next week but that’s pretty ambitious. Check back in next week to see if I accomplish my goal!
Comments:
All viewpoints are welcome but profane, threatening, disrespectful, or harassing comments will not be tolerated and are subject to moderation up to, and including, full deletion.