With the revolution of information technology greatly changing contemporary’s lifestyle, smartphone becomes a
necessity to more and more people, and individuals’ screen time is increasing as people are spending more time on
the virtual world. Through literature review, this paper proposes a hypothesis that there is a correlation between
smartphone screen time and walking step counts, which is verified through an exploratory research: data of walking
step counts and screen usage were widely collected by online and onsite questionnaires, and the analyses reveal
that 1) when daily screen time ranges from 4.99 to 15.25 hours, it is negatively correlated with daily walking step
counts; 2) the respondents’ average daily screen time is 6.3 hours, more of which (2.8 hours) is spent on social
Apps, and their average daily walking step counts are 6,750 (takes nearly 0.8 hour); 3) daily screen time and
walking step counts at weekends are less than those on weekdays; 4) people with higher education level, higher
income, or younger age use screen less daily; people with higher education level, lower exercise frequency, or
younger age walk less daily, so these factors are likely to affect the correlation between daily screen time and daily
walking step counts. The correlation of both behaviors and the affecting factors need to be further clarified, and the
impact of physical environmental elements on walking also requires more attention.
Urban greenery is closely related to people’s behaviour. With the advancement of science and technology in Artificial Intelligence, wearable sensors and cloud computing, the potential for studying the relationship between people and urban greenery through new data and technology is constantly being explored, such as assessing population exposure to urban greenery using multi-source big data. Taking one individual participant as a case study, this paper proposes and validates the effectiveness of using wearable camera (Narrative Clip 2) and machine learning (Applications Programming Interface of Microsoft Cognitive Service) to assess personal exposure to urban greenery. Microsoft API is used to identify urban greenery tags, including “flower”, “forest”, “garden”, “grass”, “green”, “plant”, “scene” and “tree”, in personal images taken by the wearable camera. Per- sonal exposure to urban greenery is assessed by calculating the frequency of the urban greenery tags in all the images taken. Furthermore, the overall evaluation and regularity of personal exposure to urban greenery (including “static exposure” and “dynamic exposure”) are explored to identify the characteristics of individual’s greenery lifelogging. This study makes a brave attempt that may contribute a new perspective in applying personal big data in studying individual behaviour.
Press by ArchDaily
2019年12月21日第八届深港城市/建筑双城双年展（深圳）（以下简称“深双”）开幕，这是目前全球唯一一个以城市和城市化作为固定主题的两年一度的展览。本届深双主题为“城市交互”，由麻省理工学院“可感知城市实验室”负责人卡洛 · 拉蒂、中国工程院院士孟建民、著名策展人与艺术评论家法比奥 · 卡瓦卢奇三人担任总策展人。展览包含“城市之眼”和“城市升维”两个板块，分别从不同角度探讨城市空间与科技创新之间不断发展的关系。本届展览汇集了来自24个国家及地区的280多位参展人，包括崔愷、王建国、王澍、张永和、刘慈欣、袁烽、MVRDV建筑事务所、藤本壮介、等国内外知名建筑师、艺术家和机构。
龙瀛团队作为首批受邀嘉宾参加了此次双年展，其项目“Digital Self, Daily Life and City Space”在展览中展出。该项目旨在探索新技术和新设备在记录个人生活和研究个人行为方面的潜力。通过邀请五位不同年龄和不同职业的志愿者佩戴可穿戴式相机一整周，对志愿者们的个人日常生活进行数字化记录，进而探寻电子产品在多大程度上改变了我们的生活。
Urban landscape is closely related to people's behaviors. With the emergence of technology, many electronic devices have been developed for self-monitoring pur-pose in studying the relationship of individual behavior and built environment, such as pedometers and arm-band sensors and wearable camera. Compared with the traditional methods using self-report, questionnaire or diary, using new tech-nologies and devices to measure and track individual behaviors and movement can obtain the first-hand data of human body passively and unconsciously. Besides that, combining with advanced computer technology of visualization, the potential of wearable devices in studying environmental exposure and psychological percep-tions has been constantly explored. Among them, wearable camera, a kind of port-able and micro photographed gadget, takes great advantage of tracking life details and recalling emotional preferences because it can take a photo every 30 seconds automatically and collect more than 1000 images each day. As a result, it become possible to build digital records of personal experience by collecting massive per-sonal image database. While this device has already been used in some experimental fields such as inter-medical science and computer science, involving interpreting personal lifelogging and treating memory disorders, but not yet been used in the study of urban space.
Taking one individual participant as an example, this paper will apply computer science in studying personal spatial behaviors and evaluating personal spatial ex-posure of greenness through analyzing personal image database collected by wear-able camera (Narrative Clip2). The data collection started from 08.10.2018 to 15.10.2018, and the participant should wear the camera from 8:00am to 11:00pm as well as keep it clipped on the same place of collar. Except private images, this process collected 8381 pictures in total and 1200-1500 photos each day in average. In order to identify and analyze database automatically, both Microsoft Cognition service API and Matlab will be used to process images and return information to identify greenery and evaluate the condition of personal exposure. Based on the Python language, Calling Microsoft Computer Vision API can extract a rich set of visual features based on the image content by identifying various “tags” appearing in each image, including environmental characters, figures and objects. This paper abstracts 41 tags related to outer space and greenness including ‘city’, ‘outdoor’, ‘flower’, ‘garden’, ‘grass’, ‘green’, ‘park’, ‘people’, ‘street’, ‘tree’ and so on, then calculates the duration of individual's exposure to the green environment. Another method to estimate personal greenness exposure is calculating the ratio of green and blue parts of each image by Matlab to verdict the continuity of the individu-al's exposure to the green environment and analyze the duration of every period of exposure. At the end, artificial audit will be applied to verify the validity of results come from Microsoft API and Matlab.
Extracting four types of tags according to the result of calling Microsoft API, in-cluding ‘city’, ‘outdoor’, ‘street’ and ‘green’. The result shows that proportions of four tags are 8.49%, 11.12%, 13.20% and 7.29% separately during the week, and all ratios witness various increase on the weekend, arriving at 12.04%, 14.88%, 15.68% and 8.32%. Compared with that, manual audit result explores that out-doors and green exposure during the week accounts for 16.88% and 14.4%, while on weekends accounts for 22.6% and 22.55%. Besides that, estimating the conti-nuity of greenery exposure by the green ratio calculated in Matlab, which shows that high greenery exposure usually appears in the commuting, eating out, going out and leisure time. It can be found that Microsoft API and Matlab color recogni-tion can basically reflect the trend of spatial exposure. However, due to the problems of lens position and usage, the quality of pictures has effects on the result.
This study has demonstrated that continuous and dynamic recordings of indi-vidual behaviors, including spatial- temporal information, is conducive to transfer people’s environmental exposure from perception to quantification. With the con-tribution of image recognition technologies and computer science, it has also been proved that state-of-the-art algorithms could pave a way to build a digital bridge between people’s behaviors and physical environment. However, some limitations still exist. It is essential to wear correctly and fixedly to obtain high qualitied pic-tures. While both Microsoft API and Matlab have its own boundedness, necessary artificial audit can explore more information and verify the result.