论文标题
“事实不是事实吗?”:分析用户验证对基于智能手机的调查中总线内/输出检测的影响
"Is not the truth the truth?": Analyzing the Impact of User Validations for Bus In/Out Detection in Smartphone-based Surveys
论文作者
论文摘要
乘客流程可以通过公共网络研究用户的行为,并协助设计新的设施和服务。通过乘客和基础设施之间的相互作用可以观察到这种流动。对于此任务,蓝牙技术和智能手机代表理想的解决方案。后一个组件允许用户的标识,身份验证和计费,而前者则允许短程隐式交互,设备对设备。为了评估这种用例的潜力,我们需要验证蓝牙信号和相关机器学习(ML)分类器如何违背现实背景的噪音。因此,我们对二元乘客国家就可以进入或出局的公共车辆(BIBO)进行建模。 Bibo标签标识了连续价值的乘客流的基本构建块。本文介绍了在半控件环境中的人机交互实验设置,其中涉及:两辆在两条路线上运行的自动驾驶汽车,服务三个公交车站和十八个用户,以及一个专有的智能手机 - 蓝图感应平台。最终的数据集包括对同一事件的多个传感器测量值和两个接地级别,第一个是参与者验证的,第二次是三个视频 - 视频胶带监视巴士和轨道。我们进行了标签 - 绑带的蒙特卡洛模拟,以模仿标签过程中的人类错误,这是在智能手机调查中发生的。接下来,我们使用此类翻转标签进行了监督ML分类器的培训。错误对模型性能偏差的影响可能很大。结果表明,ML对人类或机器错误引起的标签翻转的耐受性最高30%。
Passenger flow allows the study of users' behavior through the public network and assists in designing new facilities and services. This flow is observed through interactions between passengers and infrastructure. For this task, Bluetooth technology and smartphones represent the ideal solution. The latter component allows users' identification, authentication, and billing, while the former allows short-range implicit interactions, device-to-device. To assess the potential of such a use case, we need to verify how robust Bluetooth signal and related machine learning (ML) classifiers are against the noise of realistic contexts. Therefore, we model binary passenger states with respect to a public vehicle, where one can either be-in or be-out (BIBO). The BIBO label identifies a fundamental building block of continuously-valued passenger flow. This paper describes the Human-Computer interaction experimental setting in a semi-controlled environment, which involves: two autonomous vehicles operating on two routes, serving three bus stops and eighteen users, as well as a proprietary smartphone-Bluetooth sensing platform. The resulting dataset includes multiple sensors' measurements of the same event and two ground-truth levels, the first being validation by participants, the second by three video-cameras surveilling buses and track. We performed a Monte-Carlo simulation of labels-flip to emulate human errors in the labeling process, as is known to happen in smartphone surveys; next we used such flipped labels for supervised training of ML classifiers. The impact of errors on model performance bias can be large. Results show ML tolerance to label flips caused by human or machine errors up to 30%.