-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathtrain.json
7502 lines (7502 loc) · 882 KB
/
train.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
[
{
"arxiv_id":"2101.00183v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00183v1",
"title":"Early Prediction of Heart Disease Using PCA and Hybrid Genetic Algorithm with k-Means",
"summary":"Worldwide research shows that millions of lives lost per year because of heart disease. The healthcare sector produces massive volumes of data on heart disease that are sadly not used to locate secret knowledge for successful decision making. One of the most important aspects at this moment is detecting heart disease at an early stage. Researchers have applied distinct techniques to the UCI Machine Learning heart disease dataset. Many researchers have tried to apply some complex techniques to this dataset, where detailed studies are still missing. In this paper, Principal Component Analysis (PCA) has been used to reduce attributes. Apart from a Hybrid genetic algorithm (HGA) with k-means used for final clustering. Typically, the k-means method is using for clustering the data. This type of clustering can get stuck in the local optima because this method is heuristic. We used the Hybrid Genetic Algorithm (HGA) for data clustering to avoid this problem. Our proposed method can predict early heart disease with an accuracy of 94.06%.",
"published":"2021-01-01T00:00:00.000Z",
"authors":"['Md. Touhidul Islam', 'Sanjida Reza Rafa', 'Md. Golam Kibria']",
"arxiv_primary_category":"cs.lg",
"readability":0.9,
"actuality":4.3333333333,
"controversy":2.3333333333,
"relevance_magnitude":3.6666666667,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.5555833333
},
{
"arxiv_id":"2101.00289v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00289v1",
"title":"Design and Actuator Optimization of Lightweight and Compliant Knee Exoskeleton for Mobility Assistance of Children with Crouch Gait",
"summary":"Pediatric exoskeletons offer great promise to increase mobility for children with crouch gait caused by cerebral palsy. A lightweight, compliant and user-specific actuator is critical for maximizing the benefits of an exoskeleton to users. To date, pediatric exoskeletons generally use the same actuators as adult exoskeletons, which are heavy and resistive to natural movement. There is yet no easy way for robotic exoskeletons to accommodate the changes in design requirements that occur as a child ages. We developed a lightweight (1.65 kg unilateral mass) and compliant pediatric knee exoskeleton with a bandwidth of 22.6 Hz that can provide torque assistance to children with crouch gait using high torque density motor. Experimental results demonstrated that the robot exhibited low mechanical impedance (1.79 Nm average backdrive torque) under the unpowered condition and 0.32 Nm with zero-torque tracking control. Root mean square (RMS) error of torque tracking result is less than 0.73 Nm (5.7% with respect to 12 Nm torque). To achieve optimal age-specific performance, we proposed the first optimization framework that considered both motor and transmission of the actuator system that can produce optimal settings for children between 3 and 18 years old. The optimization generated an optimal motor air gap radius that monotonically increases with age from 0.011 to 0.033 meters, and optimal gear ratio varies from 2.6 to 11.6 (3-13 years old) and 11.6 to 10.2 (13-18 years old), leading to actuators of minimal mass.",
"published":"2021-01-01T00:00:00.000Z",
"authors":"['Sainan Zhang', 'Tzu-Hao Huang', 'Chunhai Jiao', 'Mhairi MacLean', 'Junxi Zhu', 'Shuangyue Yu', 'Hao Su']",
"arxiv_primary_category":"cs.ro",
"readability":0.79,
"actuality":2.0,
"controversy":1.3333333333,
"relevance_magnitude":2.6666666667,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":2.47225
},
{
"arxiv_id":"2101.00531v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00531v1",
"title":"Context-Aware Safe Reinforcement Learning for Non-Stationary Environments",
"summary":"Safety is a critical concern when deploying reinforcement learning agents for realistic tasks. Recently, safe reinforcement learning algorithms have been developed to optimize the agent's performance while avoiding violations of safety constraints. However, few studies have addressed the non-stationary disturbances in the environments, which may cause catastrophic outcomes. In this paper, we propose the context-aware safe reinforcement learning (CASRL) method, a meta-learning framework to realize safe adaptation in non-stationary environments. We use a probabilistic latent variable model to achieve fast inference of the posterior environment transition distribution given the context data. Safety constraints are then evaluated with uncertainty-aware trajectory sampling. The high cost of safety violations leads to the rareness of unsafe records in the dataset. We address this issue by enabling prioritized sampling during model training and formulating prior safety constraints with domain knowledge during constrained planning. The algorithm is evaluated in realistic safety-critical environments with non-stationary disturbances. Results show that the proposed algorithm significantly outperforms existing baselines in terms of safety and robustness.",
"published":"2021-01-02T00:00:00.000Z",
"authors":"['Baiming Chen', 'Zuxin Liu', 'Jiacheng Zhu', 'Mengdi Xu', 'Wenhao Ding', 'Ding Zhao']",
"arxiv_primary_category":"cs.lg",
"readability":0.83,
"actuality":4.3333333333,
"controversy":2.0,
"relevance_magnitude":3.6666666667,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.33325
},
{
"arxiv_id":"2101.00585v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00585v1",
"title":"UPSLAM: Union of Panoramas SLAM",
"summary":"We present an empirical investigation of a new mapping system based on a graph of panoramic depth images. Panoramic images efficiently capture range measurements taken by a spinning lidar sensor, recording fine detail on the order of a few centimeters within maps of expansive scope on the order of tens of millions of cubic meters. The flexibility of the system is demonstrated by running the same mapping software against data collected by hand-carrying a sensor around a laboratory space at walking pace, moving it outdoors through a campus environment at running pace, driving the sensor on a small wheeled vehicle on- and off-road, flying the sensor through a forest, carrying it on the back of a legged robot navigating an underground coal mine, and mounting it on the roof of a car driven on public roads. The full 3D maps are built online with a median update time of less than ten milliseconds on an embedded NVIDIA Jetson AGX Xavier system.",
"published":"2021-01-03T00:00:00.000Z",
"authors":"['Anthony Cowley', 'Ian D. Miller', 'Camillo Jose Taylor']",
"arxiv_primary_category":"cs.ro",
"readability":0.89,
"actuality":4.0,
"controversy":3.3333333333,
"relevance_magnitude":5.0,
"relevance_valence":4.444,
"newsworthiness_crowd_sum":4.1943333333
},
{
"arxiv_id":"2101.00633v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00633v2",
"title":"Outcome-Explorer: A Causality Guided Interactive Visual Interface for Interpretable Algorithmic Decision Making",
"summary":"The widespread adoption of algorithmic decision-making systems has brought about the necessity to interpret the reasoning behind these decisions. The majority of these systems are complex black box models, and auxiliary models are often used to approximate and then explain their behavior. However, recent research suggests that such explanations are not overly accessible to non-expert users and can lead to incorrect interpretation of the underlying model. In this paper, we show that a predictive and interactive model based on causality is inherently interpretable, does not require any auxiliary model, and allows both expert and non-expert users to understand the model comprehensively. To demonstrate our method we developed Outcome Explorer, a causality guided interactive interface, and evaluated it by conducting think-aloud sessions with three expert users and a user study with 18 non-expert users. All three expert users found our tool to be comprehensive in supporting their explanation needs while the non-expert users were able to understand the inner workings of the model easily.",
"published":"2021-01-03T00:00:00.000Z",
"authors":"['Md Naimul Hoque', 'Klaus Mueller']",
"arxiv_primary_category":"cs.hc",
"readability":0.91,
"actuality":4.3333333333,
"controversy":3.6666666667,
"relevance_magnitude":4.3333333333,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.9165833333
},
{
"arxiv_id":"2101.00646v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00646v1",
"title":"AttnMove: History Enhanced Trajectory Recovery via Attentional Network",
"summary":"A considerable amount of mobility data has been accumulated due to the proliferation of location-based service. Nevertheless, compared with mobility data from transportation systems like the GPS module in taxis, this kind of data is commonly sparse in terms of individual trajectories in the sense that users do not access mobile services and contribute their data all the time. Consequently, the sparsity inevitably weakens the practical value of the data even it has a high user penetration rate. To solve this problem, we propose a novel attentional neural network-based model, named AttnMove, to densify individual trajectories by recovering unobserved locations at a fine-grained spatial-temporal resolution. To tackle the challenges posed by sparsity, we design various intra- and inter- trajectory attention mechanisms to better model the mobility regularity of users and fully exploit the periodical pattern from long-term history. We evaluate our model on two real-world datasets, and extensive results demonstrate the performance gain compared with the state-of-the-art methods. This also shows that, by providing high-quality mobility data, our model can benefit a variety of mobility-oriented down-stream applications.",
"published":"2021-01-03T00:00:00.000Z",
"authors":"['Tong Xia', 'Yunhan Qi', 'Jie Feng', 'Fengli Xu', 'Funing Sun', 'Diansheng Guo', 'Yong Li']",
"arxiv_primary_category":"cs.lg",
"readability":0.88,
"actuality":3.0,
"controversy":3.0,
"relevance_magnitude":3.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":3.0278333333
},
{
"arxiv_id":"2101.00675v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00675v1",
"title":"Sentiment Analysis for Open Domain Conversational Agent",
"summary":"The applicability of common sentiment analysis models to open domain human robot interaction is investigated within this paper. The models are used on a dataset specific to user interaction with the Alana system (a Alexa prize system) in order to determine which would be more appropriate for the task of identifying sentiment when a user interacts with a non-human driven socialbot. With the identification of a model, various improvements are attempted and detailed prior to integration into the Alana system. The study showed that a Random Forest Model with 25 trees trained on the dataset specific to user interaction with the Alana system combined with the dataset present in NLTK Vader outperforms other models. The new system (called 'Rob') matches it's output utterance sentiment with the user's utterance sentiment. This method is expected to improve user experience because it builds upon the overall sentiment detection which makes it seem that new system sympathises with user feelings. Furthermore, the results obtained from the user feedback confirms our expectation.",
"published":"2021-01-03T00:00:00.000Z",
"authors":"['Mohamad Alissa', 'Issa Haddad', 'Jonathan Meyer', 'Jade Obeid', 'Kostis Vilaetis', 'Nicolas Wiecek', 'Sukrit Wongariyakavee']",
"arxiv_primary_category":"cs.ai",
"readability":0.88,
"actuality":3.3333333333,
"controversy":3.3333333333,
"relevance_magnitude":4.3333333333,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.58325
},
{
"arxiv_id":"2101.00744v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00744v1",
"title":"Learning to Optimize Under Constraints with Unsupervised Deep Neural Networks",
"summary":"In this paper, we propose a machine learning (ML) method to learn how to solve a generic constrained continuous optimization problem. To the best of our knowledge, the generic methods that learn to optimize, focus on unconstrained optimization problems and those dealing with constrained problems are not easy-to-generalize. This approach is quite useful in optimization tasks where the problem's parameters constantly change and require resolving the optimization task per parameter update. In such problems, the computational complexity of optimization algorithms such as gradient descent or interior point method preclude near-optimal designs in real-time applications. In this paper, we propose an unsupervised deep learning (DL) solution for solving constrained optimization problems in real-time by relegating the main computation load to offline training phase. This paper's main contribution is proposing a method for enforcing the equality and inequality constraints to the DL-generated solutions for generic optimization tasks.",
"published":"2021-01-04T00:00:00.000Z",
"authors":"['Seyedrazieh Bayati', 'Faramarz Jabbarvaziri']",
"arxiv_primary_category":"cs.lg",
"readability":0.81,
"actuality":3.6666666667,
"controversy":2.6666666667,
"relevance_magnitude":3.3333333333,
"relevance_valence":1.667,
"newsworthiness_crowd_sum":2.8334166667
},
{
"arxiv_id":"2101.00753v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00753v1",
"title":"Schemes of Propagation Models and Source Estimators for Rumor Source Detection in Online Social Networks: A Short Survey of a Decade of Research",
"summary":"Recent years have seen various rumor diffusion models being assumed in detection of rumor source research of the online social network. Diffusion model is arguably considered as a very important and challengeable factor for source detection in networks but it is less studied. This paper provides an overview of three representative schemes of Independent Cascade-based, Epidemic-based, and Learning-based to model the patterns of rumor propagation as well as three major schemes of estimators for rumor sources since its inception a decade ago.",
"published":"2021-01-04T00:00:00.000Z",
"authors":"['Rong Jin', 'Weili Wu']",
"arxiv_primary_category":"cs.si",
"readability":0.88,
"actuality":3.0,
"controversy":2.6666666667,
"relevance_magnitude":3.6666666667,
"relevance_valence":1.667,
"newsworthiness_crowd_sum":2.7500833333
},
{
"arxiv_id":"2101.00884v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00884v1",
"title":"Coreference Resolution in Research Papers from Multiple Domains",
"summary":"Coreference resolution is essential for automatic text understanding to facilitate high-level information retrieval tasks such as text summarisation or question answering. Previous work indicates that the performance of state-of-the-art approaches (e.g. based on BERT) noticeably declines when applied to scientific papers. In this paper, we investigate the task of coreference resolution in research papers and subsequent knowledge graph population. We present the following contributions: (1) We annotate a corpus for coreference resolution that comprises 10 different scientific disciplines from Science, Technology, and Medicine (STM); (2) We propose transfer learning for automatic coreference resolution in research papers; (3) We analyse the impact of coreference resolution on knowledge graph (KG) population; (4) We release a research KG that is automatically populated from 55,485 papers in 10 STM domains. Comprehensive experiments show the usefulness of the proposed approach. Our transfer learning approach considerably outperforms state-of-the-art baselines on our corpus with an F1 score of 61.4 (+11.0), while the evaluation against a gold standard KG shows that coreference resolution improves the quality of the populated KG significantly with an F1 score of 63.5 (+21.8).",
"published":"2021-01-04T00:00:00.000Z",
"authors":"['Arthur Brack', 'Daniel Uwe M\u00fcller', 'Anett Hoppe', 'Ralph Ewerth']",
"arxiv_primary_category":"cs.ir",
"readability":0.87,
"actuality":1.0,
"controversy":1.3333333333,
"relevance_magnitude":1.3333333333,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":1.4721666667
},
{
"arxiv_id":"2101.00929v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.00929v1",
"title":"Donut visualizations for network-level and regional-level overview of Spatial Social Networks",
"summary":"Spatial Social Networks (SSN) build on the node and edge structure used in Social Network Analysis (SNA) by incorporating spatial information. Thus, SSNs include both topological and spatial data. The geographic embedding of the nodes makes it impossible to move the nodes freely, rendering standard topological algorithms (e.g. force layout algorithms) used in SNA ineffective to visualize SSN sociograms. We propose a new visualization technique for SSNs that utilize the spatial and social information to provide information about the orientation and scale of connections. The donut visualization can be used to summarize the entire network or can be used on a part of the network. We demonstrate the effectiveness of the donut visualization on two standard SSNs used in literature.",
"published":"2021-01-04T00:00:00.000Z",
"authors":"['Dipto Sarkar', 'Piyush Yadav']",
"arxiv_primary_category":"cs.si",
"readability":0.76,
"actuality":2.6666666667,
"controversy":2.3333333333,
"relevance_magnitude":2.0,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.4445
},
{
"arxiv_id":"2101.01043v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.01043v1",
"title":"Caching at the Edge: Outage Probability",
"summary":"Caching at the edge of wireless networks is a keytechnology to reduce traffic in the backhaul link. However, aconcentrated amount of requests during peak-periods may causethe outage of the system, meaning that the network is not ableto serve the whole set of demands. The outage probability is afundamental metric to take into account during the networkdesign. In this paper, we derive the analytical expression ofthe outage probability as a function of the total amount ofusers requests, library size, requests distribution, cache size andcapacity constraints on the backhaul resources. In particular, wefocus on a scenario where end-users have no direct connectionto the master node which holds the complete library of contentthat can be requested. A general formulation of the outage isderived and studied for two relevant caching schemes, i.e. therandom caching scheme and the most popular caching schemes.The exact closed form expressions presented in this paper provideuseful insights on how requests, memory and resources can bebalanced when the parameters of a cache-enabled network haveto designed",
"published":"2021-01-04T00:00:00.000Z",
"authors":"['Estefan\u00eda Recayte', 'Andrea Munari']",
"arxiv_primary_category":"cs.ni",
"readability":0.81,
"actuality":3.6666666667,
"controversy":2.0,
"relevance_magnitude":4.6666666667,
"relevance_valence":1.667,
"newsworthiness_crowd_sum":3.0000833333
},
{
"arxiv_id":"2101.01078v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.01078v1",
"title":"Tensorizing Subgraph Search in the Supernet",
"summary":"Recently, a special kind of graph, i.e., supernet, which allows two nodes connected by multi-choice edges, has exhibited its power in neural architecture search (NAS) by searching for better architectures for computer vision (CV) and natural language processing (NLP) tasks. In this paper, we discover that the design of such discrete architectures also appears in many other important learning tasks, e.g., logical chain inference in knowledge graphs (KGs) and meta-path discovery in heterogeneous information networks (HINs). Thus, we are motivated to generalize the supernet search problem on a broader horizon. However, none of the existing works are effective since the supernet topology is highly task-dependent and diverse. To address this issue, we propose to tensorize the supernet, i.e., unify the subgraph search problems by a tensor formulation and encode the topology inside the supernet by a tensor network. We further propose an efficient algorithm that admits both stochastic and deterministic objectives to solve the search problem. Finally, we perform extensive experiments on diverse learning tasks, i.e., architecture design for CV, logic inference for KG, and meta-path discovery for HIN. Empirical results demonstrate that our method leads to better performance and architectures.",
"published":"2021-01-04T00:00:00.000Z",
"authors":"['Hansi Yang', 'Quanming Yao', 'James Kwok']",
"arxiv_primary_category":"cs.lg",
"readability":0.8,
"actuality":2.3333333333,
"controversy":2.3333333333,
"relevance_magnitude":4.0,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.8611666667
},
{
"arxiv_id":"2101.01337v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.01337v1",
"title":"Integration of Domain Knowledge using Medical Knowledge Graph Deep Learning for Cancer Phenotyping",
"summary":"A key component of deep learning (DL) for natural language processing (NLP) is word embeddings. Word embeddings that effectively capture the meaning and context of the word that they represent can significantly improve the performance of downstream DL models for various NLP tasks. Many existing word embeddings techniques capture the context of words based on word co-occurrence in documents and text; however, they often cannot capture broader domain-specific relationships between concepts that may be crucial for the NLP task at hand. In this paper, we propose a method to integrate external knowledge from medical terminology ontologies into the context captured by word embeddings. Specifically, we use a medical knowledge graph, such as the unified medical language system (UMLS), to find connections between clinical terms in cancer pathology reports. This approach aims to minimize the distance between connected clinical concepts. We evaluate the proposed approach using a Multitask Convolutional Neural Network (MT-CNN) to extract six cancer characteristics -- site, subsite, laterality, behavior, histology, and grade -- from a dataset of ~900K cancer pathology reports. The results show that the MT-CNN model which uses our domain informed embeddings outperforms the same MT-CNN using standard word2vec embeddings across all tasks, with an improvement in the overall micro- and macro-F1 scores by 4.97\\%and 22.5\\%, respectively.",
"published":"2021-01-05T00:00:00.000Z",
"authors":"['Mohammed Alawad', 'Shang Gao', 'Mayanka Chandra Shekar', 'S. M. Shamimul Hasan', 'J. Blair Christian', 'Xiao-Cheng Wu', 'Eric B. Durbin', 'Jennifer Doherty', 'Antoinette Stroup', 'Linda Coyle', 'Lynne Penberthy', 'Georgia Tourassi']",
"arxiv_primary_category":"cs.cl",
"readability":0.85,
"actuality":2.0,
"controversy":2.3333333333,
"relevance_magnitude":2.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.3611666667
},
{
"arxiv_id":"2101.01637v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.01637v1",
"title":"Theory-based Habit Modeling for Enhancing Behavior Prediction",
"summary":"Psychological theories of habit posit that when a strong habit is formed through behavioral repetition, it can trigger behavior automatically in the same environment. Given the reciprocal relationship between habit and behavior, changing lifestyle behaviors (e.g., toothbrushing) is largely a task of breaking old habits and creating new and healthy ones. Thus, representing users' habit strengths can be very useful for behavior change support systems (BCSS), for example, to predict behavior or to decide when an intervention reaches its intended effect. However, habit strength is not directly observable and existing self-report measures are taxing for users. In this paper, built on recent computational models of habit formation, we propose a method to enable intelligent systems to compute habit strength based on observable behavior. The hypothesized advantage of using computed habit strength for behavior prediction was tested using data from two intervention studies, where we trained participants to brush their teeth twice a day for three weeks and monitored their behaviors using accelerometers. Through hierarchical cross-validation, we found that for the task of predicting future brushing behavior, computed habit strength clearly outperformed self-reported habit strength (in both studies) and was also superior to models based on past behavior frequency (in the larger second study). Our findings provide initial support for our theory-based approach of modeling user habits and encourages the use of habit computation to deliver personalized and adaptive interventions.",
"published":"2021-01-05T00:00:00.000Z",
"authors":"['Chao Zhang', 'Joaquin Vanschoren', 'Arlette van Wissen', 'Daniel Lakens', 'Boris de Ruyter', 'Wijnand A. IJsselsteijn']",
"arxiv_primary_category":"cs.ai",
"readability":0.87,
"actuality":4.0,
"controversy":3.0,
"relevance_magnitude":4.0,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.58325
},
{
"arxiv_id":"2101.01677v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.01677v1",
"title":"Monocular Depth Estimation for Soft Visuotactile Sensors",
"summary":"Fluid-filled soft visuotactile sensors such as the Soft-bubbles alleviate key challenges for robust manipulation, as they enable reliable grasps along with the ability to obtain high-resolution sensory feedback on contact geometry and forces. Although they are simple in construction, their utility has been limited due to size constraints introduced by enclosed custom IR\/depth imaging sensors to directly measure surface deformations. Towards mitigating this limitation, we investigate the application of state-of-the-art monocular depth estimation to infer dense internal (tactile) depth maps directly from the internal single small IR imaging sensor. Through real-world experiments, we show that deep networks typically used for long-range depth estimation (1-100m) can be effectively trained for precise predictions at a much shorter range (1-100mm) inside a mostly textureless deformable fluid-filled sensor. We propose a simple supervised learning process to train an object-agnostic network requiring less than 10 random poses in contact for less than 10 seconds for a small set of diverse objects (mug, wine glass, box, and fingers in our experiments). We show that our approach is sample-efficient, accurate, and generalizes across different objects and sensor configurations unseen at training time. Finally, we discuss the implications of our approach for the design of soft visuotactile sensors and grippers.",
"published":"2021-01-05T00:00:00.000Z",
"authors":"['Rares Ambrus', 'Vitor Guizilini', 'Naveen Kuppuswamy', 'Andrew Beaulieu', 'Adrien Gaidon', 'Alex Alspach']",
"arxiv_primary_category":"cs.ro",
"readability":0.83,
"actuality":2.3333333333,
"controversy":1.6666666667,
"relevance_magnitude":3.0,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.4445
},
{
"arxiv_id":"2101.01686v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.01686v1",
"title":"Dynamic Hybrid Relation Network for Cross-Domain Context-Dependent Semantic Parsing",
"summary":"Semantic parsing has long been a fundamental problem in natural language processing. Recently, cross-domain context-dependent semantic parsing has become a new focus of research. Central to the problem is the challenge of leveraging contextual information of both natural language utterance and database schemas in the interaction history. In this paper, we present a dynamic graph framework that is capable of effectively modelling contextual utterances, tokens, database schemas, and their complicated interaction as the conversation proceeds. The framework employs a dynamic memory decay mechanism that incorporates inductive bias to integrate enriched contextual relation representation, which is further enhanced with a powerful reranking model. At the time of writing, we demonstrate that the proposed framework outperforms all existing models by large margins, achieving new state-of-the-art performance on two large-scale benchmarks, the SParC and CoSQL datasets. Specifically, the model attains a 55.8% question-match and 30.8% interaction-match accuracy on SParC, and a 46.8% question-match and 17.0% interaction-match accuracy on CoSQL.",
"published":"2021-01-05T00:00:00.000Z",
"authors":"['Binyuan Hui', 'Ruiying Geng', 'Qiyu Ren', 'Binhua Li', 'Yongbin Li', 'Jian Sun', 'Fei Huang', 'Luo Si', 'Pengfei Zhu', 'Xiaodan Zhu']",
"arxiv_primary_category":"cs.cl",
"readability":0.8,
"actuality":3.0,
"controversy":2.3333333333,
"relevance_magnitude":3.3333333333,
"relevance_valence":4.444,
"newsworthiness_crowd_sum":3.2776666667
},
{
"arxiv_id":"2101.01688v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.01688v2",
"title":"What social media told about us in the time of COVID-19: a scoping review",
"summary":"With the onset of COVID-19 pandemic, social media has rapidly become a crucial communication tool for information generation, dissemination, and consumption. In this scoping review, we selected and examined peer-reviewed empirical studies relating to COVID-19 and social media during the first outbreak starting in November 2019 until May 2020. From an analysis of 81 studies, we identified five overarching public health themes concerning the role of online social platforms and COVID-19. These themes focused on: (i) surveying public attitudes, (ii) identifying infodemics, (iii) assessing mental health, (iv) detecting or predicting COVID-19 cases, (v) analyzing government responses to the pandemic, and (vi) evaluating quality of health information in prevention education videos. Furthermore, our review highlights the paucity of studies on the application of machine learning on social media data related to COVID-19 and a lack of studies documenting real-time surveillance developed with social media data on COVID-19. For COVID-19, social media can play a crucial role in disseminating health information as well as tackling infodemics and misinformation.",
"published":"2021-01-05T00:00:00.000Z",
"authors":"['Shu-Feng Tsao', 'Helen Chen', 'Therese Tisseverasinghe', 'Yang Yang', 'Lianghua Li', 'Zahid A. Butt']",
"arxiv_primary_category":"cs.si",
"readability":0.86,
"actuality":5.0,
"controversy":3.6666666667,
"relevance_magnitude":4.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":3.9445
},
{
"arxiv_id":"2101.01993v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.01993v1",
"title":"A Survey of Deep RL and IL for Autonomous Driving Policy Learning",
"summary":"Autonomous driving (AD) agents generate driving policies based on online perception results, which are obtained at multiple levels of abstraction, e.g., behavior planning, motion planning and control. Driving policies are crucial to the realization of safe, efficient and harmonious driving behaviors, where AD agents still face substantial challenges in complex scenarios. Due to their successful application in fields such as robotics and video games, the use of deep reinforcement learning (DRL) and deep imitation learning (DIL) techniques to derive AD policies have witnessed vast research efforts in recent years. This paper is a comprehensive survey of this body of work, which is conducted at three levels: First, a taxonomy of the literature studies is constructed from the system perspective, among which five modes of integration of DRL\/DIL models into an AD architecture are identified. Second, the formulations of DRL\/DIL models for conducting specified AD tasks are comprehensively reviewed, where various designs on the model state and action spaces and the reinforcement learning rewards are covered. Finally, an in-depth review is conducted on how the critical issues of AD applications regarding driving safety, interaction with other traffic participants and uncertainty of the environment are addressed by the DRL\/DIL models. To the best of our knowledge, this is the first survey to focus on AD policy learning using DRL\/DIL, which is addressed simultaneously from the system, task-driven and problem-driven perspectives. We share and discuss findings, which may lead to the investigation of various topics in the future.",
"published":"2021-01-06T00:00:00.000Z",
"authors":"['Zeyu Zhu', 'Huijing Zhao']",
"arxiv_primary_category":"cs.ro",
"readability":0.91,
"actuality":4.3333333333,
"controversy":4.3333333333,
"relevance_magnitude":4.6666666667,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":3.8888333333
},
{
"arxiv_id":"2101.02011v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02011v1",
"title":"Towards an Abolitionist AI: the role of Historically Black Colleges and Universities",
"summary":"Abolition is the process of destroying and then rebuilding the structures that impede liberation. This paper addresses the particular case of Black folk in the United States, but is relevant to the global decolonization movement. Using notions of abolition and infrastructures of feeling developed by Ruth Wilson Gilmore, I view Historically Black Colleges and Universities ( HBCUs ) as a particular kind of abolitionist project, created for the explicit purpose of nurturing and sustaining Black excellence particularly within the sciences. I then examine how artificial intelligence (AI) in particular and computing in general have contributed to racial oppression and the further confinement and diminishing of Black existence. I conclude by examining how the space held by HBCUs in computing might contribute to a re-imagining of AI as a technology that enhances the possibility and actualization of Black life.",
"published":"2021-01-06T00:00:00.000Z",
"authors":"['Charles C. Earl']",
"arxiv_primary_category":"cs.cy",
"readability":0.87,
"actuality":4.3333333333,
"controversy":4.0,
"relevance_magnitude":4.6666666667,
"relevance_valence":4.444,
"newsworthiness_crowd_sum":4.361
},
{
"arxiv_id":"2101.02018v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02018v1",
"title":"Abusive Advertising: Scrutinizing socially relevant algorithms in a black box analysis to examine their impact on vulnerable patient groups in the health sector",
"summary":"The targeted direct-to-customer marketing of unapproved stem cell treatments by a questionable online industry is directed at vulnerable users who search the Internet in the hope of a cure. This behavior especially poses a threat to individuals who find themselves in hopeless and desperate phases in their lives. They might show low reluctance to try therapies that solely promise a cure but are not scientifically proven to do so. In the worst case, they suffer serious side-effects. Therefore, this thesis examines the display of advertisements of unapproved stem cell treatments for Parkinson's Disease, Multiple Sclerosis, Diabetes on Google's results page. The company announced a policy change in September 2019 that was meant to prohibit and ban the practices in question. However, there was evidence that those ads were still being delivered. A browser extension for Firefox and Chrome was developed and distributed to conduct a crowdsourced Black Box analysis. It was delivered to volunteers and virtual machines in Australia, Canada, the USA and the UK. Data on search results, advertisements and top stories was collected and analyzed. The results showed that there still is questionable advertising even though Google announced to purge it from its platform.",
"published":"2021-01-04T00:00:00.000Z",
"authors":"['Martin Reber']",
"arxiv_primary_category":"cs.cy",
"readability":0.92,
"actuality":3.6666666667,
"controversy":4.3333333333,
"relevance_magnitude":3.3333333333,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":3.3888333333
},
{
"arxiv_id":"2101.02026v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02026v1",
"title":"Learning about the reduction of food waste using Blockchain technology",
"summary":"Farmers need to be efficient and dedicate a lot of time in order to sustain the quality of their animals which are in their care. The most convenient and good quality - price ratio should be chosen for the feed of animals. Blockchain is used in a virtual space to store and share information over a network of users. This is done using the open source Hyperledger Fabric platform. The transactions can be viewed by all the other users in real time. These transactions are stored as JSONs inside CouchDB NoSQL database which supports queries on a large volume of data. When using this technology, the farmer can know with whom the supplier for animal feed collaborated with. The history of the transactions are not saved in just one place. In this way, it is more difficult to hack and provide implausible information. An e-learning platform was created where the farm's user can post information, respectively new blocks about the animal's birth, vaccinations, medicines, including the location of the livestock. The same e-learning platform is accessible from the mobile phone. By using the blockchain technology, anyone, including the client from the shop can know a lot about the origin of the products. Fake origins of food are much more difficult to hide. Fraud is also limited. The system monitored the traceability of dairy products inside a Romanian farm. Data about fodder provider and quality, cow productive performances and health and dairy products process were obtained and analyzed by students who will become specialists at all the levels of the food chain. Blockchain is the technology which in case of a dairy products contamination, the origin of the farm is traced in just a couple of seconds. In this way just a batch of dairy products is removed from distribution, leading to the reduction of food waste.",
"published":"2021-01-03T00:00:00.000Z",
"authors":"['Monica-Paula Marin', 'Iuliana Marin', 'Livia Vidu']",
"arxiv_primary_category":"cs.cy",
"readability":0.93,
"actuality":4.0,
"controversy":2.6666666667,
"relevance_magnitude":5.0,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.7499166667
},
{
"arxiv_id":"2101.02051v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02051v1",
"title":"Transformer-based approach towards music emotion recognition from lyrics",
"summary":"The task of identifying emotions from a given music track has been an active pursuit in the Music Information Retrieval (MIR) community for years. Music emotion recognition has typically relied on acoustic features, social tags, and other metadata to identify and classify music emotions. The role of lyrics in music emotion recognition remains under-appreciated in spite of several studies reporting superior performance of music emotion classifiers based on features extracted from lyrics. In this study, we use the transformer-based approach model using XLNet as the base architecture which, till date, has not been used to identify emotional connotations of music based on lyrics. Our proposed approach outperforms existing methods for multiple datasets. We used a robust methodology to enhance web-crawlers' accuracy for extracting lyrics. This study has important implications in improving applications involved in playlist generation of music based on emotions in addition to improving music recommendation systems.",
"published":"2021-01-06T00:00:00.000Z",
"authors":"['Yudhik Agrawal', 'Ramaguru Guru Ravi Shanker', 'Vinoo Alluri']",
"arxiv_primary_category":"cs.ir",
"readability":0.86,
"actuality":2.6666666667,
"controversy":2.0,
"relevance_magnitude":3.0,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.6111666667
},
{
"arxiv_id":"2101.02178v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02178v1",
"title":"Improving Training Result of Partially Observable Markov Decision Process by Filtering Beliefs",
"summary":"In this study I proposed a filtering beliefs method for improving performance of Partially Observable Markov Decision Processes(POMDPs), which is a method wildly used in autonomous robot and many other domains concerning control policy. My method search and compare every similar belief pair. Because a similar belief have insignificant influence on control policy, the belief is filtered out for reducing training time. The empirical results show that the proposed method outperforms the point-based approximate POMDPs in terms of the quality of training results as well as the efficiency of the method.",
"published":"2021-01-05T00:00:00.000Z",
"authors":"['Oscar LiJen Hsu']",
"arxiv_primary_category":"cs.ai",
"readability":0.88,
"actuality":2.0,
"controversy":2.0,
"relevance_magnitude":2.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.2778333333
},
{
"arxiv_id":"2101.02231v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02231v1",
"title":"Controlling Synthetic Characters in Simulations: A Case for Cognitive Architectures and Sigma",
"summary":"Simulations, along with other similar applications like virtual worlds and video games, require computational models of intelligence that generate realistic and credible behavior for the participating synthetic characters. Cognitive architectures, which are models of the fixed structure underlying intelligent behavior in both natural and artificial systems, provide a conceptually valid common basis, as evidenced by the current efforts towards a standard model of the mind, to generate human-like intelligent behavior for these synthetic characters. Sigma is a cognitive architecture and system that strives to combine what has been learned from four decades of independent work on symbolic cognitive architectures, probabilistic graphical models, and more recently neural models, under its graphical architecture hypothesis. Sigma leverages an extended form of factor graphs towards a uniform grand unification of not only traditional cognitive capabilities but also key non-cognitive aspects, creating unique opportunities for the construction of new kinds of cognitive models that possess a Theory-of-Mind and that are perceptual, autonomous, interactive, affective, and adaptive. In this paper, we will introduce Sigma along with its diverse capabilities and then use three distinct proof-of-concept Sigma models to highlight combinations of these capabilities: (1) Distributional reinforcement learning models in; (2) A pair of adaptive and interactive agent models that demonstrate rule-based, probabilistic, and social reasoning; and (3) A knowledge-free exploration model in which an agent leverages only architectural appraisal variables, namely attention and curiosity, to locate an item while building up a map in a Unity environment.",
"published":"2021-01-06T00:00:00.000Z",
"authors":"['Volkan Ustun', 'Paul S. Rosenbloom', 'Seyed Sajjadi', 'Jeremy Nuttal']",
"arxiv_primary_category":"cs.ai",
"readability":0.84,
"actuality":3.6666666667,
"controversy":2.0,
"relevance_magnitude":3.6666666667,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.3055833333
},
{
"arxiv_id":"2101.02279v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02279v1",
"title":"On State Estimation for Legged Locomotion over Soft Terrain",
"summary":"Locomotion over soft terrain remains a challenging problem for legged robots. Most of the work done on state estimation for legged robots is designed for rigid contacts, and does not take into account the physical parameters of the terrain. That said, this letter answers the following questions: how and why does soft terrain affect state estimation for legged robots? To do so, we utilized a state estimator that fuses IMU measurements with leg odometry that is designed with rigid contact assumptions. We experimentally validated the state estimator with the HyQ robot trotting over both soft and rigid terrain. We demonstrate that soft terrain negatively affects state estimation for legged robots, and that the state estimates have a noticeable drift over soft terrain compared to rigid terrain.",
"published":"2021-01-06T00:00:00.000Z",
"authors":"['Shamel Fahmi', 'Geoff Fink', 'Claudio Semini']",
"arxiv_primary_category":"cs.ro",
"readability":0.8,
"actuality":1.6666666667,
"controversy":1.6666666667,
"relevance_magnitude":2.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.1111666667
},
{
"arxiv_id":"2101.02409v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02409v1",
"title":"On the Management of Type 1 Diabetes Mellitus with IoT Devices and ML Techniques",
"summary":"The purpose of this Conference is to present the main lines of base projects that are founded on research already begun in previous years. In this sense, this manuscript will present the main lines of research in Diabetes Mellitus type 1 and Machine Learning techniques in an Internet of Things environment, so that we can summarize the future lines to be developed as follows: data collection through biosensors, massive data processing in the cloud, interconnection of biodevices, local computing vs. cloud computing, and possibilities of machine learning techniques to predict blood glucose values, including both variable selection algorithms and predictive techniques.",
"published":"2021-01-07T00:00:00.000Z",
"authors":"['Ignacio Rodriguez']",
"arxiv_primary_category":"cs.lg",
"readability":0.9,
"actuality":4.0,
"controversy":2.6666666667,
"relevance_magnitude":4.3333333333,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.72225
},
{
"arxiv_id":"2101.02434v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02434v1",
"title":"Integration of IEEE 802.1AS-based Time Synchronization in IEEE 802.11 as an Enabler for Novel Industrial Use Cases",
"summary":"Industry 4.0 introduces new use cases, with more and more mobile devices appearing in the industrial landscape. These applications require both new technologies and smooth integration into existing brownfield deployments. Emerging mobile use cases can be divided into optional mobile and mandatory mobile, where the first point considers the use of wireless communications due to soft criteria such as cost savings and the second means use cases that cannot be covered by wireline technologies due to their movement, such as AGVs. For most industrial applications, high determinism, E2E latency and synchronicity are most important. Therefore, we provide a common table, based on these requirements, listing both existing and emerging mobile use cases. Since time synchronization is particularly demanding for wireless use cases, we propose a concept for a simple but precise synchronization in IEEE 802.11 WLAN and a suitable integration using TSN in combination with OPC UA technology as examples. Furthermore, the concept is evaluated with the help of a testbed utilizing state-of-the-art hardware. This means that this concept can be directly applied in existing industry solutions. It can be shown that the concept is already suitable for a wide range of the mandatory mobile applications.",
"published":"2021-01-07T00:00:00.000Z",
"authors":"['Michael Gundall', 'Christopher Huber', 'Sergiy Melnyk']",
"arxiv_primary_category":"cs.ni",
"readability":0.89,
"actuality":3.0,
"controversy":2.6666666667,
"relevance_magnitude":3.0,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":2.7221666667
},
{
"arxiv_id":"2101.02436v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02436v1",
"title":"Feasibility Study on Virtual Process Controllers as Basis for Future Industrial Automation Systems",
"summary":"Industry 4.0 offers many possibilities for creating highly efficient and flexible manufacturing. To create such advantages, highly automated and thus digitized processes and systems are required. Here, most technologies known from the office floor are basically suitable for these tasks, but cannot meet the high demands of industrial use cases. Therefore, they cannot replace industrial technologies and devices that have performed well over decades \"out of the box\". For this reason, many technologies known from the office floor are being investigated and adapted for industrial environments. An important task is the virtualization of process controls, as more and more devices use computation offloading, e.g. due to limited resources. In this paper we extend the work on our novel architecture that enables numerous use cases and meets industrial requirements by virtualizing process controllers. In addition, a testbed based on a factory scenario is proposed to evaluate the most important features of the presented architecture.",
"published":"2021-01-07T00:00:00.000Z",
"authors":"['Michael Gundall', 'Calvin Glas', 'Hans D. Schotten']",
"arxiv_primary_category":"cs.ni",
"readability":0.93,
"actuality":3.3333333333,
"controversy":4.0,
"relevance_magnitude":4.3333333333,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":3.4721666667
},
{
"arxiv_id":"2101.02483v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02483v1",
"title":"Robust Text CAPTCHAs Using Adversarial Examples",
"summary":"CAPTCHA (Completely Automated Public Truing test to tell Computers and Humans Apart) is a widely used technology to distinguish real users and automated users such as bots. However, the advance of AI technologies weakens many CAPTCHA tests and can induce security concerns. In this paper, we propose a user-friendly text-based CAPTCHA generation method named Robust Text CAPTCHA (RTC). At the first stage, the foregrounds and backgrounds are constructed with randomly sampled font and background images, which are then synthesized into identifiable pseudo adversarial CAPTCHAs. At the second stage, we design and apply a highly transferable adversarial attack for text CAPTCHAs to better obstruct CAPTCHA solvers. Our experiments cover comprehensive models including shallow models such as KNN, SVM and random forest, various deep neural networks and OCR models. Experiments show that our CAPTCHAs have a failure rate lower than one millionth in general and high usability. They are also robust against various defensive techniques that attackers may employ, including adversarial training, data pre-processing and manual tagging.",
"published":"2021-01-07T00:00:00.000Z",
"authors":"['Rulin Shao', 'Zhouxing Shi', 'Jinfeng Yi', 'Pin-Yu Chen', 'Cho-Jui Hsieh']",
"arxiv_primary_category":"cs.lg",
"readability":0.82,
"actuality":3.6666666667,
"controversy":1.0,
"relevance_magnitude":4.3333333333,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.08325
},
{
"arxiv_id":"2101.02648v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02648v2",
"title":"Argument Schemes and Dialogue for Explainable Planning",
"summary":"Artificial Intelligence (AI) is being increasingly deployed in practical applications. However, there is a major concern whether AI systems will be trusted by humans. In order to establish trust in AI systems, there is a need for users to understand the reasoning behind their solutions. Therefore, systems should be able to explain and justify their output. In this paper, we propose an argument scheme-based approach to provide explanations in the domain of AI planning. We present novel argument schemes to create arguments that explain a plan and its key elements; and a set of critical questions that allow interaction between the arguments and enable the user to obtain further information regarding the key elements of the plan. Furthermore, we present a novel dialogue system using the argument schemes and critical questions for providing interactive dialectical explanations.",
"published":"2021-01-07T00:00:00.000Z",
"authors":"['Quratul-ain Mahesar', 'Simon Parsons']",
"arxiv_primary_category":"cs.ai",
"readability":0.96,
"actuality":4.3333333333,
"controversy":3.6666666667,
"relevance_magnitude":4.0,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.97225
},
{
"arxiv_id":"2101.02722v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02722v1",
"title":"The Distracting Control Suite -- A Challenging Benchmark for Reinforcement Learning from Pixels",
"summary":"Robots have to face challenging perceptual settings, including changes in viewpoint, lighting, and background. Current simulated reinforcement learning (RL) benchmarks such as DM Control provide visual input without such complexity, which limits the transfer of well-performing methods to the real world. In this paper, we extend DM Control with three kinds of visual distractions (variations in background, color, and camera pose) to produce a new challenging benchmark for vision-based control, and we analyze state of the art RL algorithms in these settings. Our experiments show that current RL methods for vision-based control perform poorly under distractions, and that their performance decreases with increasing distraction complexity, showing that new methods are needed to cope with the visual complexities of the real world. We also find that combinations of multiple distraction types are more difficult than a mere combination of their individual effects.",
"published":"2021-01-07T00:00:00.000Z",
"authors":"['Austin Stone', 'Oscar Ramirez', 'Kurt Konolige', 'Rico Jonschkowski']",
"arxiv_primary_category":"cs.ro",
"readability":0.86,
"actuality":4.0,
"controversy":3.0,
"relevance_magnitude":4.0,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.58325
},
{
"arxiv_id":"2101.02988v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.02988v1",
"title":"Graph embeddings for Abusive Language Detection",
"summary":"Abusive behaviors are common on online social networks. The increasing frequency of antisocial behaviors forces the hosts of online platforms to find new solutions to address this problem. Automating the moderation process has thus received a lot of interest in the past few years. Various methods have been proposed, most based on the exchanged content, and one relying on the structure and dynamics of the conversation. It has the advantage of being languageindependent, however it leverages a hand-crafted set of topological measures which are computationally expensive and not necessarily suitable to all situations. In the present paper, we propose to use recent graph embedding approaches to automatically learn representations of conversational graphs depicting message exchanges. We compare two categories: node vs. whole-graph embeddings. We experiment with a total of 8 approaches and apply them to a dataset of online messages. We also study more precisely which aspects of the graph structure are leveraged by each approach. Our study shows that the representation produced by certain embeddings captures the information conveyed by specific topological measures, but misses out other aspects.",
"published":"2021-01-08T00:00:00.000Z",
"authors":"['No\u00e9 Cecillon', 'Vincent Labatut', 'Richard Dufour', 'Georges Linares']",
"arxiv_primary_category":"cs.si",
"readability":0.86,
"actuality":4.6666666667,
"controversy":3.6666666667,
"relevance_magnitude":3.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":3.6111666667
},
{
"arxiv_id":"2101.03002v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03002v1",
"title":"Studying Leaders During Times of Crisis Using Online Social Media -- A COVID Case Study",
"summary":"Online Social media (OSM) has become a primary platform for discussion on diverse topics. Even famous and public figures often express their views on various topics through OSM platforms. Novel Coronavirus officially called COVID-19, which has become a pandemic and has created a crisis in human history, is one such topic that has attracted a lot of attention on Twitter in recent times. In this work, we analyze 29 million tweets spanning across three months, to study highly influential users, which we called as leaders. We identify these leaders using social network analysis and analyze their tweets using text analysis techniques. We group these leaders in four clusters, namely research, news, health and politics. Our analysis shows that i) all clusters show a similar amount of fear in their tweets, ii) researchers and news cluster display more sadness compared to others and, iii) health organizations and politicians try to gain public trust. The text analysis shows that the researchers are more concerned about understanding symptoms and developing vaccination; news and politicians are mainly discussing about travel and hygiene; and health organizations focuses on hygiene. Our descriptive analysis helps us to extract various features that we used to classify tweets among the four clusters with an accuracy of 96% AUC ROC.",
"published":"2021-01-08T00:00:00.000Z",
"authors":"['Rahul Goel', 'Rajesh Sharma']",
"arxiv_primary_category":"cs.si",
"readability":0.92,
"actuality":4.3333333333,
"controversy":2.6666666667,
"relevance_magnitude":4.0,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":3.4445
},
{
"arxiv_id":"2101.03072v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03072v1",
"title":"A Primer on HIBS -- High Altitude Platform Stations as IMT Base Stations",
"summary":"Mobile communication via high-altitude platforms operating in the stratosphere is an idea that has been on the table for decades. In the past few years, however, with recent advances in technology and parallel progress in standardization and regulatory bodies like 3GPP and ITU, these ideas have gained considerable momentum. In this article, we present a comprehensive overview of HIBS - High Altitude Platform Stations as IMT Base Stations. We lay out possible use cases and summarize the current status of the development, from a technological point of view as well as from standardization in 3GPP, and regarding spectrum aspects. We then present preliminary system level simulation results to shed light on the performance of HIBS. We conclude with pointing out several directions for future research.",
"published":"2021-01-08T00:00:00.000Z",
"authors":"['Sebastian Euler', 'Xingqin Lin', 'Erika Tejedor', 'Evanny Obregon']",
"arxiv_primary_category":"cs.ni",
"readability":0.89,
"actuality":4.0,
"controversy":3.6666666667,
"relevance_magnitude":4.6666666667,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":4.0555833333
},
{
"arxiv_id":"2101.03270v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03270v1",
"title":"Investigation by Driving Simulation of Tractor Overturning Accidents Caused by Steering Instability",
"summary":"Overturning tractors are the leading cause of fatalities on farms. Steering instability contributes significantly to the tractor overturning. This study investigated tractor overturning accidents caused by the steering instability using a driving simulator. The general commercial driving simulator CarSim (Mechanical Simulation Cooperation, MI, USA) was used. Tractor operations on steep passage slopes were simulated to mimic conditions present for a real accident case reported in Japan. Simulations were performed on roads with and without slopes. The tractor overturned only when on the road with the steep slope. The decrease in the vertical force on the front wheel caused the steering instability and the tractor to overturn. The steering instability caused understeer which prevents the operator from being able to control the tractor properly. Subsequently, the tractor overturned in the simulation. The tractor driving simulator was capable of reproducing the steering instability which can lead to the overturning accident.",
"published":"2021-01-09T00:00:00.000Z",
"authors":"['Masahisa Watanabe', 'Kenshi Sakai']",
"arxiv_primary_category":"cs.ro",
"readability":0.81,
"actuality":2.6666666667,
"controversy":2.0,
"relevance_magnitude":3.6666666667,
"relevance_valence":4.444,
"newsworthiness_crowd_sum":3.1943333333
},
{
"arxiv_id":"2101.03327v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03327v1",
"title":"Selection of Optimal Parameters in the Fast K-Word Proximity Search Based on Multi-component Key Indexes",
"summary":"Proximity full-text search is commonly implemented in contemporary full-text search systems. Let us assume that the search query is a list of words. It is natural to consider a document as relevant if the queried words are near each other in the document. The proximity factor is even more significant for the case where the query consists of frequently occurring words. Proximity full-text search requires the storage of information for every occurrence in documents of every word that the user can search. For every occurrence of every word in a document, we employ additional indexes to store information about nearby words, that is, the words that occur in the document at distances from the given word of less than or equal to the MaxDistance parameter. We showed in previous works that these indexes can be used to improve the average query execution time by up to 130 times for queries that consist of words occurring with high-frequency. In this paper, we consider how both the search performance and the search quality depend on the value of MaxDistance and other parameters. Well-known GOV2 text collection is used in the experiments for reproducibility of the results. We propose a new index schema after the analysis of the results of the experiments. This is a pre-print of a contribution published in Supplementary Proceedings of the XXII International Conference on Data Analytics and Management in Data Intensive Domains (DAMDID\/RCDL 2020), Voronezh, Russia, October 13-16, 2020, P. 336-350, published by CEUR Workshop Proceedings. The final authenticated version is available online at: http:\/\/ceur-ws.org\/Vol-2790\/",
"published":"2021-01-09T00:00:00.000Z",
"authors":"['Alexander B. Veretennikov']",
"arxiv_primary_category":"cs.ir",
"readability":0.88,
"actuality":3.3333333333,
"controversy":1.0,
"relevance_magnitude":4.0,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.0555833333
},
{
"arxiv_id":"2101.03441v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03441v2",
"title":"Rate Allocation and Content Placement in Cache Networks",
"summary":"We introduce the problem of optimal congestion control in cache networks, whereby \\emph{both} rate allocations and content placements are optimized \\emph{jointly}. We formulate this as a maximization problem with non-convex constraints, and propose solving this problem via (a) a Lagrangian barrier algorithm and (b) a convex relaxation. We prove different optimality guarantees for each of these two algorithms; our proofs exploit the fact that the non-convex constraints of our problem involve DR-submodular functions.",
"published":"2021-01-09T00:00:00.000Z",
"authors":"['Khashayar Kamran', 'Armin Moharrer', 'Stratis Ioannidis', 'Edmund Yeh']",
"arxiv_primary_category":"cs.ni",
"readability":0.76,
"actuality":2.6666666667,
"controversy":1.3333333333,
"relevance_magnitude":3.6666666667,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":2.8889166667
},
{
"arxiv_id":"2101.03464v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03464v1",
"title":"SPAGAN: Shortest Path Graph Attention Network",
"summary":"Graph convolutional networks (GCN) have recently demonstrated their potential in analyzing non-grid structure data that can be represented as graphs. The core idea is to encode the local topology of a graph, via convolutions, into the feature of a center node. In this paper, we propose a novel GCN model, which we term as Shortest Path Graph Attention Network (SPAGAN). Unlike conventional GCN models that carry out node-based attentions within each layer, the proposed SPAGAN conducts path-based attention that explicitly accounts for the influence of a sequence of nodes yielding the minimum cost, or shortest path, between the center node and its higher-order neighbors. SPAGAN therefore allows for a more informative and intact exploration of the graph structure and further {a} more effective aggregation of information from distant neighbors into the center node, as compared to node-based GCN methods. We test SPAGAN on the downstream classification task on several standard datasets, and achieve performances superior to the state of the art. Code is publicly available at https:\/\/github.com\/ihollywhy\/SPAGAN.",
"published":"2021-01-10T00:00:00.000Z",
"authors":"['Yiding Yang', 'Xinchao Wang', 'Mingli Song', 'Junsong Yuan', 'Dacheng Tao']",
"arxiv_primary_category":"cs.lg",
"readability":0.8,
"actuality":3.0,
"controversy":3.6666666667,
"relevance_magnitude":3.6666666667,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.5555833333
},
{
"arxiv_id":"2101.03478v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03478v1",
"title":"Activity Recognition with Moving Cameras and Few Training Examples: Applications for Detection of Autism-Related Headbanging",
"summary":"Activity recognition computer vision algorithms can be used to detect the presence of autism-related behaviors, including what are termed \"restricted and repetitive behaviors\", or stimming, by diagnostic instruments. The limited data that exist in this domain are usually recorded with a handheld camera which can be shaky or even moving, posing a challenge for traditional feature representation approaches for activity detection which mistakenly capture the camera's motion as a feature. To address these issues, we first document the advantages and limitations of current feature representation techniques for activity recognition when applied to head banging detection. We then propose a feature representation consisting exclusively of head pose keypoints. We create a computer vision classifier for detecting head banging in home videos using a time-distributed convolutional neural network (CNN) in which a single CNN extracts features from each frame in the input sequence, and these extracted features are fed as input to a long short-term memory (LSTM) network. On the binary task of predicting head banging and no head banging within videos from the Self Stimulatory Behaviour Dataset (SSBD), we reach a mean F1-score of 90.77% using 3-fold cross validation (with individual fold F1-scores of 83.3%, 89.0%, and 100.0%) when ensuring that no child who appeared in the train set was in the test set for all folds. This work documents a successful technique for training a computer vision classifier which can detect human motion with few training examples and even when the camera recording the source clips is unstable. The general methods described here can be applied by designers and developers of interactive systems towards other human motion and pose classification problems used in mobile and ubiquitous interactive systems.",
"published":"2021-01-10T00:00:00.000Z",
"authors":"['Peter Washington', 'Aaron Kline', 'Onur Cezmi Mutlu', 'Emilie Leblanc', 'Cathy Hou', 'Nate Stockham', 'Kelley Paskov', 'Brianna Chrisman', 'Dennis P. Wall']",
"arxiv_primary_category":"cs.cv",
"readability":0.88,
"actuality":4.6666666667,
"controversy":3.0,
"relevance_magnitude":4.0,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.7499166667
},
{
"arxiv_id":"2101.03641v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03641v2",
"title":"Learning Augmented Index Policy for Optimal Service Placement at the Network Edge",
"summary":"We consider the problem of service placement at the network edge, in which a decision maker has to choose between $N$ services to host at the edge to satisfy the demands of customers. Our goal is to design adaptive algorithms to minimize the average service delivery latency for customers. We pose the problem as a Markov decision process (MDP) in which the system state is given by describing, for each service, the number of customers that are currently waiting at the edge to obtain the service. However, solving this $N$-services MDP is computationally expensive due to the curse of dimensionality. To overcome this challenge, we show that the optimal policy for a single-service MDP has an appealing threshold structure, and derive explicitly the Whittle indices for each service as a function of the number of requests from customers based on the theory of Whittle index policy. Since request arrival and service delivery rates are usually unknown and possibly time-varying, we then develop efficient learning augmented algorithms that fully utilize the structure of optimal policies with a low learning regret. The first of these is UCB-Whittle, and relies upon the principle of optimism in the face of uncertainty. The second algorithm, Q-learning-Whittle, utilizes Q-learning iterations for each service by using a two time scale stochastic approximation. We characterize the non-asymptotic performance of UCB-Whittle by analyzing its learning regret, and also analyze the convergence properties of Q-learning-Whittle. Simulation results show that the proposed policies yield excellent empirical performance.",
"published":"2021-01-10T00:00:00.000Z",
"authors":"['Guojun Xiong', 'Rahul Singh', 'Jian Li']",
"arxiv_primary_category":"cs.ni",
"readability":0.87,
"actuality":3.3333333333,
"controversy":2.0,
"relevance_magnitude":4.0,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.1665833333
},
{
"arxiv_id":"2101.03654v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03654v1",
"title":"Disentangled Self-Attentive Neural Networks for Click-Through Rate Prediction",
"summary":"Click-through rate (CTR) prediction, which aims to predict the probability that whether of a user will click on an item, is an essential task for many online applications. Due to the nature of data sparsity and high dimensionality in CTR prediction, a key to making effective prediction is to model high-order feature interactions among feature fields. To explicitly model high-order feature interactions, an efficient way is to stack multihead self-attentive neural networks, which has achieved promising performance. However, one problem of the vanilla self-attentive network is that two terms, a whitened pairwise interaction term and a unary term, are coupled in the computation of the self-attention score, where the pairwise term contributes to learning the importance score for each feature interaction, while the unary term models the impact of one feature on all other features. We identify two factors, coupled gradient computation and shared transformations, impede the learning of both terms. To solve this problem, in this paper,we present a novel Disentangled Self-Attentive neural Network (DSAN) model for CTR prediction, which disentangles the two terms for facilitating learning feature interactions. We conduct extensive experiments framework using two real-world benchmark datasets. The results show that DSAN not only retains computational efficiency but obtains performance improvements over state-of-the-art baselines.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['Yanqiao Zhu', 'Yichen Xu', 'Feng Yu', 'Qiang Liu', 'Shu Wu', 'Liang Wang']",
"arxiv_primary_category":"cs.ir",
"readability":0.85,
"actuality":4.3333333333,
"controversy":3.6666666667,
"relevance_magnitude":4.6666666667,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":3.7221666667
},
{
"arxiv_id":"2101.03655v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03655v1",
"title":"Machine Learning Towards Intelligent Systems: Applications, Challenges, and Opportunities",
"summary":"The emergence and continued reliance on the Internet and related technologies has resulted in the generation of large amounts of data that can be made available for analyses. However, humans do not possess the cognitive capabilities to understand such large amounts of data. Machine learning (ML) provides a mechanism for humans to process large amounts of data, gain insights about the behavior of the data, and make more informed decision based on the resulting analysis. ML has applications in various fields. This review focuses on some of the fields and applications such as education, healthcare, network security, banking and finance, and social media. Within these fields, there are multiple unique challenges that exist. However, ML can provide solutions to these challenges, as well as create further research opportunities. Accordingly, this work surveys some of the challenges facing the aforementioned fields and presents some of the previous literature works that tackled them. Moreover, it suggests several research opportunities that benefit from the use of ML to address these challenges.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['MohammadNoor Injadat', 'Abdallah Moubayed', 'Ali Bou Nassif', 'Abdallah Shami']",
"arxiv_primary_category":"cs.lg",
"readability":0.95,
"actuality":3.0,
"controversy":3.3333333333,
"relevance_magnitude":5.0,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.8055833333
},
{
"arxiv_id":"2101.03693v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03693v1",
"title":"Exploiting a Fleet of UAVs for Monitoring and Data Acquisition of a Distributed Sensor Network",
"summary":"This study proposes an efficient data collection strategy exploiting a team of Unmanned Aerial Vehicles (UAVs) to monitor and collect the data of a large distributed sensor network usually used for environmental monitoring, meteorology, agriculture, and renewable energy applications. The study develops a collaborative mission planning system that enables a team of UAVs to conduct and complete the mission of sensors' data collection collaboratively while considering existing constraints of the UAV payload and battery capacity. The proposed mission planner system employs the Differential Evolution (DE) optimization algorithm enabling UAVs to maximize the number of visited sensor nodes given the priority of the sensors and avoiding the redundant collection of sensors' data. The proposed mission planner is evaluated through extensive simulation and comparative analysis. The simulation results confirm the effectiveness and fidelity of the proposed mission planner to be used for the distributed sensor network monitoring and data collection.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['S. MahmoudZadeh', 'A. Yazdani', 'A. Elmi', 'A. Abbasi', 'P. Ghanooni']",
"arxiv_primary_category":"cs.ro",
"readability":0.85,
"actuality":3.3333333333,
"controversy":2.0,
"relevance_magnitude":3.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.8611666667
},
{
"arxiv_id":"2101.03700v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03700v2",
"title":"AT-BERT: Adversarial Training BERT for Acronym Identification Winning Solution for SDU@AAAI-21",
"summary":"Acronym identification focuses on finding the acronyms and the phrases that have been abbreviated, which is crucial for scientific document understanding tasks. However, the limited size of manually annotated datasets hinders further improvement for the problem. Recent breakthroughs of language models pre-trained on large corpora clearly show that unsupervised pre-training can vastly improve the performance of downstream tasks. In this paper, we present an Adversarial Training BERT method named AT-BERT, our winning solution to acronym identification task for Scientific Document Understanding (SDU) Challenge of AAAI 2021. Specifically, the pre-trained BERT is adopted to capture better semantic representation. Then we incorporate the FGM adversarial training strategy into the fine-tuning of BERT, which makes the model more robust and generalized. Furthermore, an ensemble mechanism is devised to involve the representations learned from multiple BERT variants. Assembling all these components together, the experimental results on the SciAI dataset show that our proposed approach outperforms all other competitive state-of-the-art methods.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['Danqing Zhu', 'Wangli Lin', 'Yang Zhang', 'Qiwei Zhong', 'Guanxiong Zeng', 'Weilin Wu', 'Jiayu Tang']",
"arxiv_primary_category":"cs.cl",
"readability":0.83,
"actuality":1.0,
"controversy":1.0,
"relevance_magnitude":2.6666666667,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":1.7221666667
},
{
"arxiv_id":"2101.03757v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03757v2",
"title":"VaccinItaly: monitoring Italian conversations around vaccines on Twitter",
"summary":"We monitor online conversations of Italian users around vaccines on Twitter, and we provide public access to the on-going data collection which will run continuously throughout the vaccination campaign taking place in Italy. We started collecting tweets matching vaccine-related keywords (in Italian) on December 20th 2020 using Twitter APIs, capturing the Italian vaccine rollout (27th December 2020), and at the time of this writing (13th January 2020) we collected over 1.8 M tweets, with an average number of 30k tweets shared on a daily basis. We report a consistent amount of low-credibility information already circulating on Twitter alongside vaccine-related conversations, whose prevalence is smaller yet comparable to high-credibility information. We believe that our data will allow researchers to understand the interplay between public discourse on online social media and the evolution of the on-going vaccination campaign against SARS-CoV-2 taking place in Italy.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['Francesco Pierri', 'Silvio Pavanetto', 'Marco Brambilla', 'Stefano Ceri']",
"arxiv_primary_category":"cs.si",
"readability":0.91,
"actuality":5.0,
"controversy":4.6666666667,
"relevance_magnitude":5.0,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":4.2221666667
},
{
"arxiv_id":"2101.03769v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03769v1",
"title":"A Review of Evaluation Practices of Gesture Generation in Embodied Conversational Agents",
"summary":"Embodied Conversational Agents (ECA) take on different forms, including virtual avatars or physical agents, such as a humanoid robot. ECAs are often designed to produce nonverbal behaviour to complement or enhance its verbal communication. One form of nonverbal behaviour is co-speech gesturing, which involves movements that the agent makes with its arms and hands that is paired with verbal communication. Co-speech gestures for ECAs can be created using different generation methods, such as rule-based and data-driven processes. However, reports on gesture generation methods use a variety of evaluation measures, which hinders comparison. To address this, we conducted a systematic review on co-speech gesture generation methods for iconic, metaphoric, deictic or beat gestures, including their evaluation methods. We reviewed 22 studies that had an ECA with a human-like upper body that used co-speech gesturing in a social human-agent interaction, including a user study to evaluate its performance. We found most studies used a within-subject design and relied on a form of subjective evaluation, but lacked a systematic approach. Overall, methodological quality was low-to-moderate and few systematic conclusions could be drawn. We argue that the field requires rigorous and uniform tools for the evaluation of co-speech gesture systems. We have proposed recommendations for future empirical evaluation, including standardised phrases and test scenarios to test generative models. We have proposed a research checklist that can be used to report relevant information for the evaluation of generative models as well as to evaluate co-speech gesture use.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['Pieter Wolfert', 'Nicole Robinson', 'Tony Belpaeme']",
"arxiv_primary_category":"cs.hc",
"readability":0.87,
"actuality":3.6666666667,
"controversy":3.6666666667,
"relevance_magnitude":3.6666666667,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":3.4445
},
{
"arxiv_id":"2101.03785v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03785v1",
"title":"Predictive Analysis of Chikungunya",
"summary":"Chikungunya is an emerging threat for health security all over the world which is spreading very fast. Researches for proper forecasting of the incidence rate of chikungunya has been going on in many places in which DARPA has done a very extensive summarized result from 2014 to 2017 with the data of suspected cases, confirmed cases, deaths, population and incidence rate in different countries. In this project, we have analysed the dataset from DARPA and extended it to predict the incidence rate using different features of weather like temperature, humidity, dewiness, wind and pressure along with the latitude and longitude of every country. We had to use different APIs to find out these extra features from 2014-2016. After creating a pure dataset, we have used Linear Regression to predict the incidence rate and calculated the accuracy and error rate.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['Sayed Erfan Arefin', 'Tasnia Ashrafi Heya', 'Dr Moinul Zaber']",
"arxiv_primary_category":"cs.lg",
"readability":0.89,
"actuality":4.6666666667,
"controversy":2.6666666667,
"relevance_magnitude":4.6666666667,
"relevance_valence":4.444,
"newsworthiness_crowd_sum":4.111
},
{
"arxiv_id":"2101.03787v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.03787v1",
"title":"WiCV 2020: The Seventh Women In Computer Vision Workshop",
"summary":"In this paper we present the details of Women in Computer Vision Workshop - WiCV 2020, organized in alongside virtual CVPR 2020. This event aims at encouraging the women researchers in the field of computer vision. It provides a voice to a minority (female) group in computer vision community and focuses on increasingly the visibility of these researchers, both in academia and industry. WiCV believes that such an event can play an important role in lowering the gender imbalance in the field of computer vision. WiCV is organized each year where it provides a.) opportunity for collaboration with between researchers b.) mentorship to female junior researchers c.) financial support to presenters to overcome monetary burden and d.) large and diverse choice of role models, who can serve as examples to younger researchers at the beginning of their careers. In this paper, we present a report on the workshop program, trends over the past years, a summary of statistics regarding presenters, attendees, and sponsorship for the current workshop.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['Hazel Doughty', 'Nour Karessli', 'Kathryn Leonard', 'Boyi Li', 'Carianne Martinez', 'Azadeh Mobasher', 'Arsha Nagrani', 'Srishti Yadav']",
"arxiv_primary_category":"cs.cv",
"readability":0.93,
"actuality":5.0,
"controversy":2.0,
"relevance_magnitude":4.0,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.72225
},
{
"arxiv_id":"2101.04167v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04167v1",
"title":"First-Order Problem Solving through Neural MCTS based Reinforcement Learning",
"summary":"The formal semantics of an interpreted first-order logic (FOL) statement can be given in Tarskian Semantics or a basically equivalent Game Semantics. The latter maps the statement and the interpretation into a two-player semantic game. Many combinatorial problems can be described using interpreted FOL statements and can be mapped into a semantic game. Therefore, learning to play a semantic game perfectly leads to the solution of a specific instance of a combinatorial problem. We adapt the AlphaZero algorithm so that it becomes better at learning to play semantic games that have different characteristics than Go and Chess. We propose a general framework, Persephone, to map the FOL description of a combinatorial problem to a semantic game so that it can be solved through a neural MCTS based reinforcement learning algorithm. Our goal for Persephone is to make it tabula-rasa, mapping a problem stated in interpreted FOL to a solution without human intervention.",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['Ruiyang Xu', 'Prashank Kadam', 'Karl Lieberherr']",
"arxiv_primary_category":"cs.ai",
"readability":0.8,
"actuality":2.6666666667,
"controversy":2.3333333333,
"relevance_magnitude":2.6666666667,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":2.7499166667
},
{
"arxiv_id":"2101.04237v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04237v1",
"title":"Solving Common-Payoff Games with Approximate Policy Iteration",
"summary":"For artificially intelligent learning systems to have widespread applicability in real-world settings, it is important that they be able to operate decentrally. Unfortunately, decentralized control is difficult -- computing even an epsilon-optimal joint policy is a NEXP complete problem. Nevertheless, a recently rediscovered insight -- that a team of agents can coordinate via common knowledge -- has given rise to algorithms capable of finding optimal joint policies in small common-payoff games. The Bayesian action decoder (BAD) leverages this insight and deep reinforcement learning to scale to games as large as two-player Hanabi. However, the approximations it uses to do so prevent it from discovering optimal joint policies even in games small enough to brute force optimal solutions. This work proposes CAPI, a novel algorithm which, like BAD, combines common knowledge with deep reinforcement learning. However, unlike BAD, CAPI prioritizes the propensity to discover optimal joint policies over scalability. While this choice precludes CAPI from scaling to games as large as Hanabi, empirical results demonstrate that, on the games to which CAPI does scale, it is capable of discovering optimal joint policies even when other modern multi-agent reinforcement learning algorithms are unable to do so. Code is available at https:\/\/github.com\/ssokota\/capi .",
"published":"2021-01-11T00:00:00.000Z",
"authors":"['Samuel Sokota', 'Edward Lockhart', 'Finbarr Timbers', 'Elnaz Davoodi', \"Ryan D'Orazio\", 'Neil Burch', 'Martin Schmid', 'Michael Bowling', 'Marc Lanctot']",
"arxiv_primary_category":"cs.ai",
"readability":0.82,
"actuality":2.6666666667,
"controversy":3.3333333333,
"relevance_magnitude":3.6666666667,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":3.1111666667
},
{
"arxiv_id":"2101.04262v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04262v1",
"title":"Clutter Slices Approach for Identification-on-the-fly of Indoor Spaces",
"summary":"Construction spaces are constantly evolving, dynamic environments in need of continuous surveying, inspection, and assessment. Traditional manual inspection of such spaces proves to be an arduous and time-consuming activity. Automation using robotic agents can be an effective solution. Robots, with perception capabilities can autonomously classify and survey indoor construction spaces. In this paper, we present a novel identification-on-the-fly approach for coarse classification of indoor spaces using the unique signature of clutter. Using the context granted by clutter, we recognize common indoor spaces such as corridors, staircases, shared spaces, and restrooms. The proposed clutter slices pipeline achieves a maximum accuracy of 93.6% on the presented clutter slices dataset. This sensor independent approach can be generalized to various domains to equip intelligent autonomous agents in better perceiving their environment.",
"published":"2021-01-12T00:00:00.000Z",
"authors":"['Upinder Kaur', 'Praveen Abbaraju', 'Harrison McCarty', 'Richard M. Voyles']",
"arxiv_primary_category":"cs.ro",
"readability":0.81,
"actuality":3.6666666667,
"controversy":3.3333333333,
"relevance_magnitude":4.3333333333,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":3.3888333333
},
{
"arxiv_id":"2101.04283v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04283v1",
"title":"A Brief Survey of Associations Between Meta-Learning and General AI",
"summary":"This paper briefly reviews the history of meta-learning and describes its contribution to general AI. Meta-learning improves model generalization capacity and devises general algorithms applicable to both in-distribution and out-of-distribution tasks potentially. General AI replaces task-specific models with general algorithmic systems introducing higher level of automation in solving diverse tasks using AI. We summarize main contributions of meta-learning to the developments in general AI, including memory module, meta-learner, coevolution, curiosity, forgetting and AI-generating algorithm. We present connections between meta-learning and general AI and discuss how meta-learning can be used to formulate general AI algorithms.",
"published":"2021-01-12T00:00:00.000Z",
"authors":"['Huimin Peng']",
"arxiv_primary_category":"cs.ai",
"readability":0.84,
"actuality":4.6666666667,
"controversy":4.3333333333,
"relevance_magnitude":3.0,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.83325
},
{
"arxiv_id":"2101.04449v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04449v1",
"title":"Streaming VR Games to the Broad Audience: A Comparison of the First-Person and Third-Person Perspectives",
"summary":"The spectatorship experience for virtual reality (VR) games differs strongly from its non-VR precursor. When watching non-VR games on platforms such as Twitch, spectators just see what the player sees, as the physical interaction is mostly unimportant for the overall impression. In VR, the immersive full-body interaction is a crucial part of the player experience. Hence, content creators, such as streamers, often rely on green screens or similar solutions to offer a mixed-reality third-person view to disclose their full-body actions. Our work compares the most popular realizations of the first-person and the third-person perspective in an online survey (N=217) with three different VR games. Contrary to the current trend to stream in third-person, our key result is that most viewers prefer the first-person version, which they attribute mostly to the better focus on in-game actions and higher involvement. Based on the study insights, we provide design recommendations for both perspectives.",
"published":"2021-01-12T00:00:00.000Z",
"authors":"['Katharina Emmerich', 'Andrey Krekhov', 'Sebastian Cmentowski', 'Jens Krueger']",
"arxiv_primary_category":"cs.hc",
"readability":0.92,
"actuality":4.6666666667,
"controversy":3.6666666667,
"relevance_magnitude":3.6666666667,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.83325
},
{
"arxiv_id":"2101.04540v4",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04540v4",
"title":"Capturing social media expressions during the COVID-19 pandemic in Argentina and forecasting mental health and emotions",
"summary":"Purpose. We present an approach for forecasting mental health conditions and emotions of a given population during the COVID-19 pandemic in Argentina based on language expressions used in social media. This approach permits anticipating high prevalence periods in short- to medium-term time horizons. Design. Mental health conditions and emotions are captured via markers, which link social media contents with lexicons. First, we build descriptive timelines for decision makers to monitor the evolution of markers, and their correlation with crisis events. Second, we model the timelines as time series, and support their forecasting, which in turn serve to identify high prevalence points for the estimated markers. Findings. Results showed that different time series forecasting strategies offer different capabilities. In the best scenario, the emergence of high prevalence periods of emotions and mental health disorders can be satisfactorily predicted with a neural network strategy, even when limited data is available in early stages of a crisis (e.g., 7 days). Originality. Although there have been efforts in the literature to predict mental states of individuals, the analysis of mental health at the collective level has received scarce attention. We take a step forward by proposing a forecasting approach for analyzing the mental health of a given population (or group of individuals) at a larger scale. Practical implications. We believe that this work contributes to a better understanding of how psychological processes related to crisis manifest in social media, being a valuable asset for the design, implementation and monitoring of health prevention and communication policies.",
"published":"2021-01-12T00:00:00.000Z",
"authors":"['Antonela Tommasel', 'Andres Diaz-Pace', 'Juan Manuel Rodriguez', 'Daniela Godoy']",
"arxiv_primary_category":"cs.cy",
"readability":0.92,
"actuality":4.6666666667,
"controversy":3.0,
"relevance_magnitude":4.3333333333,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.83325
},
{
"arxiv_id":"2101.04547v3",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04547v3",
"title":"Of Non-Linearity and Commutativity in BERT",
"summary":"In this work we provide new insights into the transformer architecture, and in particular, its best-known variant, BERT. First, we propose a method to measure the degree of non-linearity of different elements of transformers. Next, we focus our investigation on the feed-forward networks (FFN) inside transformers, which contain 2\/3 of the model parameters and have so far not received much attention. We find that FFNs are an inefficient yet important architectural element and that they cannot simply be replaced by attention blocks without a degradation in performance. Moreover, we study the interactions between layers in BERT and show that, while the layers exhibit some hierarchical structure, they extract features in a fuzzy manner. Our results suggest that BERT has an inductive bias towards layer commutativity, which we find is mainly due to the skip connections. This provides a justification for the strong performance of recurrent and weight-shared transformer models.",
"published":"2021-01-12T00:00:00.000Z",
"authors":"['Sumu Zhao', 'Damian Pascual', 'Gino Brunner', 'Roger Wattenhofer']",
"arxiv_primary_category":"cs.cl",
"readability":0.86,
"actuality":1.3333333333,
"controversy":1.3333333333,
"relevance_magnitude":3.0,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":2.2499166667
},
{
"arxiv_id":"2101.04719v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04719v1",
"title":"Expanding Explainability: Towards Social Transparency in AI systems",
"summary":"As AI-powered systems increasingly mediate consequential decision-making, their explainability is critical for end-users to take informed and accountable actions. Explanations in human-human interactions are socially-situated. AI systems are often socio-organizationally embedded. However, Explainable AI (XAI) approaches have been predominantly algorithm-centered. We take a developmental step towards socially-situated XAI by introducing and exploring Social Transparency (ST), a sociotechnically informed perspective that incorporates the socio-organizational context into explaining AI-mediated decision-making. To explore ST conceptually, we conducted interviews with 29 AI users and practitioners grounded in a speculative design scenario. We suggested constitutive design elements of ST and developed a conceptual framework to unpack ST's effect and implications at the technical, decision-making, and organizational level. The framework showcases how ST can potentially calibrate trust in AI, improve decision-making, facilitate organizational collective actions, and cultivate holistic explainability. Our work contributes to the discourse of Human-Centered XAI by expanding the design space of XAI.",
"published":"2021-01-12T00:00:00.000Z",
"authors":"['Upol Ehsan', 'Q. Vera Liao', 'Michael Muller', 'Mark O. Riedl', 'Justin D. Weisz']",
"arxiv_primary_category":"cs.hc",
"readability":0.82,
"actuality":4.0,
"controversy":3.6666666667,
"relevance_magnitude":3.3333333333,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.72225
},
{
"arxiv_id":"2101.04794v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04794v1",
"title":"The Medical Authority of AI: A Study of AI-enabled Consumer-facing Health Technology",
"summary":"Recently, consumer-facing health technologies such as Artificial Intelligence (AI)-based symptom checkers (AISCs) have sprung up in everyday healthcare practice. AISCs solicit symptom information from users and provide medical suggestions and possible diagnoses, a responsibility that people usually entrust with real-person authorities such as physicians and expert patients. Thus, the advent of AISCs begs a question of whether and how they transform the notion of medical authority in everyday healthcare practice. To answer this question, we conducted an interview study with thirty AISC users. We found that users assess the medical authority of AISCs using various factors including automated decisions and interaction design patterns of AISC apps, associations with established medical authorities like hospitals, and comparisons with other health technologies. We reveal how AISCs are used in healthcare delivery, discuss how AI transforms conventional understandings of medical authority, and derive implications for designing AI-enabled health technology.",
"published":"2021-01-12T00:00:00.000Z",
"authors":"['Yue You', 'Yubo Kou', 'Xianghua Ding', 'Xinning Gui']",
"arxiv_primary_category":"cs.hc",
"readability":0.88,
"actuality":4.3333333333,
"controversy":4.0,
"relevance_magnitude":4.0,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":3.7778333333
},
{
"arxiv_id":"2101.04817v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04817v1",
"title":"Discrete Knowledge Graph Embedding based on Discrete Optimization",
"summary":"This paper proposes a discrete knowledge graph (KG) embedding (DKGE) method, which projects KG entities and relations into the Hamming space based on a computationally tractable discrete optimization algorithm, to solve the formidable storage and computation cost challenges in traditional continuous graph embedding methods. The convergence of DKGE can be guaranteed theoretically. Extensive experiments demonstrate that DKGE achieves superior accuracy than classical hashing functions that map the effective continuous embeddings into discrete codes. Besides, DKGE reaches comparable accuracy with much lower computational complexity and storage compared to many continuous graph embedding methods.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Yunqi Li', 'Shuyuan Xu', 'Bo Liu', 'Zuohui Fu', 'Shuchang Liu', 'Xu Chen', 'Yongfeng Zhang']",
"arxiv_primary_category":"cs.ir",
"readability":0.68,
"actuality":2.0,
"controversy":1.0,
"relevance_magnitude":3.0,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.1945
},
{
"arxiv_id":"2101.04834v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04834v1",
"title":"Whither AutoML? Understanding the Role of Automation in Machine Learning Workflows",
"summary":"Efforts to make machine learning more widely accessible have led to a rapid increase in Auto-ML tools that aim to automate the process of training and deploying machine learning. To understand how Auto-ML tools are used in practice today, we performed a qualitative study with participants ranging from novice hobbyists to industry researchers who use Auto-ML tools. We present insights into the benefits and deficiencies of existing tools, as well as the respective roles of the human and automation in ML workflows. Finally, we discuss design implications for the future of Auto-ML tool development. We argue that instead of full automation being the ultimate goal of Auto-ML, designers of these tools should focus on supporting a partnership between the user and the Auto-ML tool. This means that a range of Auto-ML tools will need to be developed to support varying user goals such as simplicity, reproducibility, and reliability.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Doris Xin', 'Eva Yiwei Wu', 'Doris Jung-Lin Lee', 'Niloufar Salehi', 'Aditya Parameswaran']",
"arxiv_primary_category":"cs.hc",
"readability":0.89,
"actuality":3.6666666667,
"controversy":2.6666666667,
"relevance_magnitude":3.6666666667,
"relevance_valence":4.444,
"newsworthiness_crowd_sum":3.611
},
{
"arxiv_id":"2101.04849v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04849v1",
"title":"Probabilistic Metric Learning with Adaptive Margin for Top-K Recommendation",
"summary":"Personalized recommender systems are playing an increasingly important role as more content and services become available and users struggle to identify what might interest them. Although matrix factorization and deep learning based methods have proved effective in user preference modeling, they violate the triangle inequality and fail to capture fine-grained preference information. To tackle this, we develop a distance-based recommendation model with several novel aspects: (i) each user and item are parameterized by Gaussian distributions to capture the learning uncertainties; (ii) an adaptive margin generation scheme is proposed to generate the margins regarding different training triplets; (iii) explicit user-user\/item-item similarity modeling is incorporated in the objective function. The Wasserstein distance is employed to determine preferences because it obeys the triangle inequality and can measure the distance between probabilistic distributions. Via a comparison using five real-world datasets with state-of-the-art methods, the proposed model outperforms the best existing models by 4-22% in terms of recall@K on Top-K recommendation.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Chen Ma', 'Liheng Ma', 'Yingxue Zhang', 'Ruiming Tang', 'Xue Liu', 'Mark Coates']",
"arxiv_primary_category":"cs.ir",
"readability":0.86,
"actuality":3.0,
"controversy":3.0,
"relevance_magnitude":4.0,
"relevance_valence":3.889,
"newsworthiness_crowd_sum":3.47225
},
{
"arxiv_id":"2101.04852v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04852v2",
"title":"Knowledge-Enhanced Top-K Recommendation in Poincar\u00e9 Ball",
"summary":"Personalized recommender systems are increasingly important as more content and services become available and users struggle to identify what might interest them. Thanks to the ability for providing rich information, knowledge graphs (KGs) are being incorporated to enhance the recommendation performance and interpretability. To effectively make use of the knowledge graph, we propose a recommendation model in the hyperbolic space, which facilitates the learning of the hierarchical structure of knowledge graphs. Furthermore, a hyperbolic attention network is employed to determine the relative importances of neighboring entities of a certain item. In addition, we propose an adaptive and fine-grained regularization mechanism to adaptively regularize items and their neighboring representations. Via a comparison using three real-world datasets with state-of-the-art methods, we show that the proposed model outperforms the best existing models by 2-16% in terms of NDCG@K on Top-K recommendation.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Chen Ma', 'Liheng Ma', 'Yingxue Zhang', 'Haolun Wu', 'Xue Liu', 'Mark Coates']",
"arxiv_primary_category":"cs.ir",
"readability":0.84,
"actuality":3.3333333333,
"controversy":2.3333333333,
"relevance_magnitude":3.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.9445
},
{
"arxiv_id":"2101.04899v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04899v2",
"title":"Experimental Evaluation of Deep Learning models for Marathi Text Classification",
"summary":"The Marathi language is one of the prominent languages used in India. It is predominantly spoken by the people of Maharashtra. Over the past decade, the usage of language on online platforms has tremendously increased. However, research on Natural Language Processing (NLP) approaches for Marathi text has not received much attention. Marathi is a morphologically rich language and uses a variant of the Devanagari script in the written form. This works aims to provide a comprehensive overview of available resources and models for Marathi text classification. We evaluate CNN, LSTM, ULMFiT, and BERT based models on two publicly available Marathi text classification datasets and present a comparative analysis. The pre-trained Marathi fast text word embeddings by Facebook and IndicNLP are used in conjunction with word-based models. We show that basic single layer models based on CNN and LSTM coupled with FastText embeddings perform on par with the BERT based models on the available datasets. We hope our paper aids focused research and experiments in the area of Marathi NLP.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Atharva Kulkarni', 'Meet Mandhane', 'Manali Likhitkar', 'Gayatri Kshirsagar', 'Jayashree Jagdale', 'Raviraj Joshi']",
"arxiv_primary_category":"cs.cl",
"readability":0.83,
"actuality":3.3333333333,
"controversy":1.6666666667,
"relevance_magnitude":2.6666666667,
"relevance_valence":2.222,
"newsworthiness_crowd_sum":2.4721666667
},
{
"arxiv_id":"2101.04992v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.04992v1",
"title":"Flatness Based Control of an Industrial Robot Joint Using Secondary Encoders",
"summary":"Due to their compliant structure, industrial robots without precision-enhancing measures are only to a limited extent suitable for machining applications. Apart from structural, thermal and bearing deformations, the main cause for compliant structure is backlash of transmission drives. This paper proposes a method to improve trajectory tracking accuracy by using secondary encoders and applying a feedback and a flatness based feed forward control strategy. For this purpose, a novel nonlinear, continuously differentiable dynamical model of a flexible robot joint is presented. The robot joint is modeled as a two-mass oscillator with pose-dependent inertia, nonlinear friction and nonlinear stiffness, including backlash. A flatness based feed forward control is designed to improve the guiding behaviour and a feedback controller, based on secondary encoders, is implemented for disturbance compensation. Using Automatic Differentiation, the nonlinear feed forward controller can be computed in a few microseconds online. Finally, the proposed algorithms are evaluated in simulations and experimentally on a real KUKA Quantec KR300 Ultra SE.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Jonas Weigand', 'Nigora Gafur', 'Martin Ruskowski']",
"arxiv_primary_category":"cs.ro",
"readability":0.79,
"actuality":3.0,
"controversy":1.3333333333,
"relevance_magnitude":2.3333333333,
"relevance_valence":2.778,
"newsworthiness_crowd_sum":2.3611666667
},
{
"arxiv_id":"2101.05004v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.05004v1",
"title":"Is the User Enjoying the Conversation? A Case Study on the Impact on the Reward Function",
"summary":"The impact of user satisfaction in policy learning task-oriented dialogue systems has long been a subject of research interest. Most current models for estimating the user satisfaction either (i) treat out-of-context short-texts, such as product reviews, or (ii) rely on turn features instead of on distributed semantic representations. In this work we adopt deep neural networks that use distributed semantic representation learning for estimating the user satisfaction in conversations. We evaluate the impact of modelling context length in these networks. Moreover, we show that the proposed hierarchical network outperforms state-of-the-art quality estimators. Furthermore, we show that applying these networks to infer the reward function in a Partial Observable Markov Decision Process (POMDP) yields to a great improvement in the task success rate.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Lina M. Rojas-Barahona']",
"arxiv_primary_category":"cs.cl",
"readability":0.87,
"actuality":4.6666666667,
"controversy":1.3333333333,
"relevance_magnitude":4.3333333333,
"relevance_valence":3.333,
"newsworthiness_crowd_sum":3.4165833333
},
{
"arxiv_id":"2101.05044v2",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.05044v2",
"title":"Publishing patterns reflect political polarization in news media",
"summary":"Digital news outlets rely on a variety of outside contributors, from freelance journalists, to political commentators, to executives and politicians. These external dependencies create a network among news outlets, traced along the contributors they share. Using connections between outlets, we demonstrate how contributors' publishing trajectories tend to align with outlet political leanings. We also show how polarized clustering of outlets translates to differences in the topics of news covered and the style and tone of articles published. In addition, we demonstrate how contributors who cross partisan divides tend to focus on less explicitly political topics. This work addresses an important gap in the media polarization literature, by highlighting how structural factors on the production side of news media create an ecosystem shaped by political leanings, independent of the priorities of any one person or organization.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Nick Hagar', 'Johannes Wachs', 'Em\u0151ke-\u00c1gnes Horv\u00e1t']",
"arxiv_primary_category":"cs.si",
"readability":0.9,
"actuality":4.6666666667,
"controversy":3.6666666667,
"relevance_magnitude":4.3333333333,
"relevance_valence":1.667,
"newsworthiness_crowd_sum":3.5834166667
},
{
"arxiv_id":"2101.05244v1",
"arxiv_url":"http:\/\/arxiv.org\/abs\/2101.05244v1",
"title":"Calibration Methods of Touch-Point Ambiguity for Finger-Fitts Law",
"summary":"Finger-Fitts law (FFitts law) is a model to predict touch-pointing times that was modified from Fitts' law. It considers the absolute touch-point precision, or a finger tremor factor sigma_a, to decrease the admissible target area and thus increase the task difficulty. Among choices such as running an independent task or performing parameter optimization, there is no consensus on the best methodology to measure sigma_a. By integrating the results of our 1D and 2D touch-pointing experiments and reanalyses of previous studies' data, we examined the advantages and disadvantages of each approach to compute sigma_a, and we found that there is no optimal choice to maximize the prediction accuracy of FFitts law.",
"published":"2021-01-13T00:00:00.000Z",
"authors":"['Shota Yamanaka', 'Hiroki Usuba']",
"arxiv_primary_category":"cs.hc",
"readability":0.83,