default search action
Shao-Yen Tseng
Person information
SPARQL queries
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
2020 – today
- 2024
- [c16]Zhipeng Cai, Matthias Mueller, Reiner Birkl, Diana Wofk, Shao-Yen Tseng, Junda Cheng, Gabriela Ben Melech Stan, Vasudev Lal, Michael Paulitsch:
L-MAGIC: Language Model Assisted Generation of Images with Coherence. CVPR 2024: 7049-7058 - [c15]Musashi Hinck, Carolin Holtermann, Matthew L. Olson, Florian Schneider, Sungduk Yu, Anahita Bhiwandiwalla, Anne Lauscher, Shao-Yen Tseng, Vasudev Lal:
Why do LLaVA Vision-Language Models Reply to Images in English? EMNLP (Findings) 2024: 13402-13421 - [i18]Musashi Hinck, Matthew L. Olson, David Cobbley, Shao-Yen Tseng, Vasudev Lal:
LLaVA-Gemma: Accelerating Multimodal Foundation Models with a Compact Language Model. CoRR abs/2404.01331 (2024) - [i17]Gabriela Ben Melech Stan, Raanan Y. Yehezkel Rohekar, Yaniv Gurwicz, Matthew Lyle Olson, Anahita Bhiwandiwalla, Estelle Aflalo, Chenfei Wu, Nan Duan, Shao-Yen Tseng, Vasudev Lal:
LVLM-Intrepret: An Interpretability Tool for Large Vision-Language Models. CoRR abs/2404.03118 (2024) - [i16]Zhipeng Cai, Matthias Mueller, Reiner Birkl, Diana Wofk, Shao-Yen Tseng, Junda Cheng, Gabriela Ben Melech Stan, Vasudev Lal, Michael Paulitsch:
L-MAGIC: Language Model Assisted Generation of Images with Coherence. CoRR abs/2406.01843 (2024) - [i15]Musashi Hinck, Carolin Holtermann, Matthew Lyle Olson, Florian Schneider, Sungduk Yu, Anahita Bhiwandiwalla, Anne Lauscher, Shao-Yen Tseng, Vasudev Lal:
Why do LLaVA Vision-Language Models Reply to Images in English? CoRR abs/2407.02333 (2024) - [i14]Neale Ratzlaff, Matthew Lyle Olson, Musashi Hinck, Shao-Yen Tseng, Vasudev Lal, Phillip Howard:
Debiasing Large Vision-Language Models by Ablating Protected Attribute Representations. CoRR abs/2410.13976 (2024) - [i13]Neale Ratzlaff, Matthew Lyle Olson, Musashi Hinck, Estelle Aflalo, Shao-Yen Tseng, Vasudev Lal, Phillip Howard:
Debias your Large Multi-Modal Model at Test-Time with Non-Contrastive Visual Attribute Steering. CoRR abs/2411.12590 (2024) - 2023
- [j4]Avinash Madasu, Estelle Aflalo, Gabriela Ben Melech Stan, Shachar Rosenman, Shao-Yen Tseng, Gedas Bertasius, Vasudev Lal:
MuMUR: Multilingual Multimodal Universal Retrieval. Inf. Retr. J. 26(1): 5 (2023) - [c14]Xiao Xu, Bei Li, Chenfei Wu, Shao-Yen Tseng, Anahita Bhiwandiwalla, Shachar Rosenman, Vasudev Lal, Wanxiang Che, Nan Duan:
ManagerTower: Aggregating the Insights of Uni-Modal Experts for Vision-Language Representation Learning. ACL (1) 2023: 14507-14525 - [c13]Avinash Madasu, Estelle Aflalo, Gabriela Ben Melech Stan, Shao-Yen Tseng, Gedas Bertasius, Vasudev Lal:
Improving Video Retrieval Using Multilingual Knowledge Transfer. ECIR (1) 2023: 669-684 - [i12]Gabriela Ben Melech Stan, Diana Wofk, Scottie Fox, Alex Redden, Will Saxton, Jean Yu, Estelle Aflalo, Shao-Yen Tseng, Fabio Nonato, Matthias Müller, Vasudev Lal:
LDM3D: Latent Diffusion Model for 3D. CoRR abs/2305.10853 (2023) - [i11]Xiao Xu, Bei Li, Chenfei Wu, Shao-Yen Tseng, Anahita Bhiwandiwalla, Shachar Rosenman, Vasudev Lal, Wanxiang Che, Nan Duan:
ManagerTower: Aggregating the Insights of Uni-Modal Experts for Vision-Language Representation Learning. CoRR abs/2306.00103 (2023) - [i10]Gabriela Ben Melech Stan, Diana Wofk, Estelle Aflalo, Shao-Yen Tseng, Zhipeng Cai, Michael Paulitsch, Vasudev Lal:
LDM3D-VR: Latent Diffusion Model for 3D VR. CoRR abs/2311.03226 (2023) - 2022
- [c12]Estelle Aflalo, Meng Du, Shao-Yen Tseng, Yongfei Liu, Chenfei Wu, Nan Duan, Vasudev Lal:
VL-InterpreT: An Interactive Visualization Tool for Interpreting Vision-Language Transformers. CVPR 2022: 21374-21383 - [c11]Yongfei Liu, Chenfei Wu, Shao-Yen Tseng, Vasudev Lal, Xuming He, Nan Duan:
KD-VLP: Improving End-to-End Vision-and-Language Pretraining with Object Knowledge Distillation. NAACL-HLT (Findings) 2022: 1589-1600 - [i9]Vin Sachidananda, Shao-Yen Tseng, Erik Marchi, Sachin Kajarekar, Panayiotis G. Georgiou:
CALM: Contrastive Aligned Audio-Language Multirate and Multimodal Representations. CoRR abs/2202.03587 (2022) - [i8]Estelle Aflalo, Meng Du, Shao-Yen Tseng, Yongfei Liu, Chenfei Wu, Nan Duan, Vasudev Lal:
VL-InterpreT: An Interactive Visualization Tool for Interpreting Vision-Language Transformers. CoRR abs/2203.17247 (2022) - [i7]Avinash Madasu, Estelle Aflalo, Gabriela Ben Melech Stan, Shao-Yen Tseng, Gedas Bertasius, Vasudev Lal:
Improving video retrieval using multilingual knowledge transfer. CoRR abs/2208.11553 (2022) - 2021
- [j3]Shao-Yen Tseng, Shrikanth Narayanan, Panayiotis G. Georgiou:
Multimodal Embeddings From Language Models for Emotion Recognition in the Wild. IEEE Signal Process. Lett. 28: 608-612 (2021) - [i6]Yongfei Liu, Chenfei Wu, Shao-Yen Tseng, Vasudev Lal, Xuming He, Nan Duan:
KD-VLP: Improving End-to-End Vision-and-Language Pretraining with Object Knowledge Distillation. CoRR abs/2109.10504 (2021) - 2020
- [c10]Sandeep Nallan Chakravarthula, Md. Nasir, Shao-Yen Tseng, Haoqi Li, Tae Jin Park, Brian R. Baucom, Craig J. Bryan, Shrikanth Narayanan, Panayiotis G. Georgiou:
Automatic Prediction of Suicidal Risk in Military Couples Using Multimodal Interaction Cues from Couples Conversations. ICASSP 2020: 6539-6543
2010 – 2019
- 2019
- [j2]Shao-Yen Tseng, Brian R. Baucom, Panayiotis G. Georgiou:
Unsupervised online multitask learning of behavioral sentence embeddings. PeerJ Comput. Sci. 5: e200 (2019) - [c9]Sandeep Nallan Chakravarthula, Haoqi Li, Shao-Yen Tseng, Maija Reblin, Panayiotis G. Georgiou:
Predicting Behavior in Cancer-Afflicted Patient and Spouse Interactions Using Speech and Language. INTERSPEECH 2019: 3073-3077 - [i5]Sandeep Nallan Chakravarthula, Haoqi Li, Shao-Yen Tseng, Maija Reblin, Panayiotis G. Georgiou:
Predicting Behavior in Cancer-Afflicted Patient and Spouse Interactions using Speech and Language. CoRR abs/1908.00908 (2019) - [i4]Prashanth Gurunath Shivakumar, Shao-Yen Tseng, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
Behavior Gated Language Models. CoRR abs/1909.00107 (2019) - [i3]Shao-Yen Tseng, Panayiotis G. Georgiou, Shrikanth S. Narayanan:
Multimodal Embeddings from Language Models. CoRR abs/1909.04302 (2019) - 2018
- [c8]Shao-Yen Tseng, Haoqi Li, Brian R. Baucom, Panayiotis G. Georgiou:
"Honey, I Learned to Talk": Multimodal Fusion for Behavior Analysis. ICMI 2018: 239-243 - [c7]Shao-Yen Tseng, Juncheng Li, Yun Wang, Florian Metze, Joseph Szurley, Samarjit Das:
Multiple Instance Deep Learning for Weakly Supervised Small-Footprint Audio Event Detection. INTERSPEECH 2018: 3279-3283 - [i2]Shao-Yen Tseng, Panayiotis G. Georgiou:
Multi-Task Unsupervised Contextual Learning for Behavioral Annotation. CoRR abs/1807.06792 (2018) - 2017
- [c6]Shao-Yen Tseng, Brian R. Baucom, Panayiotis G. Georgiou:
Approaching Human Performance in Behavior Estimation in Couples Therapy Using Deep Sentence Embeddings. INTERSPEECH 2017: 3291-3295 - [i1]Shao-Yen Tseng, Juncheng Li, Yun Wang, Joseph Szurley, Florian Metze, Samarjit Das:
Multiple Instance Deep Learning for Weakly Supervised Audio Event Detection. CoRR abs/1712.09673 (2017) - 2016
- [c5]Shao-Yen Tseng, Sandeep Nallan Chakravarthula, Brian R. Baucom, Panayiotis G. Georgiou:
Couples Behavior Modeling and Annotation Using Low-Resource LSTM Language Models. INTERSPEECH 2016: 898-902 - 2013
- [j1]Wai-Chi Fang, Hsiang-Cheh Huang, Shao-Yen Tseng:
Design of heart rate variability processor for portable 3-lead ECG monitoring system-on-chip. Expert Syst. Appl. 40(5): 1491-1504 (2013) - 2011
- [c4]Ericson Chua, Wai-Chi Fang, Chiu-Kuo Chen, Chih-Chung Fu, Shao-Yen Tseng, Shih Kang, Zong-Han Hsieh:
A highly-integrated biomedical multiprocessor system for portable brain-heart monitoring. ISCAS 2011: 1532-1535 - 2010
- [c3]Chih-Chung Fu, Chiu-Kuo Chen, Shao-Yen Tseng, Shih Kang, Ericson Chua, Wai-Chi Fang:
Portable Brain-Heart Monitoring System. FGIT-DTA/BSBT 2010: 241-250 - [c2]Shao-Yen Tseng, Wai-Chi Fang:
A Time-Frequency HRV Processor Using Windowed Lomb Periodogram. FGIT-DTA/BSBT 2010: 273-282 - [c1]Chiu-Kuo Chen, Ericson Chua, Shao-Yen Tseng, Chih-Chung Fu, Wai-Chi Fang:
Implementation of a hardware-efficient EEG processor for brain monitoring systems. SoCC 2010: 164-168
Coauthor Index
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.
Unpaywalled article links
Add open access links from to the list of external document links (if available).
Privacy notice: By enabling the option above, your browser will contact the API of unpaywall.org to load hyperlinks to open access articles. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Unpaywall privacy policy.
Archived links via Wayback Machine
For web page which are no longer available, try to retrieve content from the of the Internet Archive (if available).
Privacy notice: By enabling the option above, your browser will contact the API of archive.org to check for archived content of web pages that are no longer available. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Internet Archive privacy policy.
Reference lists
Add a list of references from , , and to record detail pages.
load references from crossref.org and opencitations.net
Privacy notice: By enabling the option above, your browser will contact the APIs of crossref.org, opencitations.net, and semanticscholar.org to load article reference information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the Crossref privacy policy and the OpenCitations privacy policy, as well as the AI2 Privacy Policy covering Semantic Scholar.
Citation data
Add a list of citing articles from and to record detail pages.
load citations from opencitations.net
Privacy notice: By enabling the option above, your browser will contact the API of opencitations.net and semanticscholar.org to load citation information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the OpenCitations privacy policy as well as the AI2 Privacy Policy covering Semantic Scholar.
OpenAlex data
Load additional information about publications from .
Privacy notice: By enabling the option above, your browser will contact the API of openalex.org to load additional information. Although we do not have any reason to believe that your call will be tracked, we do not have any control over how the remote server uses your data. So please proceed with care and consider checking the information given by OpenAlex.
last updated on 2025-01-02 18:09 CET by the dblp team
all metadata released as open data under CC0 1.0 license
see also: Terms of Use | Privacy Policy | Imprint