<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>Yang Gao</title><link>https://ygao36buffalo.github.io/</link><atom:link href="https://ygao36buffalo.github.io/index.xml" rel="self" type="application/rss+xml"/><description>Yang Gao</description><generator>Hugo Blox Builder (https://hugoblox.com)</generator><language>en-us</language><lastBuildDate>Mon, 24 Oct 2022 00:00:00 +0000</lastBuildDate><image><url>https://ygao36buffalo.github.io/media/icon_hu7729264130191091259.png</url><title>Yang Gao</title><link>https://ygao36buffalo.github.io/</link></image><item><title>🎉 Our Paper Accepted to CHI 2026!</title><link>https://ygao36buffalo.github.io/post/chi2026/</link><pubDate>Sat, 17 Jan 2026 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/post/chi2026/</guid><description>
&lt;details class="print:hidden xl:hidden" open>
&lt;summary>Table of Contents&lt;/summary>
&lt;div class="text-sm">
&lt;nav id="TableOfContents">
&lt;ul>
&lt;li>&lt;a href="#overview">Overview&lt;/a>&lt;/li>
&lt;/ul>
&lt;/nav>
&lt;/div>
&lt;/details>
&lt;h2 id="overview">Overview&lt;/h2>
&lt;p>Stay tuned for the full paper and presentation details.&lt;/p></description></item><item><title>Wrist2Finger: Sensing Fingertip Force for Force-Aware Hand Interaction with a Ring-Watch Wearable</title><link>https://ygao36buffalo.github.io/publication/gao-wrist2finger-2025/</link><pubDate>Wed, 01 Oct 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-wrist2finger-2025/</guid><description/></item><item><title>KineticsSense: A Multimodal Wearable Sensor Framework for Modeling Lower-Limb Motion Kinetics</title><link>https://ygao36buffalo.github.io/publication/gao-kinesense-2025/</link><pubDate>Wed, 03 Sep 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-kinesense-2025/</guid><description/></item><item><title>Motion2Press: Cross Model Learning from IMU to Plantar Pressure for Gait Analysis</title><link>https://ygao36buffalo.github.io/publication/gao-motion2press-2025/</link><pubDate>Wed, 03 Sep 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-motion2press-2025/</guid><description/></item><item><title>KineticsSense: A Multimodal Wearable Sensor Framework for Modeling Lower-Limb Motion Kinetics</title><link>https://ygao36buffalo.github.io/publication/10-1145-3749462/</link><pubDate>Mon, 01 Sep 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/10-1145-3749462/</guid><description/></item><item><title>Motion2Press: Cross Model Learning from IMU to Plantar Pressure for Gait Analysis</title><link>https://ygao36buffalo.github.io/publication/10-1145-3749499/</link><pubDate>Mon, 01 Sep 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/10-1145-3749499/</guid><description/></item><item><title>Family Moments</title><link>https://ygao36buffalo.github.io/gallery/family-moments/</link><pubDate>Sun, 06 Jul 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/gallery/family-moments/</guid><description>&lt;p>It&amp;rsquo;s always wonderful to spend quality time with my family. Here are a few highlights from our precious time together:&lt;/p>
&lt;figure>&lt;img src="https://ygao36buffalo.github.io/gallery/family-moments/family_photo_1.jpg">
&lt;/figure>
&lt;figure>&lt;img src="https://ygao36buffalo.github.io/gallery/family-moments/family_photo_2.jpg">
&lt;/figure>
&lt;p>I cherish these memories and look forward to many more!&lt;/p></description></item><item><title>🎉 Two Papers Newly Accepted for UbiComp 2025 Presentation!</title><link>https://ygao36buffalo.github.io/post/ubicomp2025-2/</link><pubDate>Sat, 05 Jul 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/post/ubicomp2025-2/</guid><description>
&lt;details class="print:hidden xl:hidden" open>
&lt;summary>Table of Contents&lt;/summary>
&lt;div class="text-sm">
&lt;nav id="TableOfContents">
&lt;ul>
&lt;li>&lt;a href="#overview">Overview&lt;/a>&lt;/li>
&lt;/ul>
&lt;/nav>
&lt;/div>
&lt;/details>
&lt;h2 id="overview">Overview&lt;/h2>
&lt;p>Stay tuned for the full paper and presentation details.&lt;/p></description></item><item><title>🎉 Our Paper Accepted to UIST 2025!</title><link>https://ygao36buffalo.github.io/post/uist2025/</link><pubDate>Fri, 04 Jul 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/post/uist2025/</guid><description>
&lt;details class="print:hidden xl:hidden" open>
&lt;summary>Table of Contents&lt;/summary>
&lt;div class="text-sm">
&lt;nav id="TableOfContents">
&lt;ul>
&lt;li>&lt;a href="#overview">Overview&lt;/a>&lt;/li>
&lt;/ul>
&lt;/nav>
&lt;/div>
&lt;/details>
&lt;h2 id="overview">Overview&lt;/h2>
&lt;p>Stay tuned for the full paper and presentation details.&lt;/p></description></item><item><title>EchoTouch: Low-power Face-touching Behavior Recognition Using Active Acoustic Sensing on Glasses</title><link>https://ygao36buffalo.github.io/publication/guo-echotouch-2025/</link><pubDate>Sun, 01 Jun 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/guo-echotouch-2025/</guid><description/></item><item><title>CHI 2025 &amp; Exploring Yokohama, Japan</title><link>https://ygao36buffalo.github.io/gallery/trip-to-chi2025-yokohama-japan/</link><pubDate>Tue, 29 Apr 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/gallery/trip-to-chi2025-yokohama-japan/</guid><description>&lt;p>Attending &lt;strong>CHI 2025&lt;/strong> in Yokohama, Japan, was an absolutely incredible experience – a perfect blend of diving deep into the latest Human-Computer Interaction research and exploring the captivating beauty of Japan. It was a week filled with inspiring talks, engaging discussions, and unforgettable cultural moments.&lt;/p>
&lt;h2 id="the-conference-experience-">The Conference Experience 🔬🤝&lt;/h2>
&lt;p>CHI 2025, as always, was a powerhouse of innovation and a fantastic platform for presenting cutting-edge research.&lt;/p>
&lt;p>I had the honor of presenting our work, &lt;strong>&amp;ldquo;SandTouch: Empowering Virtual Sand Art in VR with AI Guidance and Emotional Relief,&amp;rdquo;&lt;/strong> which was met with great interest and engaging questions. It was truly rewarding to share our findings and receive valuable feedback from the community.&lt;/p>
&lt;figure>&lt;img src="https://ygao36buffalo.github.io/gallery/trip-to-chi2025-yokohama-japan/my_sandtouch_presentation.jpg"
alt="Presenting &amp;lsquo;SandTouch&amp;rsquo; at CHI 2025.">&lt;figcaption>
&lt;p>Presenting &amp;lsquo;SandTouch&amp;rsquo; at CHI 2025.&lt;/p>
&lt;/figcaption>
&lt;/figure>
&lt;p>The conference featured a diverse range of talks, from fundamental research to innovative applications. Listening to fellow researchers share their groundbreaking work was truly inspiring. I particularly enjoyed connecting with researchers working on similar topics. Here&amp;rsquo;s a glimpse of one of the engaging presentations:&lt;/p>
&lt;figure>&lt;img src="https://ygao36buffalo.github.io/gallery/trip-to-chi2025-yokohama-japan/chi_other_presentation.jpg"
alt="An insightful presentation on &amp;lsquo;Conveying Dancers’ Expressiveness Through A Haptic-Installed Contemporary Dance Performance&amp;rsquo;.">&lt;figcaption>
&lt;p>An insightful presentation on &amp;lsquo;Conveying Dancers’ Expressiveness Through A Haptic-Installed Contemporary Dance Performance&amp;rsquo;.&lt;/p>
&lt;/figcaption>
&lt;/figure>
&lt;p>The conference venue itself was impressive, offering a conducive environment for learning and networking.&lt;/p>
&lt;figure>&lt;img src="https://ygao36buffalo.github.io/gallery/trip-to-chi2025-yokohama-japan/yokohama_conference_center_exterior.jpg"
alt="The sleek architecture of the CHI 2025 conference center.">&lt;figcaption>
&lt;p>The sleek architecture of the CHI 2025 conference center.&lt;/p>
&lt;/figcaption>
&lt;/figure>
&lt;h2 id="exploring-yokohamas-charms-">Exploring Yokohama&amp;rsquo;s Charms 🌃🇯🇵&lt;/h2>
&lt;p>Beyond the conference halls, Yokohama proved to be a vibrant and charming city, offering a wonderful backdrop for unwinding and exploring. Its beautiful waterfront, modern architecture, and serene streets made for a delightful experience.&lt;/p>
&lt;p>I took the chance to visit the &lt;strong>Nissan Gallery&lt;/strong>, where cutting-edge automotive design meets an immersive brand experience. The displays were truly impressive!&lt;/p>
&lt;figure>&lt;img src="https://ygao36buffalo.github.io/gallery/trip-to-chi2025-yokohama-japan/yokohama_nissan_gallery.jpg"
alt="A peek into the future of automotive design at the Nissan Gallery.">&lt;figcaption>
&lt;p>A peek into the future of automotive design at the Nissan Gallery.&lt;/p>
&lt;/figcaption>
&lt;/figure>
&lt;p>Strolling through Yokohama&amp;rsquo;s streets, I was captivated by the city&amp;rsquo;s natural beauty, including the vibrant flora that lines its avenues.&lt;/p>
&lt;figure>&lt;img src="https://ygao36buffalo.github.io/gallery/trip-to-chi2025-yokohama-japan/yokohama_street_flowers.jpg"
alt="Beautiful white azaleas blooming along a Yokohama street.">&lt;figcaption>
&lt;p>Beautiful white azaleas blooming along a Yokohama street.&lt;/p>
&lt;/figcaption>
&lt;/figure>
&lt;p>This trip was a fantastic opportunity to combine academic growth with cultural immersion. I&amp;rsquo;m already looking forward to the next CHI!&lt;/p>
&lt;hr>
&lt;p>Did you find this page helpful? Consider sharing it 🙌&lt;/p></description></item><item><title>🎉 Double Acceptance at CHI 2025, One Full Paper &amp; Late-Breaking Work!</title><link>https://ygao36buffalo.github.io/post/chi2025/</link><pubDate>Fri, 25 Apr 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/post/chi2025/</guid><description>
&lt;details class="print:hidden xl:hidden" open>
&lt;summary>Table of Contents&lt;/summary>
&lt;div class="text-sm">
&lt;nav id="TableOfContents">
&lt;ul>
&lt;li>&lt;a href="#overview">Overview&lt;/a>&lt;/li>
&lt;/ul>
&lt;/nav>
&lt;/div>
&lt;/details>
&lt;h2 id="overview">Overview&lt;/h2>
&lt;p>For more in-depth information about these exciting works, please visit their dedicated project pages:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>SandTouch:&lt;/strong> &lt;a href="https://ygao36buffalo.github.io/project/sandtouch/">Explore our full paper here&lt;/a>&lt;/li>
&lt;li>&lt;strong>From Wrist to Finger:&lt;/strong> &lt;a href="https://ygao36buffalo.github.io/project/from-wrist-to-finger/">Discover our Late-Breaking Work here&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>From Wrist to Finger: Hand Pose Tracking Using Ring-Watch Wearables</title><link>https://ygao36buffalo.github.io/project/from-wrist-to-finger/</link><pubDate>Fri, 25 Apr 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/project/from-wrist-to-finger/</guid><description>&lt;h2 id="project-overview-chi-2025-late-breaking-work">Project Overview (CHI 2025 Late-Breaking Work)&lt;/h2>
&lt;p>Hand pose tracking is a cornerstone for advancing human-computer interaction applications, from virtual reality to prosthetics control. However, existing vision-based systems and wearable devices often face limitations in portability, usability, and practicality. In this &lt;strong>CHI 2025 Late-Breaking Work&lt;/strong>, we introduce a novel multimodal hand pose tracking framework that integrates data from an IMU-equipped ring and EMG sensors embedded in a wrist-worn device.&lt;/p>
&lt;p>This work represents an important step towards enabling more practical and accessible hand tracking solutions for everyday use.&lt;/p>
&lt;h2 id="key-innovations--progress">Key Innovations &amp;amp; Progress:&lt;/h2>
&lt;h3 id="1-novel-ring-watch-wearable-design">1. Novel Ring-Watch Wearable Design&lt;/h3>
&lt;p>We propose a compact and ergonomic wearable system that combines a single IMU-equipped ring (worn on the thumb) with EMG sensors integrated into a smartwatch. This design prioritizes wearability and comfort, addressing the common limitations of bulky or multi-device tracking systems while accurately capturing intricate finger and hand motion data.&lt;/p>
&lt;h3 id="2-multi-sensor-fusion-for-precise-tracking">2. Multi-Sensor Fusion for Precise Tracking&lt;/h3>
&lt;p>Our framework leverages the complementary strengths of both motion dynamics (from the IMU) and muscle activity (from the EMG sensors). This deep learning-based sensor fusion approach achieves precise 3D hand pose reconstruction, providing robust performance even in complex or high-speed gestures. We developed a transformer-based model with time encoding and cross-modal attention mechanisms for optimal data integration.&lt;/p>
&lt;h2 id="media--resources">Media &amp;amp; Resources:&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>Paper:&lt;/strong> &lt;a href="https://doi.org/10.1145/3706599.3720220" target="_blank" rel="noopener">Extended Abstract (ACM DL)&lt;/a>&lt;/li>
&lt;li>&lt;strong>Video:&lt;/strong> &lt;a href="https://www.youtube.com/watch?v=g6bTCVMpNf4" target="_blank" rel="noopener">Project Video&lt;/a>&lt;/li>
&lt;/ul>
&lt;hr>
&lt;p>Did you find this page helpful? Consider sharing it 🙌&lt;/p></description></item><item><title>SandTouch: Empowering Virtual Sand Art in VR with AI Guidance and Emotional Relief</title><link>https://ygao36buffalo.github.io/project/sandtouch/</link><pubDate>Fri, 25 Apr 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/project/sandtouch/</guid><description>&lt;h2 id="project-overview">Project Overview&lt;/h2>
&lt;p>In today&amp;rsquo;s data-rich world, simply having data isn&amp;rsquo;t enough – we need to understand it, communicate it, and extract insights from it. This course empowers you to transform raw numbers and complex datasets into compelling, insightful visual narratives. We&amp;rsquo;ll blend art with science, learning how to design interactive visualizations that not only look good but also enable powerful data exploration and decision-making. Get ready to become a data storyteller!&lt;/p>
&lt;p>Sand painting is a unique and valuable art form, but it&amp;rsquo;s often constrained by physical equipment and a steep learning curve. To address these challenges, we developed &lt;strong>SandTouch&lt;/strong>, a novel VR sand painting system that offers an immersive and intuitive experience, closely mirroring real-world sand manipulation. This project was presented at &lt;strong>CHI 2025&lt;/strong> (the premier conference in Human-Computer Interaction).&lt;/p>
&lt;h2 id="key-features--contributions">Key Features &amp;amp; Contributions:&lt;/h2>
&lt;h3 id="1-realistic-hand-sand-interaction">1. Realistic Hand-Sand Interaction&lt;/h3>
&lt;p>We designed SandTouch to create a highly realistic and natural interaction between the user&amp;rsquo;s hands and virtual sand, allowing direct manipulation without external devices or controllers. This approach provides an intuitive, device-free interface and emphasizes the authenticity of the interaction. Our system captures fine sensations of real sand manipulation, complemented by realistic sound feedback.&lt;/p>
&lt;h3 id="2-ai-guidance-for-creative-flow">2. AI Guidance for Creative Flow&lt;/h3>
&lt;p>A pioneering aspect of SandTouch is the integration of an AI agent, powered by a large language model (LLM). This AI intelligently interprets users&amp;rsquo; creative intentions in real-time, offering contextually relevant artistic suggestions. This feature simplifies the creative process, enhances interactivity, and helps users (especially beginners) refine their artwork through technique recommendations and composition analysis.&lt;/p>
&lt;h3 id="3-emotional-relief--immersion">3. Emotional Relief &amp;amp; Immersion&lt;/h3>
&lt;p>Beyond artistic creation, SandTouch prioritizes user well-being. It incorporates calming and responsive soundscapes that react to user gestures, reinforcing a relaxing atmosphere. Comprehensive evaluations demonstrated a significant increase in user engagement and immersion, with the realistic sound feedback enhancing emotional relief and deepening the painting experience. This highlights its potential for therapeutic applications.&lt;/p>
&lt;h2 id="media--resources">Media &amp;amp; Resources:&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>Paper:&lt;/strong> &lt;a href="https://doi.org/10.1145/3706598.3714275" target="_blank" rel="noopener">Full Paper (ACM DL)&lt;/a>&lt;/li>
&lt;li>&lt;strong>Video:&lt;/strong> &lt;a href="https://www.youtube.com/watch?v=6FYOCeU0liw" target="_blank" rel="noopener">Project Video&lt;/a>&lt;/li>
&lt;/ul>
&lt;hr>
&lt;p>Did you find this page helpful? Consider sharing it 🙌&lt;/p></description></item><item><title>From Wrist to Finger: Hand Pose Tracking Using Ring-Watch Wearables</title><link>https://ygao36buffalo.github.io/publication/xiao-wrist-2025/</link><pubDate>Tue, 01 Apr 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/xiao-wrist-2025/</guid><description/></item><item><title>SandTouch: Empowering Virtual Sand Art in VR with AI Guidance and Emotional Relief</title><link>https://ygao36buffalo.github.io/publication/10-1145-3706598-3714275/</link><pubDate>Wed, 01 Jan 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/10-1145-3706598-3714275/</guid><description/></item><item><title>Wrist2Finger: Sensing Fingertip Force for Force-Aware Hand Interaction with a Ring-Watch Wearable</title><link>https://ygao36buffalo.github.io/publication/10-1145-3746059-3747767/</link><pubDate>Wed, 01 Jan 2025 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/10-1145-3746059-3747767/</guid><description/></item><item><title>🎉 New Paper Accepted to UbiComp 2025 (IMWUT 2024 May Cycle)!</title><link>https://ygao36buffalo.github.io/post/ubicomp2025-1/</link><pubDate>Thu, 21 Nov 2024 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/post/ubicomp2025-1/</guid><description>
&lt;details class="print:hidden xl:hidden" open>
&lt;summary>Table of Contents&lt;/summary>
&lt;div class="text-sm">
&lt;nav id="TableOfContents">
&lt;ul>
&lt;li>&lt;a href="#overview">Overview&lt;/a>&lt;/li>
&lt;/ul>
&lt;/nav>
&lt;/div>
&lt;/details>
&lt;h2 id="overview">Overview&lt;/h2>
&lt;p>For more in-depth information about this exciting work, please visit the dedicated project page:&lt;/p>
&lt;ul>
&lt;li>&lt;strong>PressInPose:&lt;/strong> &lt;a href="https://ygao36buffalo.github.io/project/pressinpose/">Explore our full paper here&lt;/a>&lt;/li>
&lt;/ul></description></item><item><title>PressInPose: Integrating Pressure and Inertial Sensors for Full-Body Pose Estimation in Activities</title><link>https://ygao36buffalo.github.io/project/pressinpose/</link><pubDate>Thu, 21 Nov 2024 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/project/pressinpose/</guid><description>&lt;h2 id="project-overview">Project Overview&lt;/h2>
&lt;p>Accurate human body posture assessment through wearable technology has significant implications across various fields, including sports science, clinical diagnostics, rehabilitation, and VR interaction. Traditional methods often face limitations due to complex setups or environmental constraints. To address these challenges, we developed &lt;strong>PressInPose&lt;/strong>, an innovative system that integrates pressure and inertial sensors for precise full-body pose estimation in dynamic activities. This work was published in &lt;strong>Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies (IMWUT)&lt;/strong> and will be presented at &lt;strong>UbiComp 2025&lt;/strong>.&lt;/p>
&lt;h2 id="key-innovations--contributions">Key Innovations &amp;amp; Contributions:&lt;/h2>
&lt;h3 id="1-novel-multi-sensor-fusion">1. Novel Multi-Sensor Fusion&lt;/h3>
&lt;p>PressInPose employs an advanced shoe insole embedded with pressure sensors and an Inertial Measurement Unit (IMU), coupled with a single wrist-mounted IMU. This unique multi-modal sensor fusion approach allows for a comprehensive analysis of human biomechanics, capturing intricate body dynamics that traditional single-sensor systems often miss.&lt;/p>
&lt;h3 id="2-llm-powered-virtual-data-augmentation">2. LLM-Powered Virtual Data Augmentation&lt;/h3>
&lt;p>To enhance the robustness and generalization of our system, we leveraged large language models (LLMs) to generate virtual human motion sequences. These sequences were utilized to create synthetic IMU data for data augmentation, effectively addressing the challenge of limited real-world data availability and variability, especially for complex and dynamic movements.&lt;/p>
&lt;h3 id="3-physical-kinematics-modeling--deep-learning-network">3. Physical Kinematics Modeling &amp;amp; Deep Learning Network&lt;/h3>
&lt;p>Our approach uniquely combines physical kinematics modeling based on pressure data with a multi-region human posture estimation network. This integration allows PressInPose to accurately capture interactions and dependencies between different body parts, leading to superior accuracy in pose reconstruction.&lt;/p>
&lt;h2 id="media--resources">Media &amp;amp; Resources:&lt;/h2>
&lt;ul>
&lt;li>&lt;strong>Paper:&lt;/strong> &lt;a href="%60https://doi.org/10.1145/3699773%60">Full Paper (ACM DL)&lt;/a>&lt;/li>
&lt;/ul>
&lt;hr>
&lt;p>Did you find this page helpful? Consider sharing it 🙌&lt;/p></description></item><item><title>PressInPose: Integrating Pressure and Inertial Sensors for Full-Body Pose Estimation in Activities</title><link>https://ygao36buffalo.github.io/publication/gao-pressinpose-2024/</link><pubDate>Fri, 01 Nov 2024 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-pressinpose-2024/</guid><description/></item><item><title>Data Visualization &amp; Interaction (Offered Every Fall)</title><link>https://ygao36buffalo.github.io/teaching/data-visualization-and-interaction/</link><pubDate>Sun, 01 Sep 2024 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/teaching/data-visualization-and-interaction/</guid><description>&lt;p>In our increasingly data-rich world, simply having data isn&amp;rsquo;t enough – we need to understand it, communicate it effectively, and extract actionable insights from it. This course empowers you to transform raw numbers and complex datasets into compelling, insightful visual narratives. We&amp;rsquo;ll blend the art of design with the science of data, learning how to create interactive visualizations that not only look good but also enable powerful data exploration and decision-making. Get ready to become a data storyteller!&lt;/p>
&lt;h2 id="what-youll-explore">What You&amp;rsquo;ll Explore:&lt;/h2>
&lt;h3 id="1-visual-perception--cognition-">1. Visual Perception &amp;amp; Cognition 👁️🧠&lt;/h3>
&lt;p>The human element is key! We&amp;rsquo;ll start by understanding how our eyes and brains process visual information, exploring principles of pre-attentive attributes, Gestalt psychology, and cognitive biases. This lays the groundwork for designing effective and intuitive visualizations.&lt;/p>
&lt;h3 id="2-data-fundamentals--preparation-">2. Data Fundamentals &amp;amp; Preparation 🧹&lt;/h3>
&lt;p>Before we visualize, we clean and structure! Learn about different data types (quantitative, categorical, ordinal), data quality issues, and essential techniques for wrangling, cleaning, and transforming your data to make it suitable for visualization.&lt;/p>
&lt;h3 id="3-visualizing-spatial-data-">3. Visualizing Spatial Data 🗺️&lt;/h3>
&lt;p>Explore techniques for representing geographic locations and spatial relationships. From simple scatter plots on maps to choropleth maps, heatmaps, and flow maps, we&amp;rsquo;ll learn to convey location-based insights effectively.&lt;/p>
&lt;h3 id="4-visualizing-tree-structures-">4. Visualizing Tree Structures 🌲&lt;/h3>
&lt;p>Dive into hierarchical data! Learn how to effectively visualize relationships and structures in organizational charts, file systems, category taxonomies, and other tree-like datasets using methods like treemaps, sunburst charts, and node-link diagrams.&lt;/p>
&lt;h3 id="5-visualizing-graph-structures-networks-">5. Visualizing Graph Structures (Networks) 🕸️&lt;/h3>
&lt;p>Uncover connections! Master techniques for visualizing complex networks, social graphs, co-occurrence relationships, and other graph data to reveal hidden patterns, communities, and central nodes.&lt;/p>
&lt;h3 id="6-visualizing-time-varying-data-">6. Visualizing Time-Varying Data 📈&lt;/h3>
&lt;p>See trends unfold! Learn how to effectively represent changes over time, from traditional line charts and area charts to more advanced streamgraphs, sparklines, and animated timelines that show evolution and patterns over temporal dimensions.&lt;/p>
&lt;h3 id="7-visualizing-text-data-">7. Visualizing Text Data 💬&lt;/h3>
&lt;p>Transform words into insights! Explore methods for visualizing textual information, including word clouds, n-gram charts, sentiment analysis visualizations, topic models, and network graphs derived from text.&lt;/p>
&lt;h3 id="8-geospatial-information-systems-gis--advanced-mapping-">8. Geospatial Information Systems (GIS) &amp;amp; Advanced Mapping 🌍&lt;/h3>
&lt;p>Beyond basic maps, delve into the power of GIS for analyzing and visualizing complex geographical datasets with real-world applications in urban planning, environmental science, and public health.&lt;/p>
&lt;h3 id="9-visualizing-high-dimensional-data-">9. Visualizing High-Dimensional Data ✨&lt;/h3>
&lt;p>Tame complexity! Learn strategies to visualize datasets with many attributes (dimensions), employing techniques like parallel coordinates, scatter plot matrices, dimensionality reduction (e.g., PCA, t-SNE), and glyph-based representations.&lt;/p>
&lt;h3 id="10-interactive-views--dashboard-design-">10. Interactive Views &amp;amp; Dashboard Design 🕹️&lt;/h3>
&lt;p>Bring your data to life! Explore principles of interactive design, enabling users to filter, zoom, pan, drill down, and explore data dynamically. We&amp;rsquo;ll also cover best practices for designing effective and user-friendly dashboards.&lt;/p>
&lt;hr>
&lt;div class="flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900">
&lt;span class="pr-3 pt-1 text-primary-600 dark:text-primary-300">
&lt;svg height="24" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24">&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/>&lt;/svg>
&lt;/span>
&lt;span class="dark:text-neutral-300">By the end of this course, you&amp;rsquo;ll not only be able to create beautiful and effective visualizations but also to communicate complex data stories with clarity and empower users to gain actionable insights. This is a critical skill for anyone navigating our data-driven world!&lt;/span>
&lt;/div>
&lt;hr>
&lt;p>Did you find this page helpful? Consider sharing it 🙌&lt;/p></description></item><item><title>Projects</title><link>https://ygao36buffalo.github.io/projects/</link><pubDate>Sun, 19 May 2024 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/projects/</guid><description/></item><item><title>Computer Organization (Offered Every Spring)</title><link>https://ygao36buffalo.github.io/teaching/computer-organization/</link><pubDate>Thu, 01 Feb 2024 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/teaching/computer-organization/</guid><description>&lt;p>Ever wondered what magic truly happens inside your computer when you click a mouse, launch an app, or even just boot it up? This course takes you on an exciting journey deep into the heart of modern computing systems. We&amp;rsquo;ll peel back the layers of abstraction to understand the fundamental building blocks and principles that make software run efficiently on hardware. Get ready to gain a profound appreciation for the intricate dance between bits, bytes, and transistors!&lt;/p>
&lt;h2 id="syllabus">Syllabus:&lt;/h2>
&lt;h3 id="1-introduction-to-computer-systems-">1. Introduction to Computer Systems 🚀&lt;/h3>
&lt;p>An exciting overview of the computer&amp;rsquo;s anatomy – from high-level software to low-level hardware. We&amp;rsquo;ll set the stage for our adventure into the machine&amp;rsquo;s core, understanding the big picture before diving into the details.&lt;/p>
&lt;h3 id="2-information-encoding--data-representation-">2. Information Encoding &amp;amp; Data Representation 🔡&lt;/h3>
&lt;p>How do computers understand numbers, text, images, and sound? We&amp;rsquo;ll decode the secrets of binary, hexadecimal, and various data formats that form the computer&amp;rsquo;s native language. Learn how all digital information is just a series of bits!&lt;/p>
&lt;h3 id="3-arithmetic-operations--the-alu-">3. Arithmetic Operations &amp;amp; The ALU ➕➖&lt;/h3>
&lt;p>Dive into the &amp;ldquo;brain&amp;rdquo; of the computer! Learn how the Arithmetic Logic Unit (ALU) performs lightning-fast calculations and logical operations, making complex computations possible. We&amp;rsquo;ll explore the algorithms behind addition, subtraction, multiplication, and division at the hardware level.&lt;/p>
&lt;h3 id="4-memory-systems-">4. Memory Systems 🧠&lt;/h3>
&lt;p>From lightning-fast caches to vast hard drives, explore the hierarchy and magic of how computers store and retrieve information efficiently. Understand why some memory is faster than others and how virtual memory provides an illusion of infinite storage.&lt;/p>
&lt;h3 id="5-instruction-set-architecture-isa-">5. Instruction Set Architecture (ISA) ✍️&lt;/h3>
&lt;p>This is the &amp;ldquo;vocabulary&amp;rdquo; the CPU understands! We&amp;rsquo;ll learn how software instructions are designed and executed, bridging the gap between your high-level code and the low-level hardware. Discover the fundamental operations a processor can perform.&lt;/p>
&lt;h3 id="6-control-unit-design-">6. Control Unit Design 🚦&lt;/h3>
&lt;p>The orchestrator of the entire system! Discover how the control unit directs the flow of data and instructions, ensuring every component plays its part in perfect harmony. It&amp;rsquo;s the conductor of the CPU&amp;rsquo;s orchestra.&lt;/p>
&lt;h3 id="7-inputoutput-io-systems-">7. Input/Output (I/O) Systems 🌐&lt;/h3>
&lt;p>How do computers talk to the outside world? We&amp;rsquo;ll explore the mechanisms that allow your computer to interact seamlessly with keyboards, screens, networks, and other peripheral devices, managing their diverse speeds and requirements.&lt;/p>
&lt;h3 id="8-bus-structures-">8. Bus Structures 🛣️&lt;/h3>
&lt;p>The digital highways connecting all the internal components! Understand how data travels reliably and efficiently between the CPU, memory, and I/O devices, and the different types of buses that facilitate this communication.&lt;/p>
&lt;hr>
&lt;div class="flex px-4 py-3 mb-6 rounded-md bg-primary-100 dark:bg-primary-900">
&lt;span class="pr-3 pt-1 text-primary-600 dark:text-primary-300">
&lt;svg height="24" xmlns="http://www.w3.org/2000/svg" viewBox="0 0 24 24">&lt;path fill="none" stroke="currentColor" stroke-linecap="round" stroke-linejoin="round" stroke-width="1.5" d="m11.25 11.25l.041-.02a.75.75 0 0 1 1.063.852l-.708 2.836a.75.75 0 0 0 1.063.853l.041-.021M21 12a9 9 0 1 1-18 0a9 9 0 0 1 18 0m-9-3.75h.008v.008H12z"/>&lt;/svg>
&lt;/span>
&lt;span class="dark:text-neutral-300">By the end of this course, you&amp;rsquo;ll not only understand &lt;em>how&lt;/em> computers work at their core but also &lt;em>why&lt;/em> they are designed the way they are. This foundational knowledge is essential for anyone aspiring to build, optimize, or secure computing systems!&lt;/span>
&lt;/div>
&lt;hr>
&lt;p>Did you find this page helpful? Consider sharing it 🙌&lt;/p></description></item><item><title>Experience</title><link>https://ygao36buffalo.github.io/experience/</link><pubDate>Tue, 24 Oct 2023 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/experience/</guid><description/></item><item><title>SignRing: Continuous American Sign Language Recognition Using IMU Rings and Virtual IMU Data</title><link>https://ygao36buffalo.github.io/publication/li-signring-2023/</link><pubDate>Fri, 01 Sep 2023 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/li-signring-2023/</guid><description/></item><item><title>An End-to-End Energy-Efficient Approach for Intake Detection With Low Inference Time Using Wrist-Worn Sensor</title><link>https://ygao36buffalo.github.io/publication/wei-end-end-2023/</link><pubDate>Tue, 01 Aug 2023 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/wei-end-end-2023/</guid><description/></item><item><title>SmartASL: \"Point-of-Care\" Comprehensive ASL Interpreter Using Wearables</title><link>https://ygao36buffalo.github.io/publication/jin-smartasl-2023/</link><pubDate>Thu, 01 Jun 2023 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/jin-smartasl-2023/</guid><description/></item><item><title>EarCommand: \"Hearing\" Your Silent Speech Commands In Ear</title><link>https://ygao36buffalo.github.io/publication/jin-earcommand-2022/</link><pubDate>Fri, 01 Jul 2022 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/jin-earcommand-2022/</guid><description/></item><item><title>PPGface: Like What You Are Watching? Earphones Can \"Feel\" Your Facial Expressions</title><link>https://ygao36buffalo.github.io/publication/choi-ppgface-2022/</link><pubDate>Fri, 01 Jul 2022 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/choi-ppgface-2022/</guid><description/></item><item><title>Detecting Screen Presence with Activity-Oriented RGB Camera in Egocentric Videos</title><link>https://ygao36buffalo.github.io/publication/adate-detecting-2022/</link><pubDate>Tue, 01 Mar 2022 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/adate-detecting-2022/</guid><description/></item><item><title>SonicFace: Tracking Facial Expressions Using a Commodity Microphone Array</title><link>https://ygao36buffalo.github.io/publication/gao-sonicface-2021/</link><pubDate>Wed, 01 Dec 2021 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-sonicface-2021/</guid><description/></item><item><title>SonicASL: An Acoustic-based Sign Language Gesture Recognizer Using Earphones</title><link>https://ygao36buffalo.github.io/publication/jin-sonicasl-2021/</link><pubDate>Tue, 01 Jun 2021 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/jin-sonicasl-2021/</guid><description/></item><item><title>Voice In Ear: Spoofing-Resistant and Passphrase-Independent Body Sound Authentication</title><link>https://ygao36buffalo.github.io/publication/gao-voice-2021/</link><pubDate>Mon, 01 Mar 2021 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-voice-2021/</guid><description/></item><item><title>ThermoTag: A Hidden ID of 3D Printers for Fingerprinting and Watermarking</title><link>https://ygao36buffalo.github.io/publication/gao-thermotag-2021/</link><pubDate>Fri, 01 Jan 2021 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-thermotag-2021/</guid><description/></item><item><title>EchoWhisper: Exploring an Acoustic-based Silent Speech Interface for Smartphone Users</title><link>https://ygao36buffalo.github.io/publication/gao-echowhisper-2020/</link><pubDate>Tue, 01 Sep 2020 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-echowhisper-2020/</guid><description/></item><item><title>Heart Monitor Using Flexible Capacitive ECG Electrodes</title><link>https://ygao36buffalo.github.io/publication/gao-heart-2020/</link><pubDate>Wed, 01 Jul 2020 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-heart-2020/</guid><description/></item><item><title>EarEcho: Using Ear Canal Echo for Wearable Authentication</title><link>https://ygao36buffalo.github.io/publication/gao-earecho-2019/</link><pubDate>Sun, 01 Sep 2019 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-earecho-2019/</guid><description/></item><item><title>Watching and Safeguarding Your 3D Printer: Online Process Monitoring Against Cyber-Physical Attacks</title><link>https://ygao36buffalo.github.io/publication/gao-watching-2018/</link><pubDate>Sat, 01 Sep 2018 00:00:00 +0000</pubDate><guid>https://ygao36buffalo.github.io/publication/gao-watching-2018/</guid><description/></item></channel></rss>